-
Notifications
You must be signed in to change notification settings - Fork 0
Description
Hello,
Thank you so much for releasing your code! It's incredibly helpful.
I have a few questions regarding the fine-tuning procedure used for the "hearing anything anywhere" dataset, as I would like to replicate and better understand your approach:
Fine-tuning Procedure:
- Could you provide more details on the fine-tuning process?
- Are any layers frozen during fine-tuning? If so, which ones?
- What learning rate did you used? For how many steps?
Dataset Handling:
The "hearing anything anywhere" dataset does not contain panoramic depth maps. How do you handle the geometric context in the absence of this data? Is there a particular method or workaround you’ve implemented?
Room Configurations:
Are you using only the "base" room configurations, or do you also incorporate the other room variations such as Translation, Rotation, or Panel?
Reproducing Results:
Is there any additional information, configuration setting, or insight that would be helpful for successfully reproducing your results?
I really appreciate any clarifications you can provide!