Xiulong Liu, Anurag Kumar, Paul Calamia, Sebasti`a V. Amengual, Calvin Murdock , Ishwarya Ananthabhotla , Philip Robinson , Eli Shlizerman , Vamsi Krishna Ithapu , Ruohan Gao
(If you find this project helpful, please give us a star ⭐ on this GitHub repository to support us.)
xRIR is a novel and generalizable framework for cross-room RIR prediction. The approach demonstrates strong performance not only on large-scale synthetic dataset but also achieves decent performance when adapted to real acoustic scenes. This repository contains the unofficial implementation of the CVPR 2025 paper.
Clone the repository and create a conda environment:
git clone https://github.com/DragonLiu1995/xRIR_code.git
conda create -n xRIR python=3.8
conda activate xRIR
pip install torch==2.0.1+cu117 torchvision==0.15.2+cu117 torchaudio==2.0.2+cu117 \
-f https://download.pytorch.org/whl/torch_stable.html
Install dependencies: pip install -r requirements.txt
Check out the official dataset repository at for details: https://github.com/facebookresearch/AcousticRooms. Download all and unzip all *.zip files to a data folder.
Here we provide checkpoints for xRIR under 8-shot scenario for both seen and unseen splits in AcousticRooms dataset. To evaluate the model:
export PYTHONPATH=$PYTHONPATH:[repo_directory]
- Run:
python eval_unseen.py
for unseen test split, and
python eval_seen.py
for seen test split.
Download our pretrained model checkpoints from here
If you have any questions or need further assistance, feel free to reach out to us:
- Xiulong Liu: [email protected]
If you use this code for your research, please cite our work:
@inproceedings{liu2025hearing,
title={Hearing Anywhere in Any Environment},
author={Liu, Xiulong and Kumar, Anurag and Calamia, Paul and Amengual, Sebastia V and Murdock, Calvin and Ananthabhotla, Ishwarya and Robinson, Philip and Shlizerman, Eli and Ithapu, Vamsi Krishna and Gao, Ruohan},
booktitle={Proceedings of the Computer Vision and Pattern Recognition Conference},
pages={5732--5741},
year={2025}
}