The official PyTorch implementation of the paper
SoundSil-DS: Deep Denoising and Segmentation of Sound-field Images with Silhouettes
The SoundSil-DS is a deep learning model for noise reduction in sound-field images with object silhouettes measured by optical methods, such as interferometer and holography. This is a continuous work of Deep Sound-Field Denoiser. It treats the complex-valued amplitude of the sound field in the frequency domain as a 2-channel image consisting of real and imaginary parts, and performs noise reduction and object-silhouette segmentation using a network based on CascadedGaze. The network has been trained using a sound-field image dataset we created using 2D acoustic simulations. The dataset includes noisy data with additive Gaussian white noise.
Our code is based on CascadedGaze.
-
Download trained weights and place them in the 'trained_weights' directory.
-
Download dataset and place them in the 'dataset' directory. When you run evaluation.py, 'evaluatation' directory from the dataset is required. When you run train.py, 'training' and 'validation' directories are required.
-
Install dependencies.
pip install -r requirements.txt
demo.ipynb provides a simple demo including loading a sound field from evaluation dataset, performing denoising and segmentation by pretrained weights, and displaying true, noisy, and denoised images.
To evaluate metrics and save denoised data on evaluation dataset, run
python evaluate.py --config config.yml
You can specify parameters for evaluation by the properties below 'eval' section in config.yml. The evaluation results will be saved into 'save_dir' directory of 'evaluation' section in the yaml file, a sub directory with the timestamp as its name will be automatically generated.
To train your model, run
python train.py --config config.yml
You can specify parameters for training by the properties below 'train' and 'validation' sections in config.yml.
Read the NTTSoftwareLicenseAgreement.pdf.
If you use SoundSil-DS, or this codebase in your work, please consider citing this work:
@misc{tanigawa2024soundsilds,
title={SoundSil-DS: Deep Denoising and Segmentation of Sound-field Images with Silhouettes},
author={Risako Tanigawa and Kenji Ishikawa and Noboru Harada and Yasuhiro Oikawa},
year={2024},
eprint={2411.07517},
archivePrefix={arXiv},
primaryClass={eess.SP},
url={https://arxiv.org/abs/2411.07517},
}