This repository contains codes for multi-modality learning from the Multimodal Cognition group of Ant Group that have been integrated into AntMMF. AntMMF encapsulates standard multimodal functionalities including dataset management, data processing, training workflows, models, and modules, while also enabling custom extensions of these components.
- March, 2024: M2-RAAP was accepted by SIGIR 2024.
- February, 2024: release the code of bilingual multimodal CLIP-M2-Encoder, which was trained on our BM-6B bilingual dataset.
- December, 2023: release the code of SNP-S3, DMAE, and CNVid-3.5M.
- June, 2023: SNP-S3 was accepted by IEEE T-CSVT 2023.
- May, 2023: DMAE was accepted by ACM MultiMedia 2023.
- March, 2023: CNVid-3.5M was accepted by CVPR 2023.
- Dataset
- CNVid-3.5M (CVPR-2023): A large-scale public Chinese video-text pretraining dataset.
- Pretraining Methods
- SNP-S3 (IEEE T-CSVT 2023): Semantic enhancement for video pretraining.
- DMAE (ACM MM-2023): Dual-Modal attention-enhanced Text-Video Retrieval with triplet partial margin contrastive learning.
- EVE: Efficient zero-shot video editing.
- Please follow the steps below to initialize the environment of the AntMMF.
# Build a new environment.
conda create -n antmmf python=3.8
source activate antmmf
# Clone this project.
cd /YourPath/
git clone https://github.com/alipay/Ant-Multi-Modal-Framework
# Install the required packages.
cd antmmf
pip install -r requirements.txt
If you find AntMMF useful for your work, please consider citing:
@misc{qp2023AntMMF,
author = {Qingpei, Guo and Xingning, Dong and Xiaopei, Wan and Xuzheng, Yu and Chen, Jiang and Xiangyuan, Ren and Kiasheng, Yao and Shiyu, Xuan},
title = {AntMMF: Ant Multi-Modal Framework},
howpublished = {\url{https://github.com/alipay/Ant-Multi-Modal-Framework}},
year = {2023}
}
This project is licensed under the Apache 2.0 license, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Our code is based on FAIR mmf. We thank the authors for their wonderful open-source efforts.
🙋 For help or issues with this codebase, please submit an issue.
⭐ We are hiring, if you are interested in our work, please feel free to contact Qingpei Guo([email protected]).