Skip to content

Commit

Permalink
Summary
Browse files Browse the repository at this point in the history
  • Loading branch information
DmitryRyumin committed Feb 12, 2024
1 parent 1545b93 commit 12af6e7
Show file tree
Hide file tree
Showing 2 changed files with 3 additions and 3 deletions.
2 changes: 1 addition & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -348,7 +348,7 @@ Contributions to improve the completeness of this list are greatly appreciated.
<a href="https://github.com/DmitryRyumin/CVPR-2023-Papers/blob/main/sections/2023/main/multimodal-learning.md"><img src="https://img.shields.io/badge/61-1D7FBF" alt="Open Code"></a>
</td>
<td>
<a href="https://github.com/DmitryRyumin/CVPR-2023-Papers/blob/main/sections/2023/main/multimodal-learning.md"><img src="https://img.shields.io/badge/50-FF0000" alt="Videos"></a>
<a href="https://github.com/DmitryRyumin/CVPR-2023-Papers/blob/main/sections/2023/main/multimodal-learning.md"><img src="https://img.shields.io/badge/51-FF0000" alt="Videos"></a>
</td>
</tr>
<tr>
Expand Down
4 changes: 2 additions & 2 deletions sections/2023/main/multimodal-learning.md
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@

## Multimodal Learning

![Section Papers](https://img.shields.io/badge/Section%20Papers-89-42BA16) ![Preprint Papers](https://img.shields.io/badge/Preprint%20Papers-75-b31b1b) ![Papers with Open Code](https://img.shields.io/badge/Papers%20with%20Open%20Code-61-1D7FBF) ![Papers with Video](https://img.shields.io/badge/Papers%20with%20Video-50-FF0000)
![Section Papers](https://img.shields.io/badge/Section%20Papers-89-42BA16) ![Preprint Papers](https://img.shields.io/badge/Preprint%20Papers-75-b31b1b) ![Papers with Open Code](https://img.shields.io/badge/Papers%20with%20Open%20Code-61-1D7FBF) ![Papers with Video](https://img.shields.io/badge/Papers%20with%20Video-51-FF0000)

| **Title** | **Repo** | **Paper** | **Video** |
|-----------|:--------:|:---------:|:---------:|
Expand All @@ -39,7 +39,7 @@
| CNVid-3.5M: Build, Filter, and Pre-Train the Large-Scale Public Chinese Video-Text Dataset | [![GitHub](https://img.shields.io/github/stars/CNVid/CNVid-3.5M?style=flat)](https://github.com/CNVid/CNVid-3.5M) | [![thecvf](https://img.shields.io/badge/pdf-thecvf-7395C5.svg)](https://openaccess.thecvf.com/content/CVPR2023/papers/Gan_CNVid-3.5M_Build_Filter_and_Pre-Train_the_Large-Scale_Public_Chinese_Video-Text_CVPR_2023_paper.pdf) | [![YouTube](https://img.shields.io/badge/YouTube-%23FF0000.svg?style=for-the-badge&logo=YouTube&logoColor=white)](https://www.youtube.com/watch?v=ywJfAg4wvr0) |
| Egocentric Audio-Visual Object Localization | [![GitHub](https://img.shields.io/github/stars/WikiChao/Ego-AV-Loc?style=flat)](https://github.com/WikiChao/Ego-AV-Loc) | [![thecvf](https://img.shields.io/badge/pdf-thecvf-7395C5.svg)](https://openaccess.thecvf.com/content/CVPR2023/papers/Huang_Egocentric_Audio-Visual_Object_Localization_CVPR_2023_paper.pdf) <br /> [![arXiv](https://img.shields.io/badge/arXiv-2303.13471-b31b1b.svg)](http://arxiv.org/abs/2303.13471) | [![YouTube](https://img.shields.io/badge/YouTube-%23FF0000.svg?style=for-the-badge&logo=YouTube&logoColor=white)](https://www.youtube.com/watch?v=0-_XJJ1JLmM) |
| Learning Visual Representations via Language-Guided Sampling | [![GitHub](https://img.shields.io/github/stars/mbanani/lgssl?style=flat)](https://github.com/mbanani/lgssl) | [![thecvf](https://img.shields.io/badge/pdf-thecvf-7395C5.svg)](https://openaccess.thecvf.com/content/CVPR2023/papers/Banani_Learning_Visual_Representations_via_Language-Guided_Sampling_CVPR_2023_paper.pdf) <br /> [![arXiv](https://img.shields.io/badge/arXiv-2302.12248-b31b1b.svg)](http://arxiv.org/abs/2302.12248) | [![YouTube](https://img.shields.io/badge/YouTube-%23FF0000.svg?style=for-the-badge&logo=YouTube&logoColor=white)](https://www.youtube.com/watch?v=95I8DmUoJ2s) |
| Unite and Conquer: Plug and Play Multi-Modal Synthesis using Diffusion Models | [![GitHub](https://img.shields.io/github/stars/Nithin-GK/UniteandConquer?style=flat)](https://github.com/Nithin-GK/UniteandConquer) | [![thecvf](https://img.shields.io/badge/pdf-thecvf-7395C5.svg)](https://openaccess.thecvf.com/content/CVPR2023/papers/Nair_Unite_and_Conquer_Plug__Play_Multi-Modal_Synthesis_Using_Diffusion_CVPR_2023_paper.pdf) <br /> [![arXiv](https://img.shields.io/badge/arXiv-2212.00793-b31b1b.svg)](http://arxiv.org/abs/2212.00793) | :heavy_minus_sign: |
| Unite and Conquer: Plug and Play Multi-Modal Synthesis using Diffusion Models | [![GitHub Page](https://img.shields.io/badge/GitHub-Page-159957.svg)](https://nithin-gk.github.io/projectpages/Multidiff) <br /> [![GitHub](https://img.shields.io/github/stars/Nithin-GK/UniteandConquer?style=flat)](https://github.com/Nithin-GK/UniteandConquer) | [![thecvf](https://img.shields.io/badge/pdf-thecvf-7395C5.svg)](https://openaccess.thecvf.com/content/CVPR2023/papers/Nair_Unite_and_Conquer_Plug__Play_Multi-Modal_Synthesis_Using_Diffusion_CVPR_2023_paper.pdf) <br /> [![arXiv](https://img.shields.io/badge/arXiv-2212.00793-b31b1b.svg)](http://arxiv.org/abs/2212.00793) | [![YouTube](https://img.shields.io/badge/YouTube-%23FF0000.svg?style=for-the-badge&logo=YouTube&logoColor=white)](https://www.youtube.com/watch?v=N4EOwnhNzIk) |
| iQuery: Instruments As Queries for Audio-Visual Sound Separation | [![GitHub](https://img.shields.io/github/stars/JiabenChen/iQuery?style=flat)](https://github.com/JiabenChen/iQuery) | [![thecvf](https://img.shields.io/badge/pdf-thecvf-7395C5.svg)](https://openaccess.thecvf.com/content/CVPR2023/papers/Chen_iQuery_Instruments_As_Queries_for_Audio-Visual_Sound_Separation_CVPR_2023_paper.pdf) <br /> [![arXiv](https://img.shields.io/badge/arXiv-2212.03814-b31b1b.svg)](http://arxiv.org/abs/2212.03814) | [![YouTube](https://img.shields.io/badge/YouTube-%23FF0000.svg?style=for-the-badge&logo=YouTube&logoColor=white)](https://www.youtube.com/watch?v=EZ9CgknV9Z4) |
| Diverse Embedding Expansion Network and Low-Light Cross-Modality Benchmark for Visible-Infrared Person Re-Identification | [![GitHub](https://img.shields.io/github/stars/ZYK100/LLCM?style=flat)](https://github.com/ZYK100/LLCM) | [![thecvf](https://img.shields.io/badge/pdf-thecvf-7395C5.svg)](https://openaccess.thecvf.com/content/CVPR2023/papers/Zhang_Diverse_Embedding_Expansion_Network_and_Low-Light_Cross-Modality_Benchmark_for_Visible-Infrared_CVPR_2023_paper.pdf) <br /> [![arXiv](https://img.shields.io/badge/arXiv-2303.14481-b31b1b.svg)](http://arxiv.org/abs/2303.14481) | [![YouTube](https://img.shields.io/badge/YouTube-%23FF0000.svg?style=for-the-badge&logo=YouTube&logoColor=white)](https://www.youtube.com/watch?v=oMIRqc-Fq5c) |
| PiMAE: Point Cloud and Image Interactive Masked Autoencoders for 3D Object Detection | [![GitHub](https://img.shields.io/github/stars/BLVLab/PiMAE?style=flat)](https://github.com/BLVLab/PiMAE) | [![thecvf](https://img.shields.io/badge/pdf-thecvf-7395C5.svg)](https://openaccess.thecvf.com/content/CVPR2023/papers/Chen_PiMAE_Point_Cloud_and_Image_Interactive_Masked_Autoencoders_for_3D_CVPR_2023_paper.pdf) <br /> [![arXiv](https://img.shields.io/badge/arXiv-2303.08129-b31b1b.svg)](http://arxiv.org/abs/2303.08129) | [![YouTube](https://img.shields.io/badge/YouTube-%23FF0000.svg?style=for-the-badge&logo=YouTube&logoColor=white)](https://www.youtube.com/watch?v=rcs8DYAwugQ) |
Expand Down

0 comments on commit 12af6e7

Please sign in to comment.