From 044508b05fc381cbb10e16d2868d4559e623b4de Mon Sep 17 00:00:00 2001
From: Tai-Wang
Date: Fri, 14 Jun 2024 17:48:39 +0800
Subject: [PATCH] Update MMScan in the README
---
README.md | 48 ++++++++++--------------------------------------
1 file changed, 10 insertions(+), 38 deletions(-)
diff --git a/README.md b/README.md
index b3eced6..613fe2b 100644
--- a/README.md
+++ b/README.md
@@ -1,47 +1,11 @@
EmbodiedScan: A Holistic Multi-Modal 3D Perception Suite Towards Embodied AI
-
- Tai Wang*
- Xiaohan Mao*
- Chenming Zhu*
- Runsen Xu
- Ruiyuan Lyu
- Peisen Li
- Xiao Chen
-
- Wenwei Zhang
- Kai Chen
- Tianfan Xue
- Xihui Liu
- Cewu Lu
- Dahua Lin
- Jiangmiao Pang
-
- Shanghai AI Laboratory Shanghai Jiao Tong University The University of Hong Kong
-
- The Chinese University of Hong Kong Tsinghua University
-
-
-
+This repository contains EmbodiedScan-series works for holistic multi-modal 3D perception, currently including [EmbodiedScan](https://tai-wang.github.io/embodiedscan/) & [MMScan](https://tai-wang.github.io/mmscan/).
-[![arXiv](https://img.shields.io/badge/arXiv-2312.16170-blue)](https://arxiv.org/abs/2312.16170)
-[![](https://img.shields.io/badge/Paper-%F0%9F%93%96-blue)](./assets/EmbodiedScan.pdf)
-[![](https://img.shields.io/badge/Project-%F0%9F%9A%80-blue)](https://tai-wang.github.io/embodiedscan)
-
## 🤖 [Demo](https://tai-wang.github.io/embodiedscan)
@@ -81,6 +45,7 @@ Building upon this database, we introduce a baseline framework named Embodied
## 🔥 News
+- \[2024-06\] The report of our follow-up work with the most-ever hierarchical grounded language annotations, [MMScan](https://tai-wang.github.io/mmscan/), has been released. Welcome to talk with us about EmbodiedScan and MMScan at Seattle, CVPR 2024!
- \[2024-04\] We release all the baselines with pretrained models and logs. Welcome to try and play with them on our demo data! Note that we rename some keys in the multi-view 3D detection and visual grounding model. Please re-download the pretrained models if you just use our code for inference.
- \[2024-03\] The challenge test server is also online [here](https://huggingface.co/spaces/AGC2024/visual-grounding-2024). Looking forward to your strong submissions!
- \[2024-03\] We first release the data and baselines for the challenge. Please fill in the [form](https://docs.google.com/forms/d/e/1FAIpQLScUXEDTksGiqHZp31j7Zp7zlCNV7p_08uViwP_Nbzfn3g6hhw/viewform?usp=sf_link) to apply for downloading the data and try our baselines. Welcome any feedback!
@@ -297,18 +262,25 @@ Please see the [paper](./assets/EmbodiedScan.pdf) for more details of our benchm
- \[x\] Release codes for our baselines and benchmarks.
- \[ \] Release codes for all the other methods.
- \[ \] Full release and further updates.
+- \[ \] Release MMScan data and codes.
## 🔗 Citation
If you find our work helpful, please cite:
```bibtex
-@inproceedings{wang2023embodiedscan,
+@inproceedings{embodiedscan,
title={EmbodiedScan: A Holistic Multi-Modal 3D Perception Suite Towards Embodied AI},
author={Wang, Tai and Mao, Xiaohan and Zhu, Chenming and Xu, Runsen and Lyu, Ruiyuan and Li, Peisen and Chen, Xiao and Zhang, Wenwei and Chen, Kai and Xue, Tianfan and Liu, Xihui and Lu, Cewu and Lin, Dahua and Pang, Jiangmiao},
year={2024},
booktitle={IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
}
+@inproceedings{mmscan,
+ title={MMScan: A Multi-Modal 3D Scene Dataset with Hierarchical Grounded Language Annotations},
+ author={Lyu, Ruiyuan and Wang, Tai and Lin, Jingli and Yang, Shuai and Mao, Xiaohan and Chen, Yilun and Xu, Runsen and Huang, Haifeng and Zhu, Chenming and Lin, Dahua and Pang, Jiangmiao},
+ year={2024},
+ booktitle={arXiv},
+}
```
If you use our dataset and benchmark, please kindly cite the original datasets involved in our work. BibTex entries are provided below.