FVPLab / Argus-3D

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Argus-3D: Pushing Auto-regressive Models for 3D Shape Generation at Capacity and Scalability

Paper | Project Page

Installation

You can create an anaconda environment called argus-3d using

conda env create -f environment.yaml
conda activate argus-3d

Next, compile the extension modules. You can do this via

python setup.py build_ext --inplace

Generation

Download stage1 checkpoint and place it into output/PR256_ED512_EN8192.

Download stage2 checkpoint and place it into output/PR256_ED512_EN8192/class-guide/transformer3072_24_32.

Then you can try class-guide generation by run:

python generate_class-guide.py --batch_size 16 --cate chair

This script should create a folder output/PR256_ED512_EN8192/class-guide/transformer3072_24_32/class_cond where the output meshes are stored.

Note: Our model requires significant memory, and it's recommended to run it on a GPU with high VRAM capacity (40GB or above). Generating a single mesh on the A100 (80GB) takes approximately 50 seconds on average, while on V100 (32GB) it takes ~6 minutes.

Dataset

The occupancies, point clouds, and supplementary rendered images based on the Objaverse dataset can be downloaded from https://huggingface.co/datasets/BAAI/Objaverse-MIX

Coming Soon

  • Image-guide generation
  • Text-guide generation
  • Training code

Shout-outs

Thanks to everyone who makes their code and models available.

Thanks for open-sourcing!

BibTeX

@inproceedings{inproceedings,
      author = {Luo, Simian and Qian, Xuelin and Fu, Yanwei and Zhang, Yinda and Tai, Ying and Zhang, Zhenyu and Wang, Chengjie and Xue, Xiangyang},
      year = {2023},
      month = {10},
      pages = {14093-14103},
      title = {Learning Versatile 3D Shape Generation with Improved Auto-regressive Models},
      doi = {10.1109/ICCV51070.2023.01300}
}
@misc{qian2024pushing,
      title={Pushing Auto-regressive Models for 3D Shape Generation at Capacity and Scalability}, 
      author={Xuelin Qian and Yu Wang and Simian Luo and Yinda Zhang and Ying Tai and Zhenyu Zhang and Chengjie Wang and Xiangyang Xue and Bo Zhao and Tiejun Huang and Yunsheng Wu and Yanwei Fu},
      year={2024},
      eprint={2402.12225},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}

About


Languages

Language:Python 75.7%Language:C++ 10.0%Language:C 6.8%Language:Cython 4.3%Language:Mako 3.2%