USTCzzl / HRG_Net

high resolution network for robotic grasping

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

HRG_Net

To reduce the spatial dimensional inaccuracy due to upsampling in the traditional CNN framework, we develop a novel grasping visual architecture referred to as High resolution grasp nerual network (HRG-Net), a parallel-branch structure that always maintains a high-resolution representation and repeatedly exchanges information across resolutions.

This repository contains the implementation of the High Resolution Grasp Nerual Network(HRG-Net) from the paper: A Robotic Visual Grasping Design: Rethinking Convolution Neural Network with High-Resolutions

arxiv|video

Contact

Any questions or comments contact mail.

If you use this work, please cite:

@article{zhou2022robotic,
title={A Robotic Visual Grasping Design: Rethinking Convolution Neural Network with High-Resolutions},
author={Zhou, Zhangli and Wang, Shaochen and Chen, Ziyang and Cai, Mingyu and Kan, Zhen},
journal={arXiv preprint arXiv:2209.07459},
 year={2022}
}

Installation

conda create -n hrgnet python==3.8
conda activate hrgnet
pip install -r requirements.txt / conda env create -f environment.yaml

Datasets

Cornell | Jacquard | multiobject(Just for test)

Training

python train_hrgnet.py

Evaluation/Visualisation

For validation and visualization purposes, we provide our previously trained model

python evaluation_grasp.py    # For Cornell and Jacquard dataset

python evaluation_heatmap.py  # For Cornell and Jacquard dataset

python multi_grasp_visualization.py # For multiobject dataset

Acknowledgement

Code heavily inspired and modified from https://github.com/dougsm/ggcnn. The code for the experiments related to the robot in the physical environment will be released later. For more of our recent work please follow link

About

high resolution network for robotic grasping

License:MIT License


Languages

Language:Python 100.0%