**Acknowledgements**

This work was supported by the Tsinghua University Initiative Scientific Research Program No. 2019Z08QCX15.

**References**

**14**(1):1-8

pp. 1957-1964

1703.09312. 2017:1-12

1709.06670. 2017:1-16

[6] Mahler J, Matl M, Liu X, Li A, Gealy D, Goldberg K. Dexnet 3.0: Computing robust robot vacuum suction grasp targets in point clouds using a new analytic model and deep learning. arXiv preprint arXiv:

[7] Zhang T, McCarthy Z, Jowl O, Lee D,

manipulation tasks from virtual reality

International Conference on Robotics and Automation (ICRA). IEEE; 2018.

Chen X, Goldberg K, et al. Deep imitation learning for complex

teleoperation. In: 2018 IEEE

pp. 1-8

**129**

2014;**30**(2):289-309

[1] Bohg J, Morales A, Asfour T, Kragic D. Data-driven grasp synthesis survey. IEEE Transactions on Robotics.

*Visual-Tactile Fusion for Robotic Stable Grasping DOI: http://dx.doi.org/10.5772/intechopen.91455*

> [8] Vlack K, Mizota T, Kawakami N, Kamiyama K, Kajimoto H, Tachi S. Gelforce: A vision-based traction field computer interface. In: CHI'05

Extended Abstracts on Human Factors in Computing Systems. ACM; 2005.

[9] Goldfeder C, Ciocarlie M, Dang H, Allen PK. The Columbia grasp database. In: IEEE International Conference on Robotics and Automation, 2009. ICRA'09. IEEE; 2009. pp. 1710-1716

[10] Sermanet P, Lynch C, Chebotar Y, Hsu J, Jang E, Schaal S, et al. Timecontrastive networks: Self-supervised learning from video. arXiv preprint

[11] Pinto L, Gupta A. Supersizing selfsupervision: Learning to grasp from 50k tries and 700 robot hours. In: 2016 IEEE International Conference on Robotics and Automation (ICRA). IEEE; 2016.

arXiv:1704.06888. 2017:1-15

[12] Li J, Dong S, Adelson E. Slip detection with combined tactile and visual information. arXiv preprint arXiv:1802.10153. 2018:1-6

[13] Kobayashi F, Kojima F, Nakamoto H, Kida Y, Imamura N, Shirasawa H. Slip detection with multiaxis force/torque sensor in universal robot hand. International Journal of Applied Electromagnetics and Mechanics. 2012;**39**(1–4):1047-1054

[14] Heyneman B, Cutkosky MR. Slip classification for dynamic tactile array sensors. The International Journal of Robotics Research. 2016;**35**(4):404-421

[15] Chebotar Y, Hausman K, Su Z, Molchanov A, Kroemer O, Sukhatme G, et al. BiGS: BioTac grasp stability dataset. In: ICRA 2016 Workshop on Grasping and Manipulation Datasets;

2016

pp. 3406-3413

pp. 1154-1155

[2] Levine S, Pastor P, Krizhevsky A, Ibarz J, Quillen D. Learning hand-eye coordination for robotic grasping with deep learning and large-scale data collection. The International Journal of Robotics Research. 2018;**37**(4–5):421-436

[3] Guo D, Sun F, Kong T, Liu H. Deep vision networks for real-time robotic grasp detection. International Journal of Advanced Robotic Systems. 2016;

[4] Mahler J, Pokorny FT, Hou B, Roderick M, Laskey M, Aubry M, et al. Dex-Net 1.0: A cloud-based network of 3D objects for robust grasp planning using a multi-armed bandit model with correlated rewards. In: 2016 IEEE International Conference on Robotics and Automation (ICRA). IEEE; 2016.

[5] Mahler J, Liang J, Niyaz S, Laskey M, Doan R, Liu X, et al. Dex-net 2.0: Deep learning to plan robust grasps with synthetic point clouds and analytic grasp metrics. arXiv preprint arXiv:
