Skip to content

gaocong13/Learning-to-See-Forces

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

15 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Learning-to-See-Forces

Project | Paper | Slide

Pytorch implementation of surgical force prediction using da Vinci robot. This paper is awarded Best Paper Award (second place).

Learning to See Forces: Surgical Force Prediction with RGB-Point Cloud Temporal Convolutional Networks
Cong Gao, Xingtong Liu, Michael Peven, Mathias Unberath, and Austin Reiter
Computer Assisted and Robotic Endoscopy, MICCAI workshop 2018.

Obtaining haptic feedback is one of the major limitations during robot assisted surgical procedures. We propose the use of “visual cues” to infer forces from tissue deformation. Endoscopic video is a passive sensor that is freely available, in the sense that any minimally-invasive procedure already utilizes it. To this end, we employ deep learning to infer forces from video as an attractive low-cost and accurate alternative to typically complex and expensive hardware solutions.

Setup

Prerequisites

  • Linux or OSX
  • NVIDIA GPU + CUDA

Getting Started

git clone https://github.com/gaocong13/Learning-to-See-Forces.git
cd Learning-to-See-Forces/src
  • Extract PointNet feature and VGG16 feature from last layers:
python extract_ptnet_feature.py
python extract_vgg_feature.py
  • Train the model
mkdir ../model ../result
python trainTCN_main.py

Datasets

An example dataset is provided (./data/). It contains RGB image data (./data/rgbimage1-1), depth image data (./data/depthimage1-1) and measured grountruth force (./data/Force_ori). All are sorted and synchronized accordingly. Please contact cgao11@jhu.edu if you want to use the full dataset.

Models

PointNet pretrained model is uploaded (./data/pt_net_00041.pt). Please contact cgao11@jhu.edu if you want to use the full extracted features and final TCN model.

Citation

If you use this code for your research, please cite our paper Learning to See Forces: Surgical Force Prediction with RGB-Point Cloud Temporal Convolutional Networks :

@incollection{gao2018learning,
title={Learning to See Forces: Surgical Force Prediction with RGB-Point Cloud Temporal Convolutional Networks},
author={Gao, Cong and Liu, Xingtong and Peven, Michael and Unberath, Mathias and Reiter, Austin},
booktitle={OR 2.0 Context-Aware Operating Theaters, Computer Assisted Robotic Endoscopy, Clinical Image-Based Procedures, and Skin Image Analysis},
pages={118--127},
year={2018},
publisher={Springer}
}

Acknowledgments

This work was funded by an Intuitive Surgical Sponsored Research Agreement.

About

pytorch implementation of "Learning to See Forces" project

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages