SurgSync: Time-Synchronized Multi-modal Data Collection Framework and Dataset for Surgical Robotics

Currently under review, submitted in 2025

Overview

This research represents Multi-Modal Time-Synchronized Data Collection and Post-Processing Framework for Surgical Robotics, along with datasets and AI-driven applications (stay tuned!).

Description
SurgSync Overview

Key Contributions

  • Lead and manage 8-person cross-functional team, delivering IRB-compliant data collection and analysis that accelerated publication timeline and demo readiness
  • Build time-synchronized (multi-modality sync time latency within 10ms) multi-modal data collection pipeline (vision + kinematics + sensor) and large (100+ instances) ex-vivo dataset
  • Enable a novel kinematic projection approach and downstream optical flow/depth estimation using deep learning
  • Design and implement a custom capacitive contact sensor to acquire the ground truth of tool-tissue contact
  • Integrate a modern chip-on-tip endoscope with the dVRK seamlessly, enabling high-quality image data acquisition
  • Design a novel data annotation application with graphic user interface (GUI) using PyQt for manual label annotations
  • The data collection pipeline has been employed for efforts on Open-H-Embodiment

Future/Ongoing Work

  • Investigate surgical robot tool-tissue contact detection using a multi-modal deep learning approach.

Tags: , , , , , , ,