Hiwi (Student Assistant Researcher) in KI Fabrik

Published:

1. Multi-Modal Demonstration Collection

I collected human demonstration data in a multi-modal fashion to enable skill planning in robotic manipulation using Large Language Models (LLMs). This includes:

  • Data Collection: Capturing human motion, force feedback, and vision-based interactions.
  • Skill Representation: Translating collected data into a structured format for robot skill learning.

2. Skill-Based Robot Control (MIOS)

MIOS (Skill-based Robot Control Platform) was extended by programming new skills and controllers:

  • Motion Primitives: Developing reusable skill modules for diverse robotic tasks.
  • New Controller: Designing controllers for robust execution of skills.

3. Haptic Sensing Integration

To enhance manipulation, external force sensing was integrated into MIOS, allowing:

  • Force Feedback Processing: Utilizing external force-torque sensors to enhance control precision.

4. Robot Vision System Implementation

Several vision-based techniques were implemented for perception tasks:

  • Hand-Eye Calibration: Establishing accurate alignment between camera and robot.
  • 3D Reconstruction: Utilizing BundleSDF for environment mapping.
  • 6D Pose Estimation: Leveraging FoundationPose for object localization.

5. Web Development

I contributed to web development for Project LEMMo-Plan: LLM-Enhanced Learning from Mutli-Modal Demonstration for Planning Sequential Contact-Rich Manipulation Tasks:

  • Project Website: Developed and maintained LEMMo-Plan.

6. Contribution to Deformable Linear Object Assembly

I continued my work on deformable linear object assembly by implementing:

  • Virtual-Tactile-Based Robot Position Local Correction: Ensuring accurate assembly using tactile feedback.
  • Real Robot Experiments: Conducted hardware experiments to validate the proposed approach.
  • Submitted to IROS 2025.