CyLab Spatial Computing Interface
CyLab Logo

May 2023 - Aug 2023

CyLab, Carnegie Mellon University

Spatial Computing Researcher

Designing immersive interfaces for real-time human-computer interaction in spatial environments

Project Overview

Working with Professor Anthony Rowe at CyLab, I contributed to cutting-edge research in real-time human pose estimation and tracking. This project served as the foundation for my journey into Computer Vision and Human-Computer Interaction (HCI).

My work focused on creating an immersive, real-time VR/AR experience by integrating depth cameras, LiDARs, 3D modeling, and game engines. The system streamed a live representation of a workspace environment, including digital twins of robotic arms and human subjects, allowing users to remotely interact with the environment as if physically present.

This research explored the future of remote collaboration, teleoperation, and spatial computing interfaces, with applications in fields ranging from remote surgery to hazardous environment operations.

My Role

  • Human Pose Detection & Tracking
  • 3D Environment Visualization
  • VR/AR Interface Design
  • Real-time Data Processing

Technologies Used

Unity3DOpenCVPythonC#Azure KinectOculus SDK

The Challenge

Creating immersive spatial computing interfaces presents several significant technical and design challenges:

  • 1Real-time Performance: Processing and transmitting 3D data with minimal latency is crucial for immersive experiences
  • 2Accurate Tracking: Human pose estimation must be precise enough for meaningful interaction
  • 3Intuitive Interfaces: Users need to understand how to interact with virtual objects naturally
  • 4System Integration: Multiple hardware and software components must work together seamlessly
Spatial computing challenges

Technical Approach

01. Human Pose Detection & Tracking

Capturing Human Movement in Real-time

I developed a system that uses depth cameras and computer vision algorithms to detect and track human subjects in real-time. Key components included:

  • Multi-camera calibration for comprehensive spatial coverage
  • Skeletal tracking with 25 joint points for detailed pose estimation
  • Temporal filtering to reduce jitter and improve tracking stability
  • Occlusion handling to maintain tracking when body parts are hidden
Human pose detection system
3D environment visualization
02. 3D Environment Visualization

Creating Digital Twins of Physical Spaces

I created a system to generate and update digital representations of physical environments in real-time:

  • LiDAR-based 3D scanning to create accurate spatial maps
  • Dynamic object tracking to update the position of movable items
  • Texture mapping from RGB cameras to enhance visual fidelity
  • Optimized mesh generation for real-time rendering performance
03. VR/AR Interface Design

Creating Intuitive Spatial Interactions

I designed the interface that allowed users to interact with the virtual environment:

  • Gesture recognition for natural interaction with virtual objects
  • Spatial UI elements that integrate with the 3D environment
  • Visual feedback systems to indicate system state and interaction possibilities
  • Multi-user presence representation for collaborative scenarios
VR/AR interface design

User Experience Design

The final spatial computing interface integrated intuitive gesture controls, real-time feedback systems, and multi-user presence to create an immersive environment for remote collaboration and interaction.

Gesture Recognition

I designed a set of intuitive hand gestures for manipulating virtual objects, calibrated to be easily learned yet distinct enough to prevent accidental activation.

Visual Feedback

I implemented a comprehensive visual feedback system using color, animation, and spatial cues to indicate interactive elements and system status.

Multi-User Presence

I created avatar representations for remote collaborators with real-time pose updates and proximity-based interaction capabilities.

Results & Impact

The spatial computing interface demonstrated significant potential for remote operation and collaboration scenarios. Key achievements included:

  • 1Low Latency Interaction: Achieved end-to-end latency of under 100ms, creating a responsive feel
  • 2Accurate Pose Tracking: Skeletal tracking with average joint position error of less than 2cm
  • 3Intuitive Control: 90% success rate for first-time users completing a series of manipulation tasks
  • 4Scalable Architecture: System successfully tested with up to 4 simultaneous remote users
<100ms

End-to-end system latency

90%

First-time user task success rate

<2cm

Average joint position error

4+

Simultaneous remote users

"This spatial computing interface represents a significant advancement in remote collaboration technology. The system's ability to create an immersive, responsive environment opens up new possibilities for applications in hazardous environments, remote surgery, and distributed team collaboration."

AL

Prof. Anthony Rowe

Professor, CyLab

Potential Applications

Remote Surgery

Enabling surgeons to perform procedures from a distance with precise control and spatial awareness.

Hazardous Environment Operations

Allowing operators to control robots in dangerous locations while maintaining situational awareness.

Distributed Team Collaboration

Creating shared virtual workspaces for teams to collaborate on physical or digital projects remotely.

Lessons Learned

  • Latency Management: Even small delays can significantly impact user experience in spatial computing. Optimizing every part of the pipeline is essential.
  • Intuitive Gestures: Natural gestures that map to physical world interactions are more quickly adopted than abstract control schemes.
  • Sensor Fusion: Combining data from multiple sensors provides more robust tracking but requires careful calibration and synchronization.
  • User Comfort: Extended use of VR interfaces requires careful attention to ergonomics and cognitive load to prevent fatigue.

Future Work

  • Haptic Feedback: Integrating force feedback gloves to provide tactile sensation when interacting with virtual objects.
  • AI-Assisted Interaction: Implementing intelligent systems to predict user intent and assist with complex manipulations.
  • Mobile Deployment: Adapting the system for use with lightweight, mobile VR/AR headsets for field operations.
  • Cross-Platform Support: Extending the interface to work across different VR/AR hardware platforms.

Explore My Other CMU Projects

Check out my work with Xu Lab and the Abstract Lab at Carnegie Mellon.

© 2023 Sushmith Thuluva. All rights reserved.