Skip to content

ojusxe/lumo-playground

Repository files navigation

Lumo Playground

7th semester minor project @BTech @IKGPTU @CS Control 3D models using hand gestures and voice commands in real-time. lumo's playground


An interactive web app built with threejs, mediapipe computer vision, web speech API, and rosebud AI. While most applications of computer vision technologies like OpenCV have been through Python bindings, this project demonstrates the power and accessibility of JavaScript bindings for the same advanced computer vision capabilities, bringing real-time hand tracking directly to web browsers without additional installations.
  • Pinch fingers to control the 3D model, follow the instructions given
  • [WIP @custom-avatar] drag/drop a new 3D model onto the page to import it (GLB/GLTF format)
  • [WIP @speech-recognition] say "drag", "rotate", "scale", or "animate" to change the interaction mode

Requirements and permissions

Modern web browser with WebGL support & camera / microphone access

Technologies

  • Three.js for 3D rendering
  • MediaPipe for hand tracking and gesture recognition
  • Web Speech API for speech recognition
  • HTML5 Canvas for visual feedback
  • JavaScript for real-time interaction

Credits

About

enter lumo's playground where only hands are allowed. control 3D characters with hand gestures with presets of movements and actions https://lumo-playground.vercel.app

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors