Show HN: Controlling 3D models with voice and hand gestures

3 weeks ago 1

Control 3D models using hand gestures and voice commands in real-time.

An interactive web app built with threejs, mediapipe computer vision, web speech API, and rosebud AI.

  • Say "drag", "rotate", "scale", or "animate" to change the interaction mode
  • Pinch fingers to control the 3D model
  • Drag/drop a new 3D model onto the page to import it (GLTF format only for now)

Video | Live Demo

  • Modern web browser with WebGL support
  • Camera / microphone access
  • Three.js for 3D rendering
  • MediaPipe for hand tracking and gesture recognition
  • Web Speech API for speech recognition
  • HTML5 Canvas for visual feedback
  • JavaScript for real-time interaction
# Clone this repository git clone https://github.com/collidingScopes/3d-model-playground # Navigate to the project directory cd 3d-model-playground # Serve with your preferred method (example using Python) python -m http.server

Then navigate to http://localhost:8000 in your browser.

MIT License

You might also like some of my other open source projects:

If you found this tool useful, feel free to buy me a coffee.

My name is Alan, and I enjoy building open source software for computer vision, games, and more. This would be much appreciated during late-night coding sessions!

Buy Me A Coffee

Read Entire Article