JavaScript Quickies: Controlling 3D Objects with Hands 🤯

Hey guys, what’s up? We all at Team XenoX are really excited to inform you that we are starting a new series of articles called Javascript Quickies. These will be quick experiments that you guys can do in Javascript to explore something new in technology. Thanks to Javascript, we can just plug in various modules and create anything. The only limit is your imagination.

The Idea 💡

We all have our favorite moments from Sci-Fi movies. These moments are extra special for us developers because we can’t help but wonder how all the cool sci-fi tricks that we see on the screen could be turned into reality. Whenever I see something like that, my mind immediately jumps into top gear and I start thinking about all the technical possibilities. There’s a child-like fascination attached to it that I absolutely love.
I remember watching Iron Man as a teen and being completely amazed by the scene where he interacts with holographic objects in his lab. As I recalled that scene, I got to thinking whether I could create something similar, something that sparked the same kind of joy.

Of course, we don’t have all that juicy tech to create the exact same effect, at least not yet. But we can surely try to make something almost as cool from what we already have. So I made this cool little project over the weekend to share with you guys.
Buckle up, Avengers! Let’s create The Hand Trick.

Requirements 🧺

I made this using Vanilla Javascript. So you should have a basic understanding of Javascript to understand this tutorial. Other than that, I have used two libraries in this:
1. Three.js 👉🏼 link
2. Handtrack.js 👉🏼 link
That’s it.

Let’s code now 👩🏽‍💻

The HTML side of the code is very simple. We are just including the libraries here and adding the div to render the camera feed in the browser:

<html lang=”en">
<meta charset="UTF-8" />
<meta name="viewport" content="width=device-width, initial-scale=1.0" />
<meta http-equiv="X-UA-Compatible" content="ie=edge" />
<title>The Hand Trick</title>
<link rel="stylesheet" href="css/style.css" />
<!– Video for handtracker –>
<div class="tracker">
<video id="myvideo"></video>
<canvas id="canvas" class="border"></canvas>
<button id="trackbutton" disabled onclick="toggleVideo()">Button</button>
<div id="updatenote">hello</div>
<div class="data">
<div class="hand-1">
<p id="hand-x">X: <span>0</span></p>
<p id="hand-y">Y: <span>0</span></p>
<script src="js/three.js"></script>
<script src=""></script>
<script src="js/scene.js"></script>

Once that is done, let’s quickly jump back to the Javascript side of things. If you know Three.js, you can skip this part. For others, we are creating a scene here by setting the details about it.
// Setting scene for 3D Object
var scene = new THREE.Scene();
var camera = new THREE.PerspectiveCamera(
window.innerWidth / window.innerHeight,
var vector = new THREE.Vector3();
var renderer = new THREE.WebGLRenderer();
renderer.setSize(window.innerWidth, window.innerHeight);

After that, let’s create a 3D Object to render on the scene. Here we will be defining the Geometry of the box and the type of the mesh’s material.
// Creating 3D object
var geometry = new THREE.BoxGeometry(1, 2, 1);
var material = new THREE.MeshBasicMaterial({
color: "rgba(3, 197, 221, 0.81)",
wireframe: true,
wireframeLinewidth: 1

var cube = new THREE.Mesh(geometry, material);

camera.position.z = 5;

This step is optional if you want to rotate your object in 3D. It just looks cooler this way.
// Optional animation to rotate the element
var animate = function() {
cube.rotation.x += 0.01;
cube.rotation.y += 0.01;
renderer.render(scene, camera);


That’s all we need to do with Three.js. Now let’s play around with Handtrack.js
// Creating Canavs for video Input
const video = document.getElementById("myvideo");
const handimg = document.getElementById("handimage");
const canvas = document.getElementById("canvas");
const context = canvas.getContext("2d");
let trackButton = document.getElementById("trackbutton");
let updateNote = document.getElementById("updatenote");

let imgindex = 1;
let isVideo = false;
let model = null;

// Params to initialize Handtracking js
const modelParams = {
flipHorizontal: true,
maxNumBoxes: 1,
iouThreshold: 0.5,
scoreThreshold: 0.7

handTrack.load(modelParams).then(lmodel => {
model = lmodel;
updateNote.innerText = "Loaded Model!";
trackButton.disabled = false;

We are defining parameters here to load Handtrack js but this step is optional; you can pass an empty object also. The handTrack.load() method will help you load a model. Once the handtrack js is loaded, let’s write functions to load video stream in the canvas defined in the html. For that, we are using the handTrack.startVideo() method.
// Method to start a video
function startVideo() {
handTrack.startVideo(video).then(function(status) {
if (status) {
updateNote.innerText = "Video started. Now tracking";
isVideo = true;
} else {
updateNote.innerText = "Please enable video";

// Method to toggle a video
function toggleVideo() {
if (!isVideo) {
updateNote.innerText = "Starting video";
} else {
updateNote.innerText = "Stopping video";
isVideo = false;
updateNote.innerText = "Video stopped";

Now we can write the code to get prediction data from the handtrack.js
//Method to detect movement
function runDetection() {
model.detect(video).then(predictions => {
model.renderPredictions(predictions, canvas, context, video);
if (isVideo) {

The Real Trick 🧙🏼‍♂️

All the above code can basically be copy-pasted from the documentation of the library. But the real challenge was to integrate both to get the desired result.
The trick is to track the coordinates of the hand on video canvas and make changes with respect to it on to the Three js Object.
The prediction object from model.detect() method returns the following object:
"bbox": [x, y, width, height],
"class": "hand",
"score": 0.8380282521247864

bbox gives you the value coordinates, width and height of the box drawn around the hand. But the coordinates are not for the center point. To calculate it for the center point, we are using this simple formula:
let midvalX = value[0] + value[2] / 2;
let midvalY = value[1] + value[3] / 2;

Another problem is that the scale of the object’s canvas and tracker’s canvas is huge. Also, the center point origin of both the sources is not center. To take care of that, first we have to shift the coordinates​ so that the origin point of video canvas can be center.
Once that’s done, taking care of the scale issue is easy. So the final result will be something like this.
//Method to detect movement
function runDetection() {
model.detect(video).then(predictions => {
model.renderPredictions(predictions, canvas, context, video);
if (isVideo) {
if (predictions.length > 0) {

//Method to Change prediction data into useful information
function changeData(value) {
let midvalX = value[0] + value[2] / 2;
let midvalY = value[1] + value[3] / 2;

document.querySelector(".hand-1 #hand-x span").innerHTML = midvalX;
document.querySelector(".hand-1 #hand-y span").innerHTML = midvalY;

moveTheBox({ x: (midvalX – 300) / 600, y: (midvalY – 250) / 500 });

//Method to use prediction data to render cube accordingly
function moveTheBox(value) {
cube.position.x = ((window.innerWidth * value.x) / window.innerWidth) * 5;
cube.position.y = -((window.innerHeight * value.y) / window.innerHeight) * 5;
renderer.render(scene, camera);

Well, that’s it. You can now control the 3D object with your hand. I have made the code public on Github so go check it out. Clone it, run it, and have fun with it.


🖐🏼 Controlling 3D object with your hand

The Hand Trick
🖐🏼 Controlling 3D object with you hand

View on GitHub

Wrap Up 👋🏻

The story has just begun. This was the first tutorial in the series, and I have plans to take this experiment one step further. I’d love to have some contributors. If you’d like to contribute to the project, just generate a pull request at XenoX Multiverse and I’ll contact you.
Team XenoX started as a small team of devs working on open-source projects for the fun of it. But over the months, it has grown bigger and stronger. This is why I’ve created XenoX Multiverse, the home of all open-source initiatives by Team XenoX.If you want to be one of us, just write your name and start contributing!

Before I Go

We have a Telegram channel for now! Get the best of on the go, along with external articles, videos, and polls that we send daily!
👉🏼 Link

Time for me to go now. That’s all, folks! Remember that this is just a quick experiment to help you get your creative juices flowing. You can add more cool features to this, and if you feel like it’s running sluggishly, you can optimize the code later. The point is to learn something new as quickly as possible. Hope you liked this post.
See ya later!