LJ/www
2020-09-30 23:40:39 +02:00
..
build [fix] there should be a js file for auduno/clmtrackr 2020-09-27 18:52:29 +02:00
css newUI, python3,... 2020-09-19 14:28:56 +02:00
knobs newUI, python3,... 2020-09-19 14:28:56 +02:00
models Missing files 2020-09-21 11:06:30 +02:00
trckr conf template and bugfixs 2020-09-29 22:40:10 +02:00
align.html Scene change new UI 2020-09-30 23:40:39 +02:00
android-chrome-192x192.png newUI, python3,... 2020-09-19 14:28:56 +02:00
android-chrome-512x512.png newUI, python3,... 2020-09-19 14:28:56 +02:00
apple-touch-icon-80x80.png newUI, python3,... 2020-09-19 14:28:56 +02:00
apple-touch-icon.png newUI, python3,... 2020-09-19 14:28:56 +02:00
auralls.html conf template 2020-09-28 01:04:05 +02:00
favicon-16x16.png newUI, python3,... 2020-09-19 14:28:56 +02:00
favicon-32x32.png newUI, python3,... 2020-09-19 14:28:56 +02:00
favicon.ico Missing files 2020-09-21 11:06:30 +02:00
index.html conf template and bugfixs 2020-09-29 22:40:10 +02:00
launch.png newUI, python3,... 2020-09-19 14:28:56 +02:00
LJ.js UI updates 2020-09-27 18:52:29 +02:00
README.md conf template and bugfixs 2020-09-29 22:40:10 +02:00
selector.min.css Missing files 2020-09-21 11:06:30 +02:00
selector.min.js Missing files 2020-09-21 11:06:30 +02:00
simu.html conf template and bugfixs 2020-09-29 22:40:10 +02:00
touch-icon-ipad-retina.png newUI, python3,... 2020-09-19 14:28:56 +02:00
touch-icon-ipad.png newUI, python3,... 2020-09-19 14:28:56 +02:00
touch-icon-iphone-retina.png newUI, python3,... 2020-09-19 14:28:56 +02:00
webaudio-controls.js newUI, python3,... 2020-09-19 14:28:56 +02:00
webcomponents-lite.js newUI, python3,... 2020-09-19 14:28:56 +02:00
webcomponents-lite.js.map newUI, python3,... 2020-09-19 14:28:56 +02:00

= Web Interface

You can load index.html file from your browser or have a webserver of your choice pointing to this directory.

Webserver is mandatory if you want :

  • to remotely control LJ : imagine LJ can be installed in a dedicated computer/container with no easy access.
  • to use the face tracking, say from a smartphone. That's Lasercam (a clmtrackr plugin).

== Simu

A laser simulator. Choose lasernumber and it will display redis points for current scene/lasernumber

== clmtrackr

npm version

tracked face

clmtrackr is a javascript library for fitting facial models to faces in videos or images. It currently is an implementation of constrained local models fitted by regularized landmark mean-shift, as described in Jason M. Saragih's paper. clmtrackr tracks a face and outputs the coordinate positions of the face model as an array, following the numbering of the model below:

facemodel_numbering

Reference - Overview

The library provides some generic face models that were trained on the MUCT database and some additional self-annotated images. Check out clmtools for building your own models.

For tracking in video, it is recommended to use a browser with WebGL support, though the library should work on any modern browser.

For some more information about Constrained Local Models, take a look at Xiaoguang Yan's excellent tutorial, which was of great help in implementing this library.

Examples

Usage

Download the minified library clmtrackr.js, and include it in your webpage.

/* clmtrackr libraries */
<script src="js/clmtrackr.js"></script>

The following code initiates the clmtrackr with the default model (see the reference for some alternative models), and starts the tracker running on a video element.

<video id="inputVideo" width="400" height="300" autoplay loop>
  <source src="./media/somevideo.ogv" type="video/ogg"/>
</video>
<script type="text/javascript">
  var videoInput = document.getElementById('inputVideo');
  
  var ctracker = new clm.tracker();
  ctracker.init();
  ctracker.start(videoInput);
</script>

You can now get the positions of the tracked facial features as an array via getCurrentPosition():

<script type="text/javascript">
  function positionLoop() {
    requestAnimationFrame(positionLoop);
    var positions = ctracker.getCurrentPosition();
    // positions = [[x_0, y_0], [x_1,y_1], ... ]
    // do something with the positions ...
  }
  positionLoop();
</script>

You can also use the built in function draw() to draw the tracked facial model on a canvas :

<canvas id="drawCanvas" width="400" height="300"></canvas>
<script type="text/javascript">
  var canvasInput = document.getElementById('drawCanvas');
  var cc = canvasInput.getContext('2d');
  function drawLoop() {
    requestAnimationFrame(drawLoop);
    cc.clearRect(0, 0, canvasInput.width, canvasInput.height);
    ctracker.draw(canvasInput);
  }
  drawLoop();
</script>

See the complete example here.

Development

First, install node.js with npm.

In the root directory of clmtrackr, run npm install then run npm run build. This will create clmtrackr.js and clmtrackr.module.js in build folder.

To test the examples locally, you need to run a local server. One easy way to do this is to install http-server, a small node.js utility: npm install -g http-server. Then run http-server in the root of clmtrackr and go to https://localhost:8080/examples in your browser.

License

clmtrackr is distributed under the MIT License