No description, website, or topics provided.
C++ Other Processing Java C Python Other
Switch branches/tags
Nothing to show
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.

readme.md

ROKOKO

This is the software repository for the ROKOKO project. A lot of it is still in development, nothing is guaranteed to work and stuff will break all the time. Also, the current design has some pretty weird design choices (like connecting 18 individual sensors with USB cables in USB hubs) -- these are of course not final, but were made in order to make our prototyping process as fast as possible.

License

The license still has to be determined. For now, we reserve all rights to all original works in this repository (this will change once we've had that discussion). In the meantime you're welcome to download it, build it, tinker around with it for non-commercial use.

The ArUco library is released under the BSD license.

Introduction

ROKOKO aims to create a wireless motion capture system that tracks the motion of the actor, their position on the stage as well as their facial expression. This is going to be used for children's theatre by the company of the same name.

The motion capture suit prototype consists of a number of Arduino micros with an IMU attached and a Raspberry Pi receiving the data over USB-serial interface and forwarding it as OSC. The position of the actor is determined by the ArUco augmented reality library. The stage is looking rather trippy:

ArUco pattern

... but that is because the ArUco library is capable of determining the position of a camera filming such a pattern.

Motion capture

Firmware

The Arduinos run the code in ./Arduino/compactImuClient. We're building the code with an Arduino distribution that makes use of GCC version 4.8.1 in order to save space. At the time of writing, the main Arduino distribution produces a binary that is too large for the Arduino Micros (it's called arduino-avr-toolchain-nightly-gcc-4.8.1-linux64.tgz).

The Raspberry Pi is running an Archlinux distro. The c program in ./imu-udp-bridge-c runs on startup and connects to the sensors.

Recipient on the actor

The Arduinos are connected to the recipient (currently a Raspberry Pi) using USB and an awful lot of hubs. The program in ./imu-udp-bridge-c/ polls the Arduinos for data and relays it over OSC to the IP specified in recipient.txt on UDP port 14040.

The data is received by our Unity-based renderer and applied to the relevant bones.

Calibration

The IMUs need to be calibrated to work properly.

To start the accelerometer calibration routine send an 'a' on the serial port. Then carefully rotate the sensor around itself to max out the x, y and z axes in 1g.

To calibrate the magnetometer send an 'm' on the serial port and repeat as above but be extra carefull that you are not close to electronic or magnetic disturbances in the room. Step away from computers, refrigerators, radiator, any big metal objects etc.

To set the unique identifier of a sensor send a serial 'I' followed by an ID of sixteen (16) characters, no more, no less. Pad with spaces.

Position on the stage

The position on the stage is determined by the ArUco software. ArUco is an academic project for augmented reality applications, that is superimposing 3D objects onto the real world, as seen through the 'lens' of one's mobile phone. We use it to get the coordinates of the camera, calculated from the perspective of the markers on the stage floor. The coordinates are given as a pair of vectors called T and R, respectively. From these we can determine the Translation and Rotation of the camera.

The application in ./aruco/utils/rokoko_position_streamer.cpp implements this using the ArUco library, and streams the vector pair over OSC (on UDP port 14040, as the sensors).

Facial expression

We also need to keep track of the eyes and facial expression of the actor, in order to have the 3D model reflect that, too. There is still a lot of research being made in that regard, and currently we're looking at two different models:

  • FaceTracker is a markerless face tracking implementation. The license disallows commercial use of the software, but is unclear whether the ROKOKO project constitutes commercial use when no money is earned on selling mocap suits and we're sharing all the other code.
  • eyeLike is a library that tracks eye movements in a webcam image. There is a fork of that project here that, besides eyes, tracks green markers in the face and streams the coordinates of those over OSC.

Other tools, legacy / WIP - not complete

  • ./legacy/imuOscVisualizer is a somewhat incomplete openFrameworks tool for visualizing the OSC data. We primarily use Unity for the visualization now.
  • ./legacy/imuTestVisualizer is an openframeWorks skeleton animation test connecting directly to the sensors.
  • ./legacy/camcalib is for making opencv instrinsics for the camera