PapARt is a software development kit (SDK) that enables the creation of interactive projection mapping. Today, it is developed by the Inria Potioc team. It comes from the augmented physical drawing tools created by Jeremy Laviole, which are documented in his PhD thesis (free to read).
PapARt 1.4 - Release Candidate. (July 2018)
This new release brings many new features. It is now easier to place PaperScreen on the table with the new TableScreen class.
The color tracking and particularly the circular tracking is quite robust and enable the creation of physical interfaces with a high detection rate. There will be a complete tutorial on how to create a mixed reality interface with touch and circle tracking.
We work to improve the current API, as it will be part of the coming Nectar platform. The main motivation for Nectar to push further the possibilites of SAR with PapARt. The rendering will not be limited to Processing for rendering with the Unity3D Nectar plugin. The plugin is in internal test/development phase, and is already quite promising.
The 1.4 version and development versions are hosted on gitlab. You can request access if you collaborate with RealityTech, or use RealityTech Hardware platforms.
The 1.3 version, sister of 1.4 will be free and publicly available on github.
Version 1.1 and 1.2 (January 2018)
The first 2018 releases are 1.1 and 1.2.There are two major updates:
- (1.1) ColorTracking: The library enables color tracking. The system learns to recognize and track five colors which can be used to activate buttons, or identify objects.
- (1.2) Hand recognition and tracking is improved to segment the arm - hands and fingers. The API is in progress and will evolve. This version is distributed with RealityTech's hardware.
- Easier to compile thanks to the release of JavaCV/JavaCPP 1.4.
- Support of intel Realsense cameras (SR300 and F200).
- Support of Orbbec cameras (Astra S).
- JRubyArt support is getting extended.
- Community and commercial support is moved from the wiki to the forum.
Version 1.0 (summer 2017)
The first big release is ready. If you want to try it out download our precompiled version from the example repository.
This repository is for the development of the library.
You may want to go to the PapARt-Examples repository to see how to use it or discover the features and demos.
It enables the creation of Augmented Reality (AR) applications in Processing. Like most AR toolkit, it is vision based and it detects the world using color cameras. In addition to this, PapARt also uses a depth camera.
We generally use pre-calibrated (intrinsics parameters) and PapARt enables the extrinsic calibration: how cameras are located relatively from one to another. It also provides simple and unprecise tools to create intrinsic calibration.
It uses tracking libraries such as ARToolkit and OpenCV, and can be extended. The strength of this library is the creation of interactive projection (also called spatial augmented reality in research). In addition to cameras, PapARt calibrates the projector’s extrinsics to create projector/camera systems also called ProCams.
Interactivity is increased thanks to an object and hand tracking enabled by the depth camera.
More information about the research project here:
PapARt is large library, and work with many different systems:
- webcams and professionnal cameras (PointGrey cameras).
- depthCameras: Kinect Xbox360, Kinect xbox one, Intel Realsense.
- Projector/camera/depth camera systems (the main purpose of the library).
How to contribute
The open source release is new (end of August 2016), feel free to fork, star, and file issues for this sources. You can contribute your examples to the example repository to share your creations with the community.
The distribution got better, and the next steps would be to create versions on Android and/or on Raspberry PI.
PapARt is an open source software owned by Inria, Bordeaux University, CNRS, and RealityTech distributed under the LGPL license.