Skip to content
MixedRealityToolkit-Unity uses code from the base MixedRealityToolkit repository and makes it easier to consume in Unity.
Branch: mrtk_release
Clone or download
wiwei Merge pull request #4879 from wiwei/versionupdate
Update the version to RC2.1
Latest commit ab6bfc0 Jun 13, 2019

README.md

What is the Mixed Reality Toolkit

MRTK is a Microsoft driven open source project.

MRTK-Unity provides a set of foundational components and features to accelerate MR app development in Unity. The latest Release of MRTK (V2) supports HoloLens/HoloLens 2, Windows Mixed Reality, and OpenVR platforms.

  • Provides the basic building blocks for unity development on HoloLens, Windows Mixed Reality, and OpenVR.
  • Showcases UX best practices with UI controls that match Windows Mixed Reality and HoloLens Shell.
  • Enables rapid prototyping via in-editor simulation that allows you to see changes immediately.
  • Is extensible. Provides devs ability to swap out core components and extend the framework.
  • Supports a wide range of platforms, including
    • Microsoft HoloLens
    • Microsoft HoloLens 2
    • Windows Mixed Reality headsets
    • OpenVR headsets (HTC Vive / Oculus Rift)

Build Status

Branch CI Status Docs Status
mrtk_development CI Status Docs Status

Required Software

Windows SDK 18362+ Windows SDK 18362+ Unity Unity 2018.4.x Visual Studio 2017 Visual Studio 2017 Simulator (optional) Simulator (optional)
To build apps with MRTK v2, you need the Windows 10 May 2019 Update SDK.
To run apps for Windows Mixed Reality immersive headsets, you need the Windows 10 Fall Creators Update.
The Unity 3D engine provides support for building mixed reality projects in Windows 10 Visual Studio is used for code editing, deploying and building UWP app packages The Emulators allow you test your app without the device in a simulated environment

Feature Areas

Input System
 
Hand Tracking (HoloLens 2) Eye Tracking (HoloLens 2) Voice Commanding Gaze + Select (HoloLens) Teleportation
 
UI Controls
 
Solver and Interactions Controller Visualization Spatial Understanding Diagnostic Tool
 
MRTK Standard Shader

Getting Started with MRTK

Please check out the Getting Started Guide

More Documentation

Find this readme, other documentation articles and the MRTK api reference on our MRTK Dev Portal on github.io.

UI and Interaction Building blocks

Button Button Bounding Box Bounding Box Manipulation Handler Manipulation Handler
A button control which supports various input methods including HoloLens 2's articulated hand Standard UI for manipulating objects in 3D space Script for manipulating objects with one or two hands
Slate Slate System Keyboard System Keyboard Interactable Interactable
2D style plane which supports scrolling with articulated hand input Example script of using the system keyboard in Unity A script for making objects interactable with visual states and theme support
Solver Solver Object Collection Object Collection Tooltip Tooltip
Various object positioning behaviors such as tag-along, body-lock, constant view size and surface magnetism Script for lay out an array of objects in a three-dimensional shape Annotation UI with flexible anchor/pivot system which can be used for labeling motion controllers and object.
App Bar App Bar Pointers Pointers Fingertip Visualization Fingertip Visualization
UI for Bounding Box's manual activation Learn about various types of pointers Visual affordance on the fingertip which improves the confidence for the direct interaction
Slider Slider MRTK Standard Shader MRTK Standard Shader Hand Joint Chaser Hand Joint Chaser
Slider UI for adjusting values supporting direct hand tracking interaction MRTK's Standard shader supports various Fluent design elements with performance Demonstrates how to use Solver to attach objects to the hand joints
Eye Tracking: Target Selection Eye Tracking: Target Selection Eye Tracking: Navigation Eye Tracking: Navigation Eye Tracking: Heat Map Eye Tracking: Heat Map
Combine eyes, voice and hand input to quickly and effortlessly select holograms across your scene Learn how to auto scroll text or fluently zoom into focused content based on what you are looking at Examples for logging, loading and visualizing what users have been looking at in your app

Example Scenes

Explore MRTK's various types of interactions and UI controls in this example scene.

You can find other example scenes under Assets/MixedRealityToolkit.Examples/Demos folder.

Example Scene

Engage with the Community

  • Join the conversation around MRTK on Slack.

  • Ask questions about using MRTK on Stack Overflow using the MRTK tag.

  • Search for known issues or file a new issue if you find something broken in MRTK code.

  • Join our weekly community shiproom to hear directly from the feature team. (link coming soon)

  • Deep dive into project plan and learn how you can contribute to MRTK in our wiki.

  • For issues related to Windows Mixed Reality that aren't directly related to the MRTK, check out the Windows Mixed Reality Developer Forum.

This project has adopted the Microsoft Open Source Code of Conduct. For more information see the Code of Conduct FAQ or contact opencode@microsoft.com with any additional questions or comments.

Useful resources on the Mixed Reality Dev Center

Discover Discover Design Design Develop Develop Distribute) Distribute
Learn to build mixed reality experiences for HoloLens and immersive headsets (VR). Get design guides. Build user interface. Learn interactions and input. Get development guides. Learn the technology. Understand the science. Get your app ready for others and consider creating a 3D launcher.

Useful resouces on Azure

Spatial Anchors
Spatial Anchors
Speech Services Speech Services Vision Services Vision Services
Spatial Anchors is a cross-platform service that allows you to create Mixed Reality experiences using objects that persist their location across devices over time. Discover and integrate Azure powered speech capabilities like speech to text, speaker recognition or speech translation into your application. Identify and analyze your image or video content using Vision Services like computer vision, face detection, emotion recognition or video indexer.

Learn more about MRTK Project

You can find our planning material on our wiki under Project Management Section. You can always see the items the team is actively working on in the Iteration Plan issue.

How to Contribute

View the How To Contribute wiki page for the most up to date instructions on contributing to the Mixed Reality Toolkit!

For details on the different branches used in the Mixed Reality Toolkit repositories, check this Branch Guide here.

You can’t perform that action at this time.