Skip to content
master
Switch branches/tags
Code

Latest commit

 

Git stats

Files

Permalink
Failed to load latest commit information.
Type
Name
Latest commit message
Commit time
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

microllaborators ๐Ÿ‘ฉโ€๐Ÿ‘ฆโ€๐Ÿ‘ฆ๐Ÿ”ฌ๐Ÿ‘ฉโ€๐Ÿซ

put microscope images into AR for advanced collaboration ๐Ÿ”ฎ๐Ÿ”ฌ๐Ÿ‘ฉโ€๐Ÿซ

Teacher/student classroom scenario for gamified collaboration on three dimensional microscopic images. By opening the website on a device with a camera (iOS not supported ๐Ÿ˜“) pupils and the teacher can interact visually, from remote locations and in realtime.

safari fails

The idea is that the teacher can ask a question regarding the current image and the students can answer it by placing markers/annotations on the image. The teacher can see all annotations and validate the students response.

Note: This codebase was created as part of the Zeiss Hackathon 2017-04-29 in Munich. Don't blame us for the ugly code!

โœ๏ธ authors

๐Ÿ‘พ๐Ÿฆ„ usage

To use the augmented reality feature you need to get your hands on a markerimage of the zeiss marker. Just print it out or open it on your laptop. After you got a marker just head over to the website, see below. Currently there are two separate views implemented for student and teachers. In a production environment there would be authentication & authorization concepts in place that automatically route the user to the correct view.

๐Ÿ‘จโ€๐ŸŽ“ student

https://zeisshack.surge.sh

AR marker

๐Ÿ‘ฉโ€๐Ÿซ teacher

https://zeisshack.surge.sh/teacher

running the server

You can start a local development server with server-side autorestart and access it at https://localhost:8083

yarn
npm start

deploying the website

If you are Tim or Andreas you can deploy the code to surge at zeisshack.surge.sh

npm run deploy

If you're someone else just put the public folder on any static hosting. All the dynamic features are client only and realized over firebase.

๐Ÿ“ฝ demo / workflow

This section describes how the solution would be used in a real environment.

preparation

Teaching environment e.g. school with a teacher using a microscope that allows digital image capturing. Students use a smartphone with camera and internet/wifi.

The students open the student website and place the AR marker as paper on their desk (might be handed out by teacher).

teacher image selection

providing the question

The teacher goes to the teacher website and uploads/selects and image from his microscope he wants to discuss. He inputs the question in the top of the website and it gets broadcasted to all students.

teacher image selection

gamification

The students see the new image and an updated question. Each student can place a marker, if he misplaces it he can replace it by clicking the screen again. By moving their phone around they can look at the image from different angles with a 3d effect.

student places marker

student AR

evaluation of responses

While the students place their markers the teacher gets all answers in real time in his view and can select the correct region he expects to be placed. An overview shows him how many students answered correctly. This allows him to get feedback from the whole class without them influencing each other.

teacher results

โš–๏ธ License

MIT

๐Ÿ—‚ Used puzzle pieces / Attribution

  • AR.js: helped us a lot to get off the ground
  • firebase: allowed us to do realtime collaboration without backend
  • three.js: used for 3D markers
  • material design lite: fancy design
  • Zeiss: thanks for hosting and guiding us during the hackathon

About

microllaborators ๐Ÿ‘ฉโ€๐Ÿ‘ฆโ€๐Ÿ‘ฆ๐Ÿ”ฎ๐Ÿ”ฌ๐Ÿ‘ฉโ€๐Ÿซ - the revolution in teaching

Topics

Resources

License

Releases

No releases published

Packages

No packages published