An augmented reality app to locate services around you that are open
JavaScript CSS HTML
Switch branches/tags
Nothing to show
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Permalink
Failed to load latest commit information.
css
fonts
icons
img
js
less
markers
sass
style
.gitignore
Readme.md
awe-jsartoolkit-dependencies.js
awe.marker_ar.js
geomap.html

Readme.md

R U Open

![ru open logo](img/RU Open.png)

An augmented reality app to locate services around you that are open.

Description

Why R U Open

Our aim is to help a user find food and entertainment services in the immediate vicinity. The heart of the service is an Augmented Reality (AR) app which would use business and dining permits datasets (we have used Outdoor dining permits - Adelaide City Council found at data.sa.gov.au for proof-of-concept) to determine which business locations are open at the time the app is being used in the user's immediate vicinity.

The app would also allow relevant businesses to subscribe and provide information on menu and current campaigns (e.g. special offers) which can be displayed when the user taps on the marker designating the business location on the AR app.

Augmented reality based geolocation, routing and navigation:

We have developed the proof-of-concept of a browser-based Augmented Reality app to locate services which are open at the time of use (of the app) around the user. This kind of app would be normally developed using native APIs of the mobile phone OS (Android/iOS/Windows Mobile) and is extremely complex. The key concept is to use the mobile device sensors, mainly accelerometer, magnetometer, GPS, camera to establish the current location, orientation and point-of-view (in terms of camera viewport) of the device. On top, of the camera feed on the phone screen, data is superimposed in transparent/semi-transparent layers. This is done by creating virtual markers based on geolocated dataset(s) and using algorithms to determine the approximate distance and direction of heading for the marker in relation to the user's (and phone camera's) point-of-view. Navigation features can be added by using the phone's motion sensors to determine direction of movement as well.

Since our proof-of-concept has been developed within the time constraints of Unleashed 2016, Adelaide and Govhack, it has been developed using HTML5 geolocation (GPS), the orientation API (accelerometer) and WebGL (GPU). This PoC currently requires a standards compliant mobile browser (e.g. Firefox on Android). iOS does not currently support WebGL or WebRTC and has not implemented the DeviceOrientation API correctly. Please see this post to the W3C GeoLocation Working Group link for more detailed information. [Awe.js link] (https://github.com/buildar/awe.js/) has been used for building the app. A transparent overlay of Google Maps has been used to display roads and locate markers on top of the camera feed for simplicity.

Team name: Bob's gang

Team members: Matt Cejko, Martin Nobis, Muhammad Pavel, Mizanur Rahman, Asheshwor Shrestha, Oshim Somers

Source repository url: https://github.com/Unleashed2016/ruopen

Video url: https://youtu.be/al08lttiCS0

Hackerspace url: https://2016.hackerspace.govhack.org/content/r-u-open

Local Event Location:

Adelaide

Datasets Used:

APIs Used

  • Awe.js: Used to simplify using HTML5 geolocation, orientation and WebGL

  • Three.js: Used as a cool WebGL library

  • Google Maps JS API: Used to superimpose a layer of Roads and Streets with names on top of the camera feed.

  • Google Maps Directions API: Used for routing and navigation from user's current location to destination (location of restaurant, club, etc.)

  • Backendless: mBaaS service for storing business related data (e.g. for subscribing businesses to provide details of the business, special offers, campaigns, etc). Ideally a custom developed backend service would be used.