Date post: | 22-Jul-2016 |
Category: |
Documents |
Upload: | rilwan-kujenya |
View: | 217 times |
Download: | 2 times |
HYDRAAR Visualizes the Information Emergencies Need on the Scene
AN AUGMENTED REALITY APP CAN AID EMERGENCY PROFESSIONALS
Vector graphic wireframe design.This vector wraps the user’s FOV.
Radial compass shows blips for tag locations.
Tick for compass bar
Navigational meter showing altitude and pitch
Blips for tag locations on the compass bar also showing distances
Tags visualized in user’s FOV
Walking Mode: the basic view displays simply a wireframe with compass on the top and tags and blips in FOV
Glance Mode: this interactive view will place numerous tags including smartphone and device locations in FOV. Many functions of this app are operated in this mode.
Full Map View: Allows the user to see locations on a map
MOCKUP FOR HYDRAARUI/UX WIREFRAME
BOSTON ARCHITECTURAL COLLEGE
RILWAN KUJENYA May 22, 20l5
Bachelor of Design Studies, conc. Digital Design and VisualizationFinal Review: May 14, 2015
Advised by Josh CastellanoINSTRUCTED BY CLAUDETTE L’HULLIER & JOSHUA WHITE
HYDRAAR: AN EXPERIMENTAL AUGMENTED REALITY APP CAN AID
EMERGENCY PROFESSIONALS
MODES
List of App Functions
HYDRAAR, is capable of assisting the user with coordinates and navigation, object detection and monitoring, device and smartphone connectivity and remote control via bluetooth and radio, and connectivity with other HYDRAAR users.
To see demo please follow this link:
A screenshot of my app is pictured here belowThis image shows the basic visual cue of the heads up display.
https://youtu.be/T-MHdIJyhVI
Bright Vector-base wireframe graphic language.
My objective was to design an app based on Wikitude as a platform. Then, to clean it up and make it far more realizable. Essentially, I want to Create a HUD UI design that is more ledgible. I will also how my app connects with the technologies that support it.
x
All of the systems exist today for my proposed app to work. If these highlighted sets of data are collected for the app, the app should have the functionality to aid emeemergency professionals.
Sensors detection particle content and radiation levels in the air will also be able to report it collected data to the app.
Monitoring of combustion gases and preemptive re conditons to de ne alert zones.
Vital signs monitoring from wearable device apps like HealthKit (Apple) for health monitoring.
Monitoring of buiding vibrations and material condi-tions with sensors installed in within buildings.
Control doors, gates, security cameras, and device power switches remotely.
Detect Smartphones (Windows, iOS, Android, Black-berry, and Linux-Based GUIs) via Bluetooth interface.
In the user’s Feild of View, FOV, outlines will wrap flame engulfed areas.
Temperature readings as collected from thermostats as well, as long as they operating.
HVAC sensors will share their readings.
Smart Appliances, like Samsung Zipel, will be directly powered off
BT4 power supplies are detected and remotely controlled.
Activity will be monitored and the phone, and all of its sensors, can be directly interacted with. This includes GPS and phonecalls.
Mobile phones will be detected.
App allows all of these functions
O2 leves are shared
O2 tank meter can be visualized in the FOV upon user’s command. Alerts will pop up in levels are becoming critically low.t
This will allow the user to make effective rescue decisions
Tile is an of a simple BT4tracking device, such a deviceoffers accurate positioning, and can also pick up gyroscopic, accelaration, and altitude sensors. The app will be able to determine vibrations and seismic conditions
Such a device maybe pre-installed in buildings, or dropped into building by working firefighters.
Imaging prisms
Wearable devices worn with gear
Feauture Image:This is what the app (or wikiworld) will look like when you find it
World Icon:.PNG files: large 512x512 and small 32x32
Wikitude Browser:I ran the app in real-time and this is what I got simple from uploading a kml file. I have much to expand on
RILWAN KUJENYA WEEK 2Pinup Progress
TESTING AUGMENTED
REALITY TAGS; SEEING
POTENTIAL
TAGS
Tags associate visually displayed information to points, or points of interests that are inputed into a map or augmented reality file (KML, or ARML)
In my app the crew members of the user’s emergency team are tagged. I know this is possible because I tested tagging doors in my home.
You can find my demo WikiWorld by searching for “AR EXIT”
This is how you make things appear in AR browsers. Interactions
will be operated by the users touch-free smart glove
Interactive information displays here
Interactive Menu: this a key visual display in this app, this will display most textual information the user needs.
MENU
Operating The Apps Fuctionalities
Feel free to access the video via QR scan with your smartphone.
Center crosshair
Concept
Project
BasisINTERNET OF THINGS EXISTING AR BROWSERS
Thesis
RILWAN KUJENYA BACHELORS DESIGN STUDIES DIGITAL DESIGN AND VISUALIZATIONBOSTON ARCHITECTURAL COLLEGE - 2015