+ All Categories
Home > Engineering > SIXTH SENSE TECHNOLOGY REPORT

SIXTH SENSE TECHNOLOGY REPORT

Date post: 18-May-2015
Category:
Upload: jismi-jacob
View: 1,723 times
Download: 9 times
Share this document with a friend
Popular Tags:
43
SIXTH SENSE TECHNOLOGY Seminar report submitted in partial fulfillment of the requirements for the award of the degree of Bachelor of Technology (B. Tech) In Information Technology Of University of Calicut By JISMI K JACOB Under the Guidance of Ms. JISNA V A Asst.Professor, IT Department
Transcript
Page 1: SIXTH SENSE TECHNOLOGY REPORT

SIXTH SENSE TECHNOLOGY

Seminar report submitted in partial fulfillment of the

requirements for the award of the degree of

Bachelor of Technology (B. Tech)

In

Information Technology

Of

University of Calicut

By

JISMI K JACOB

Under the Guidance of

Ms. JISNA V A

Asst.Professor, IT Department

April 2014

Department of Information Technology

JYOTHI ENGINEERING COLLEGE,

Cheruthuruthy,Thrissur – 679 531.

Page 2: SIXTH SENSE TECHNOLOGY REPORT

DEPARTMENT OF INFORMATION TECHNOLOGY

JYOTHI ENGINEERING COLLEGE

CHERUTHURUTHY, THRISSUR- 679 531

April 2014

CERTIFICATE

This is to certify that the seminar report entitled “SIXTH SENSE

TECHNOLOGY” being submitted by Ms. JISMI K JACOB in partial

fulfillment of the requirements for the award of degree of Bachelor of

Technology of University of Calicut is bonafide record of work carried out at

Department of Information Technology, JECC by her during the period

December 2013 – April 2014 under our supervision.

Seminar Guide Seminar Coordinator Ms. DIVYA M MENON

Ms. JISNA V A Ms. SABNA A B H.O.D in charge,

Asst.Professor Asst.Professor IT Department

Page 3: SIXTH SENSE TECHNOLOGY REPORT

ABSTRACT

Sixth Sense technology is a technology with which a system could be trained to

recognize and percept real world objects and react as desired. Sixth Sense technology bridges

the gap between the physical world and the digital world, bringing intangible, digital

information out into the tangible world, and allowing us to interact with this information via

natural hand gestures. Sixth Sense Technology is implemented in 'Sixth Sense/WUW (wear

your world) using gesture recognition, augmented reality, computer vision and radio

frequency identification. It’s just born concept which allows user to connect with the internet

seamlessly. Without use of keyboard, mouse we can see videos access, change, move data

simply .But this concept bottle necks lead to modification of the same by using commands

instead of gestures. Sixth Sense technology could be integrated with voice recognition.

Bluetooth device and laser projectors could be used.

Page 4: SIXTH SENSE TECHNOLOGY REPORT

INDEX

Contents Page no:

1. INTRODUCTION 1

2. OVERVIEW 3

3. ORIGIN OF IDEA 4

3.1 What is Sixth Sense Technology? 5

3.2 Earlier Sixth Sense Prototype 5

3.3 Recent Prototype 6

4. DETAILED DESCRIPTION 8

4.1 Components 8

4.2 Working of Sixth Sense Technology 11

4.3 Related Technologies 13

4.4 Applications 19

4.5 Advantages 23

4.6 Future Enhancements 23

5. CONCLUSION 24

6. BIBLIOGRAPHY 25

Page 5: SIXTH SENSE TECHNOLOGY REPORT

List of Figures

3.1 Six Senses 5

3.2 Earlier Prototype 6

3.3 Recent Prototypes 7

4.1.1 Camera 8

4.1.2 Projector 9

4.1.3 Mirror 9

4.1.4 Smartphone 10

4.1.5 Color Marker 11

4.1.6 Microphone 11

4.2 Working of the Sixth Sense 12

4.3.1 Augmented reality 15

4.3.2 Computer vision 15

4.3.3 Hand gestures 17

4.3.4 Components of RFID 18

Page 6: SIXTH SENSE TECHNOLOGY REPORT

CHAPTER 1

INTRODUCTION

We use our five natural senses to perceive any information; that information helps us

make decisions and choose the right actions to take. But arguably the most useful information

that can help us make the right decision is not naturally perceivable with our five senses,

namely the data, information and knowledge that mankind has accumulated about everything

and which is increasingly all available online. Although the miniaturization of computing

devices allows us to carry computers in our pockets, keeping us continually connected to the

digital world, there is no link between our digital devices and our interactions with the

physical world. Information is confined traditionally on paper or digitally on a screen. Sixth

Sense Technology bridges this gap, bringing intangible, digital information out into the

tangible world, and allowing us to interact with this information via natural hand gestures.

‘Sixth Sense’ frees information from its confines by seamlessly integrating it with reality, and

thus making the entire world your computer. WUW was developed by Pranav Mistry, a Ph. D

student at Fluid Interfaces Group at the MIT Media Lab. The Sixth Sense prototype

implements several applications that demonstrate the usefulness, viability and flexibility of

the system acts as the computer and your connection to the Cloud, all the information stored

on the web.

Sixth Sense recognizes the objects around you, displaying information automatically

and letting you access it in any way you want, in the simplest way possible. The device

brings us closer to reality and assists us in making right decisions by providing the relevant

information, thereby, making the entire world a computer. The technology is mainly based on

hand gesture recognition, image capturing, processing, and manipulation, etc. The software of

the technology uses the video stream, which is captured by the camera, and also tracks the

location of the tips of the fingers to recognize the gestures. This process is done using some

techniques of computer vision. He invented ‘ Sixth Sense / WUW ( Wear UR World)’ which

is a wearable gestural , user friendly interface which links the physical world around us with

digital information and uses hand gestures to interact with them. This technology is a

revolutionary way to interface the physical world with digital information. Modern

technologies include the touch screen techniques which is used widely and it makes ease of

operation and saves utilization time.

Page 7: SIXTH SENSE TECHNOLOGY REPORT

This paper deals with the latest technology called the sixth sense. It’s a wearable

interface that augments the physical world around us with the digital information. It’s just

born concept which allows user to connect with the internet seamlessly. Without use of

keyboard, mouse we can see videos access, change, move data simply .But this concept bottle

necks lead to modification of the same by using commands instead of gestures. Speech IC is

used as a database for commands which will be initially trained for storage. It performs the

corresponding commands accessing the operation from the mobile device connected to it and

action is projected using a projector over any surface.

CHAPTER 2

Page 8: SIXTH SENSE TECHNOLOGY REPORT

OVERVIEW

Previously many technologies evolved such as augmented reality which is to add

information and meaning to real object or place. Unlike virtual reality, augmented reality

does not create a simulation of reality instead it takes a real object or space as the foundation

and incorporates technologies that add contextual data to deepen a person understanding of

the subject. It’s a term for live direct or indirect view of a physical real world environment

whose elements are augmented by virtual computer generated imagery. Gesture recognition is

a term with a goal of interpreting human gestures through mathematical gestures and

mathematical algorithms. Computer vision is the science and technology of machines that is

concerned with the theory behind artificial systems that extract information from the images.

As a technological discipline, computer vision seeks to apply its theories and models

to the construction of computer vision systems. The examples include the controlling

processes, detecting events, organising information, modelling objects or environments and

interaction. Recently speech integrated circuits evolved which is used widely in car

automation and home appliances. It eases the operation and saves the utilization time of the

manual operations performed by the human’s every day. The speech recognition process is

performed by a software component known as speech recognition engine. The primary

function of this is to process the spoken input and translate it into text which the application

understands. The application then can do one of the two things, 1.The application can

interpret the result of the recognition as a command, in this case application is a command

and control application. If the application handles the recognized text as simply text, then it’s

considered as dictation application. When the user says something, it is known as utterance.

An utterance is a stream of speech between two periods of silence. The speech IC can be used

for all sorts of data, statistical models, and algorithms to convert spoken input into text.

CHAPTER 3

Page 9: SIXTH SENSE TECHNOLOGY REPORT

ORIGIN OF IDEA

This technology is a revolutionary way to interface the physical world with digital

information. Modern technologies include the touch screen techniques which is used widely

and it makes ease of operation and saves utilization time. Sixth sense is a wearable gestural

interface that augments the physical world around us with digital information and lets us use

natural hand gestures to interact with that information. But the bottle necks of this method

such as position of camera, for capturing gestures interprets the accuracy in the projected

output, lead to use of commands instead of hand gestures. The position of camera is a major

constraint in the image capturing and projected output efficiency and accuracy. Therefore the

actions which we regularly perform in our daily life, are converted to commands and are

trained to a speech IC .They are stored as a database in the integrated circuit and

corresponding actions are performed when the speech is recognized from the user. It’s a hi-

tech device seamlessly integrate Analog information with our every day physical world. The

voice is directly performed into operation within fractions of seconds, and the action is

projected on the surface. It’s a portable device and eases the operation which we regularly

perform. Basically the sixth sense technology concept involves the use of hand gestures .the

finger tip will contain colored markers and hence gestures performed will be captured by the

camera. Then it’s given to the mobile device for the corresponding action to be performed.

The action is projected on the surface through the projector. Software algorithms and

computer vision technologies will be used to enable the action from the mobile device for the

corresponding gesture captured in the camera. This gesture based technology is used for

variety of applications like performing basic actions, locating points in the map, watching

video in news paper, dialing number in hand etc. The slight modification of this method leads

to the use of commands that is analog information into real world. The analog data is

converted into digital and performed as action, as all times the hand gestures cannot be used.

This was how the wearable device is fit to the human body. Here color markers are used in

the finger tips.

In our technology we use commands for performing the same operations. Many high

technology speech integrated circuits evolved which makes our operation enhanced with

more advanced features. To ensure accurate gesture recognition and an intuitive interface a

number of constraints are applied. A region in the front of the projection screen is defined as

Page 10: SIXTH SENSE TECHNOLOGY REPORT

the active zone and the gestures are ignored, if the gestures are performed out of this area.

Gestures are also defined by a set start posture, end posture and dynamic motion between the

start and end postures. Perhaps the use of gestures is most powerful when combined with

other input modalities, especially voice. Allowing combined voice and gestural input has

several tangible advantages. The first is purely practical-ease of expression .Ease corresponds

to the efficiency with which commands can be remembered and expressiveness, size of

command vocabulary.

3.1 WHAT IS SIXTH SENSE?

Sixth sense in scientific or non scientific terms defined as Extra Sensory Perception. It

involves the reception of information not gained through any of the five senses. Namely the

data, information and knowledge that mankind has accumulated about everything available

online.

Figure 3.1 six senses

3.2 EARLIER SIXTH SENSE PROTOTYPE

Steve Mann is the father of Sixth Sense Technology who made of wearable

computer in 1990. He implemented the neck worn projector with a camera system. Maes’

MIT group, which includes seven graduate students, were thinking about how a person could

be more integrated into the world around them and access information without having to do

something like take out a phone. They initially produced a wristband that would read an

Radio Frequency Identification tag to know, for example, which book a user is holding in a

store. They also had a ring that used infrared to communicate by beacon to supermarket smart

shelves to give you information about products. As we grab a package of macaroni, the ring

would glow red or green to tell us if the product was organic or free of peanut traces —

whatever criteria we program into the system. They wanted to make information more useful

to people in real time with minimal effort in a way that doesn’t require any behavior changes.

The wristband was getting close, but we still had to take out our cell phone to look at the

Page 11: SIXTH SENSE TECHNOLOGY REPORT

information. That’s when they struck on the idea of accessing information from the internet

and projecting it. So someone wearing the wristband could pick up a paperback in the

bookstore and immediately call up reviews about the book, projecting them onto a surface in

the store or doing a keyword search through the book by accessing digitized pages on

Amazon or Google books. They started with a larger projector that was mounted on a helmet.

But that proved cumbersome if someone was projecting data onto a wall then turned to speak

to friend — the data would project on the friend’s face.

Figure 3.2: Earlier prototype

3.3 RECENT PROTOTYPE

WUW was developed by Pranav Mistry, a Ph. D student at Fluid Interfaces Group at

the MIT Media Lab. The Sixth Sense prototype implements several applications that

demonstrate the usefulness, viability and flexibility of the system acts as the computer and

your connection to the Cloud, all the information stored on the web. The key here is that

Sixth Sense recognizes the objects around you, displaying information automatically and

letting you access it in any way you want, in the simplest way possible. The device brings us

closer to reality and assists us in making right decisions by providing the relevant

information, thereby, making the entire world a computer. The technology is mainly based on

hand gesture recognition, image capturing, processing, and manipulation, etc. The software of

the technology uses the video stream, which is captured by the camera, and also tracks the

location of the tips of the fingers to recognize the gestures. This process is done using some

techniques of computer vision. He invented ‘ Sixth Sense / WUW ( Wear UR World)’ which

is a wearable gestural , user friendly interface which links the physical world around us with

digital information and uses hand gestures to interact with them.

Page 12: SIXTH SENSE TECHNOLOGY REPORT

Figure 3.3 Recent prototypes

CHAPTER 4

DETAILED DESCRIPTION

Page 13: SIXTH SENSE TECHNOLOGY REPORT

4.1 COMPONENTS

The hardware components are coupled in a pendant like mobile wearable device. The

components are:

1. Camera

2. Projector

3. Mirror

4. Mobile component

5. Color markers

6. Microphone

4.1.1 Camera

Camera captures an object in view and tracks the user’s hand gestures. It sends the

data to smart phone. Camera recognizes and tracks user's hand gestures and physical objects

using computer-vision based techniques. Sixth Sense system implements a gestural camera

that takes photos of the scene the user is looking at by detecting the ‘framing’ gesture. It acts

as a digital eye, connecting you to the world of digital information.

Figure 4.1.1: Camera

4.1.2 Projector

A tiny LED projector displays data sent from the smart phone on any surface in view-object,

wall, or person. The projector projects visual information enabling surfaces, walls and

physical objects around us to be used as interfaces. The

projector projects visual information enabling surfaces,

walls and physical objects around us to be used as

interfaces. We want this thing to merge with the physical world in

a real physical sense. You are touching that object and

projecting info onto that object. The information will look like it

is part of the object. The projector itself contains a battery inside, with 3 hours of battery life.

Page 14: SIXTH SENSE TECHNOLOGY REPORT

Figure 4.1.2: Projector

4.1.3 Mirror

The mirror reflects the projection coming out from the projector and thus helps in

projecting onto the desired locations on walls or surfaces. The user manually can change the

tilt of the mirror to change the location of the projection. For example in application where

the user wants the projection to go on the ground instead of the surface in front, he can

change the tilt of the mirror to change the projection. Thus, the mirror in the Sixth Sense

helps in overcoming the limitation of the limited projection space of the projector.

Figure 4.1.3 Mirror

4.1.4 Mobile component

The Sixth Sense system uses a mobile computing device in user’s pocket as the

processing device. The software program enabling all the features of the system runs on this

computing device. This device can be a mobile phone or a small laptop computer. The

camera, the projector and the microphone are connected to this device using wired or wireless

connection. The detail of the software program that runs on

this device is provided in next section. The mobile computing device is also connected to the

Internet via 3G network or wireless connection.

Page 15: SIXTH SENSE TECHNOLOGY REPORT

A Web-enabled smart phone in the user's pocket processes the video data, using

vision algorithms to identify the object. Other software searches the Web and interprets the

hand gestures.

Figure 4.1.4: Smartphone

4.1.5 Color marker

Color marker is at the tip of the user’s fingers. Marking the user’s fingers with red, yellow,

green, and blue tape helps the webcam recognize gestures. The camera tracks the movements

of the color markers. The movements and arrangements of these makers are interpreted into

gestures that act as interaction instructions for the projected application interfaces.

The software program processes the video stream data captured by the camera and

tracks the locations of the colored markers (visual tracking fiducials) at the tip of the user’s

fingers using simple computer-vision techniques. The movements and arrangements of these

fiducials are interpreted into gestures that act as interaction instructions for the projected

application interfaces. The maximum number of tracked fingers is only constrained by the

number of unique fiducials, thus Sixth Sense also supports multi-touch and multi-user

interaction.

Page 16: SIXTH SENSE TECHNOLOGY REPORT

Figure 4.1.5: Color marker

4.1.6 Microphone

The microphone is an optional component of the Sixth Sense. It is required when

using a paper as a computing interface. When the user wants to use a sheet of paper as an

interactive surface, he or she clips the microphone to the paper. The microphone attached this

way captures the sound signals of user’s touching the paper. This data is passed to computing

device for processing. Later, combined with the tracking information about user’s finger, the

system is able to identify precise touch events on the paper. Here, the sound signal captured

by the microphone provides time information whereas the camera performs tracking. The

applications enabled by this technique are explained earlier.

Figure 4.1.6 Microphone

4.2 WORKING OF SIXTH SENSE TECHNOLOGY

The hardware that makes Sixth Sense work is a pendant like mobile wearable

interface. It has a camera, a mirror and a projector and is connected wirelessly to a Bluetooth

smart phone that can slip comfortably into one’s pocket. The camera recognizes individuals,

images, pictures, gestures one makes with their hands. Information is sent to the Smartphone

for processing. The downward-facing projector projects the output image on to the mirror.

Mirror reflects image on to the desired surface. Thus, digital information is freed from its

confines and placed in the physical world.

Page 17: SIXTH SENSE TECHNOLOGY REPORT

Figure 4.2: working of the sixth sense

The entire hardware apparatus is encompassed in a pendant-shaped mobile wearable

device. Basically the camera recognizes individuals, images, pictures, gestures one makes

with their hands and the projector assists in projecting any information on whatever type of

surface is present in front of the person. The usage of the mirror is significant as the projector

dangles pointing downwards from the neck. To bring out variations on a much higher plane,

in the demo video which was broadcasted to showcase the prototype to the world, Mistry uses

colored caps on his fingers so that it becomes simpler for the software to differentiate

between the fingers, demanding various applications.

The software program analyses the video data caught by the camera and also tracks

down the locations of the colored markers by utilizing single computer vision techniques.

One can have any number of hand gestures and movements as long as they are all reasonably

identified and differentiated for the system to interpret it, preferably through unique and

varied fiducials. This is possible only because the ‘Sixth Sense’ device supports multi-touch

and multi-user interaction.

The technology is mainly based on hand gesture recognition, image capturing,

processing, and manipulation, etc. The map application lets the user navigate a map displayed

on a nearby surface using hand gestures, similar to gestures supported by multi-touch based

Page 18: SIXTH SENSE TECHNOLOGY REPORT

systems, letting the user zoom in, zoom out or pan using intuitive hand movements. The

drawing application lets the user draw on any surface by tracking the fingertip movements of

the user’s index finger.

4.3 RELATED TECHNOLOGIES

Previously many technologies evolved such as augmented reality which is to add

information and meaning to real object or place. Unlike virtual reality, augmented reality

does not create a simulation of reality instead it takes a real object or space as the foundation

and incorporates technologies that add contextual data to deepen a person understanding of

the subject. It’s a term for live direct or indirect view of a physical real world environment

whose elements are augmented by virtual computer generated imagery. Sixth Sense

technology takes a different approach to computing and tries to make the digital aspect of our

lives more intuitive, interactive and, above all, more natural. We shouldn’t have to think

about it separately. It’s a lot of complex technology squeezed into a simple portable device.

When we bring in connectivity, we can get instant, relevant visual information projected on

any object we pick up or interact with The technology is mainly based on hand augmented

reality, gesture recognition, computer vision based algorithm etc.

4.3.1 AUGMENTED REALITY

Augmented reality is a term for a live direct or indirect view of a physical real world

environment whose elements are augmented by virtual computer generated imagery.

Augmented reality, blurs the line between what's real and what's computer-generated by

enhancing what we see, hear, feel and smell. Augmented reality is one of the newest

innovations in the electronics industry. It superimposes graphics, audio and other sense

enhancements from computer screens onto real time environments. Augmented reality goes

far beyond the static graphics technology of television where the graphics imposed do not

change with the perspective. Augmented reality systems superimpose graphics for every

perspective and adjust to every movement of the user's head and eyes. The basic idea of

augmented reality is to superimpose graphics, audio and other sensory enhancements over a

real world environment in real time. However, augmented reality is more advanced than any

technology you've seen in television broadcasts, although some new TV effects come close,

Page 19: SIXTH SENSE TECHNOLOGY REPORT

such as RACEf/x and the super-imposed first down line on televised U.S. football games,

both created by Sport vision. But these systems display graphics for only one point of view.

You can even point the phone at a building, and Layar will tell you if any companies in that

building are hiring, or it might be able to find photos of the building on Flickr. Augmented

reality adds graphics, sounds, haptic feedback and smell to the natural world as it exists.

Everyone from tourists, to soldiers, to someone looking for the closest subway stop can now

benefit from the ability to place computer-generated graphics in their field of vision.

The main hardware components for augmented reality are: display, tracking, input

devices, and computer. Combination of powerful CPU, camera, accelerometers, GPS and

solid state compass are often present in modern Smartphone, which make them prospective

platforms. There are three major display techniques for Augmented Reality:

1. Head Mounted Displays

2. Handheld Displays

3. Spatial Displays

1. Head Mounted Displays

A Head Mounted Display (HMD) places images of both the physical world and

registered virtual graphical objects over the user's view of the world. The HMD's are either

optical see through or video see-through in nature.

2. Handheld Displays

Handheld Augment Reality employs a small computing device with a display that fits in a

user's hand. All handheld AR solutions to date have employed video see through techniques

to overlay the graphical information to the physical world. Initially handheld AR employed

sensors such as digital compasses and GPS units for its six degree of freedom tracking

sensors.

3. Spatial Displays

Instead of the user wearing or carrying the display such as with head mounted displays or

handheld devices; Spatial Augmented Reality (SAR) makes use of digital projectors to

display graphical information onto physical objects.

Page 20: SIXTH SENSE TECHNOLOGY REPORT

Figure 4.3.1 Augmented reality

4.3.2 COMPUTER VISION

Computer vision is the science and technology of machines that see. It is concerned

with the theory behind artificial systems that extract information from images. An image is a

huge array of gray level (brightness) values of individual pixels. Taken individually, these

numbers are almost meaningless, because they contain very little information about the scene

. A robot needs information like "object ahead", "table to the left", or "person approaching" to

perform its tasks. The conversion of this huge amount of low level information into usable

high level information is the subject of computer vision. Earlier algorithms were too

computationally expensive to run in real-time, but also required any type of memory and

modeling. We concentrate on two types of images frequently used in computer vision:

Intensity images Photograph like images encoding light intensities, Range images encoding

shape and distance (sonar and laser).

Figure 4.3.2: Computer vision

Page 21: SIXTH SENSE TECHNOLOGY REPORT

4.3.3 GESTURE RECOGNTION

Gesture recognition is a topic in computer science and language technology with the

goal of interpreting human gestures via Mathematical algorithms. Gestures can originate from

any bodily motion or state but commonly originate from the face or hand. Current focuses in

the field include emotion recognition from the face and hand gesture recognition. The

keyboard and mouse are currently the main interfaces between man and computer. Humans

communicate mainly by vision and sound, therefore, a man-machine interface would be more

intuitive if it made greater use of vision and audio recognition. Another advantage is that the

user not only can communicate from a distance, but need have no physical contact with the

computer. However, unlike audio commands, a visual system would be preferable in noisy

environments or in situations where sound would cause a disturbance. Many approaches have

been made using cameras and computer vision algorithms to interpret sign language.

However, the identification and recognition of posture, gait, proxemics, and human behaviors

is also the subject of gesture recognition techniques. Gesture recognition can be seen as a way

for computers to begin to understand human body language, thus building a richer bridge

between machines and humans than primitive text user interfaces or even GUIs (graphical

user interfaces), which still limit the majority of input to keyboard and mouse. Gesture

recognition enables humans to interface with the machine (HMI) and interact naturally

without any mechanical devices. Using the concept of gesture recognition, it is possible to

point a finger at the computer screen so that the cursor will move accordingly. Gesture

recognition is useful for processing information from humans which is not conveyed through

speech or type. As well, there are various types of gestures which can be identified by

computers.

Figure 4.3.3: Hand gestures

Page 22: SIXTH SENSE TECHNOLOGY REPORT

4.3.4 RADIO FREQUENCY IDENTIFICATION

Radio Frequency Identification is basically an electronic tagging technology that

allows the detection, tracking of tags and consequently the objects that they are affixed to

Radio frequency identification (RFID) is a technology that uses communication via radio

waves to exchange data between a reader and an electronic tag attached to an object, for the

purpose of identification and tracking. Some tags can be read from several meters away and

beyond the line of sight of the reader. The application of bulk reading enables an almost

parallel reading of tags. Radio frequency identification (RFID) is a generic term that is used

to describe a system that transmits the identity (in the form of a unique serial number) of an

object or person wirelessly, using radio waves. It's grouped under the broad category of

automatic identification technologies. Radio-frequency identification involves interrogators

(also known as readers), and tags (also known as labels). Most RFID tags contain at least two

parts. One is an integrated circuit for storing and processing information, modulating and

demodulating a radio-frequency (RF) signal, and other specialized functions. The other is an

antenna for receiving and transmitting the signal. RFID is in use all around us. If you have

ever chipped your pet with an ID tag, used EZ Pass through a toll booth, or paid for gas using

Speed Pass, you've used RFID. In addition, RFID is increasingly used with biometric

technologies for security.

The idea of Sixth Sense is to use Radio Frequency Identification technology in

conjunction with a bunch of other enterprise systems such as the calendar system or online

presence that can track user activity. Here, we consider an enterprise setting of the future

where people (or rather their employee badges) and their personal objects such as books,

laptops, and mobile phones are tagged with cheap, passive RFID tags, and there is good

coverage of RFID readers in the workplace. Sixth Sense incorporates algorithms that start

with a mass of undifferentiated tags and automatically infer a range of information based on

an accumulation of observations. The technology is able to automatically differentiate

between people tags and object tags, learn the identities of people, infer the ownership of

objects by people, learn the nature of different zones in a workspace (e.g., private office

versus conference room), and perform other such inferences. By combining information from

these diverse sources, Sixth Sense records all tag-level events in a raw database. The

inference algorithms consume these raw events to infer events at the level of people, objects,

Page 23: SIXTH SENSE TECHNOLOGY REPORT

and workspace zones, which are then recorded in a separate processed database. Applications

can either poll these databases by running SQL queries or set up triggers to be notified of

specific events of interest. Sixth Sense infers when a user has interacted with an object, for

example, when you pick up your mobile phone. It is a platform in that its programming

model makes the inferences made automatically available to applications via a rich set of

APIs. To demonstrate the capabilities of the platform, the researchers have prototyped a few

applications using these APIs, including a misplaced object alert service, an enhanced

calendar service, and rich annotation of video with physical events.

Figure 4.3.4: Components of RFID

4.4 APPLICATIONS

The basic operations such as enabling clock, inbox, browsing, searching gallery,

calendar, seeing contact list etc are performed regularly in the mobile every time. These

operations can be stored as commands in the IC and then can be accessed on the screen or

over any surface using our technology within fractions of seconds.

4.4.1 MAKE A CALL

The sixth sense can project the keypad onto one’s hand which can be used as a virtual

screen to make a call. You can use the Sixth Sense to project a keypad onto your hand, then

use that virtual keypad to make a call. Calling a number also will not be a great task with the

Page 24: SIXTH SENSE TECHNOLOGY REPORT

introduction of Sixth Sense Technology. No mobile device will be required, just type in the

number with your palm acting as the virtual keypad. The keys will come up on the fingers.

The fingers of the other hand will then be used to key in the number and call.

Figure 4.4.1: Making a call

4.4.2 CHECK THE TIME

When you draw a circle on your wrist, a virtual watch appears that gives you the correct

time. Sixth Sense all we have to do is draw a circle on our wrist with our index finger to get a

virtual watch that gives us the correct time. The computer tracks the red marker cap or piece

of tape, recognizes the gesture, and instructs the projector to flash the image of a watch onto

his wrist.

Figure 4.4.2: Checking time

4.4.3 ZOOMING FEATURES

The user can zoom in or zoom out by using their intuitive hand movements.

Page 25: SIXTH SENSE TECHNOLOGY REPORT

Figure 4.4.3: zooming

4.4.4 GET PRODUCT INFORMATION

The sixth sense uses image recognition or a marker technology to recognize the products

we pick up and gives us the information on those products. For example, if you're trying to

shop "green" and are looking for paper towels with the least amount of bleach in them, the

system will scan the product you pick up off the shelf and give you guidance on whether this

product is a good choice for you.

Figure 4.4.4: Product information

4.4.5 GET FLIGHT UPDATES

By using our boarding pass, the system will recognize it and would let you know whether

your flight is on time and if the gate has changed.

Page 26: SIXTH SENSE TECHNOLOGY REPORT

Figure 4.4.5: Flight updates

4.4.6 TAKE A PICTURE

If you fashion your index finger and thumbs into a square (“framing” gesture), the

system snaps a photo at that time. After taking the desired number of photos, we can project

them onto a surface to view them.

Figure 4.4.6: Take pictures

4.4.7 CALL UP A MAP

It is helpful to call up the map of our choice and then use our thumbs and index

fingers to navigate the map. The sixth sense also implements map which lets the user display

the map on any physical surface and find his destination and he can use his thumbs and index

fingers to navigate the map, for example, to zoom in and out and do other controls.

Figure 4.4.7: Virtual map

4.4.8 CHECK THE EMAIL

Draw an @ symbol on any surface helps to check the email using the web service.

4.4.9 CREATE MULTIMEDIA READING EXPERIENCES

The Sixth Sense system also augments physical objects the user is interacting with by

projecting more information about these objects projected on them. For example, a

Page 27: SIXTH SENSE TECHNOLOGY REPORT

newspaper can show live video news or dynamic information can be provided on a regular

piece of paper. Thus a piece of paper turns into a video display.

Figure 4.4.9: Multimedia reading

4.4.10 FEED INFORMATION ON PEOPLE

It helps to display the relevant information about a person we are looking at. Sixth

Sense also is capable of "a more controversial use”. When you go out and meet someone,

projecting relevant information such as what they do, where they work, and also it could

display tags about the person floating on their shirt. It could be handy if it displayed their

facebook relationship status so that you knew not to waste your time.

Figure 4.4.10: info about person

4.5 ADVANTAGES

Portable, supports multi-touch and multi-user interaction, connectedness between

world and information, cost effective, data access directly from machine in real time, mind

map the idea anywhere, assists us in making right decisions, supports multi touch and multi

user interaction, the device serves the purpose of a computer plus saves time spent on

Page 28: SIXTH SENSE TECHNOLOGY REPORT

searching information. Sixth Sense also recognizes user’s freehand gestures (postures) and

saves electricity.

4.6 FUTURE ENHANCEMENTS

Imagine the world where Sixth Sense Technology is applied everywhere. In

educational field, the number of hardware components could be reduced. Usage of papers and

electricity could decrease. Students could use any wall or any surface wherever they are to

carry out activities that are done in a PC. Security will be assured for everyone. It could be

helpful in rendering defense services. In medical field, it could be implied to check the

genuinity of drugs. It could be implemented to monitor the agricultural lands. Blind people

could be able to read books and recognize objects. It could be used for the betterment of

handicapped people. Sixth sense could make the world magical.

Page 29: SIXTH SENSE TECHNOLOGY REPORT

CHAPTER 5

CONCLUSION

Sixth Sense technology recognizes the objects around us, displaying information

automatically and letting us to access it in any way we need. The Sixth Sense prototype

implements several applications that demonstrate the usefulness, viability and flexibility of

the system. Allowing us to interact with this information through natural hand gestures. The

potential of becoming the ultimate "transparent" user interface for accessing information

about everything around us. Currently the prototype of the device costs around $350 to build.

It could change the way we interact with the real world and truly give everyone complete

awareness of the environment around us. The Sixth Sense prototype implements several

applications that demonstrate the usefulness, viability and flexibility of the system. It will

definitely revolutionize the world.

The Sixth Sense software will be open source. As far as this seems to be a little set of

items, there will not be user interfaces or much advanced programs for the users. There will

be much harder and secured coding inside the device to make sure the security of the

software. It will be interesting to know the new language for coding for a sixth sense device.

Page 30: SIXTH SENSE TECHNOLOGY REPORT

BIBLIOGRAPHY

[1] Kirishima, T. Sato, K. Chihara, K.Dept. of Electr. Eng., Nara Nat. Coll. of Technol.,

Japan Robotics, “Gesture Spotting and Recognition for Human–Robot Interaction”, IEEE

Transactions on Volume: 23, Issue:2 pp256 – 270., April 2007.

[2] Alon,J.Athitsos, V.Quan, YuanSclarof,” A Unified Framework for Gesture Recognition

and Spatiotemporal Gesture Segmentation”, IEEE transactions on Pattern Analysis and

Machine Intelligence, Volume: 31, Issue:9 pp 1685 - 1699 ., Sept. 2009.

[3] Gomez, A.M. Peinado, A.M. Sanchez, V. Rubio, A.J.dept. eoria de la Senal,” Recognition

of coded speech transmitted over wireless channels Wireless Communications”, IEEE

Transactions on Volume: 5, Issue: 9, pp-2555 – 2562., Sept. 2006.

[4] Evans, J.R. Tjoland, W.A. Allred, L.G.Ogden Air Logistics Center, Hill AFB,” Achieving

a hands-free computer interface using voice recognition and speech synthesis”, IEEE

Volume: 15, Issue:1, pp 14 16., Jan 2000.

[5] Pelaez-Moreno, C. Gallardo-Antolin, A. Diaz-de-Maria,” Recognizing voice over IP: a

robust front-end for speech recognition on the world wide webMultimedia”, IEEE

Transactions on Volume: 3, Issue:2, pp-209 – 218., Jun 2001.


Recommended