ArticlePDF Available

AURA: A mobile platform for object and location annotation

Authors:

Abstract and Figures

In this paper, we describe a system used to link online content to physical objects implemented with commercially available pocket computers using integrated bar code scanners, wireless networks, and web services. We discuss our design goals and technical architecture and describe applications that have been constructed on this architecture. We describe the role of the related web site to create communities around scans collected by the handhelds.
Content may be subject to copyright.
AURA: A mobile platform for object and location
annotation
Marc Smith, Duncan Davenport, Howard Hwa
Microsoft Research
One Microsoft Way
Redmond, WA 98052 USA
+1 425 706 6896
{masmith, duncand, a-hhwa}@microsoft.com
ABSTRACT
In this paper, we describe a system used to link online
content to physical objects implemented with commercially
available pocket computers using integrated bar code
scanners, wireless networks, and web services. We discuss
our design goals and technical architecture and describe
applications that have been constructed on this architecture.
We describe the role of the related web site to create
communities around scans collected by the handhelds.
Keywords
Laminated reality, mobile object annotation, communities,
mobile devices, bar codes, machine readable object tags,
wireless networks
INTRODUCTION
Every object has a story to tell. However, labels and signs
can only tell part of this story; there is always an enormous
amount more to learn than will fit on a label. Mobile
devices are changing this, allowing physical objects to to be
linked to associated online content. This dramatically
expands the space for commentary and services related to
the places, products, and objects that physically surround
us.
The technical process of linking physical objects to online
content has become increasingly straight-forward. Adding
a tag reading device to a network connected portable
computer shortens the gap between physical objects and
places, and the digital information related to them. This
enables wirelessly networked devices to cheaply and
accurately recognize a wide range of objects and places,
and offer access to information and services pertaining to
those objects. It seems reasonable that some form or forms
of tag detectors will eventually be common features of most
networked information devices. Currently cameras and bar
code readers are widely available for cell phones and
pocket computers.
We created just such a system that combined widely
available wirelessly networked Pocket PC handheld
computers with a laser scanner for reading bar codes.
Client software was created to integrate these components
and connect them with servers available over the public
Internet.
The resulting system has applications in many settings.
Meta-data about objects with UPC codes, found on almost
all consumer products in the United States, can be drawn
from publicly accessible online data services. These
services often provide the name of the object or product, its
size (if it has one) and the name of its manufacturer in
exchange for the object’s bar coded identifier. Our system
uses such a data service to retrieve meta-data that is then
used to construct queries for search engines that yield useful
and highly relevant results. Scanned objects quickly link
back to the web sites for their manufacturers or online
commerce sites that offer those objects for sale. Similarly,
books often bear an ISBN number in the form of a bar
code. These numbers can be used in queries to online book
sellers, making the services offered there like book reviews,
lists of related books, and, of course, purchasing available
with just one scan and a tap.
Figure 1. Mobile device hardware platforms
composed of a Toshiba e740 and a Socket
Compact Flash Bar Code Scanner.
RELATED WORK
Several projects have explored the ways objects and places
can be linked to online content and services. Ljungstrand,
et al. (2000) have built the WebSticker system to link
barcodes to web pages. This was predominantly a desktop
bound system. There is a large body of work on “context-
aware” computing (Schilit, et al., 1994). Context-
awareness refers to the identification of a user’s proximate
environment for the delivery of computing content or
services. Xerox’s PARCTAB system uses custom built
infrared transceivers to help palm-sized computers to
identify their physical environments (Want, et al., 1995).
The Cyberguide uses Palm PDA’s to provide map guides to
tourists (Abowd, et al., 1997). Positioning in Cyberguide is
provided by a combination of custom applications based on
infrared sensing (for indoor) and GPS (for outdoor). MIT
LCS’ Cricket System deploys custom built RF and
ultrasound beacons for indoor navigation (Priyantha, et al.
2000).
The CoolTown Project at HP is building context-awareness
technologies to provide web presences for people, places
and things (Kindberg, et al., 2000). Similar to the MIT
Project Oxygen (MIT, 2002), CoolTown’s main goal is to
enable future “nomadic computing” such that computing
resources follow the human user and customize the human-
computing interaction based on the local human
environment.
Our approach is a more modest and potentially more
broadly deployable in the short term. Our goal is to enable
a light weight way to both access information about
physical objects and places and to add annotations to them.
This focus is different from, but complementary to, efforts
to link physical devices, like printers or projectors to device
based user interfaces.
HARDWARE PLATFORM
The mobile component of our system integrates three core
hardware features: a laser bar code scanner, a wireless
network connection, and a PDA. There are a number of
alternative sensors that could be usefully integrated into this
system, including GPS and wireless network signal strength
detection for location information and readers for the
emerging technology of RFID tags. To date we have only
made use of bar code readers but the system architecture is
extensible, allowing these or other emerging sensor
technologies to generate information that can be used to
identify objects or places.
SERVER
The server is comprised of three components: a web
service, runtime, and local and remote data stores. The
Web Service is the channel the client uses to communicate
with the backend server. This is accomplished entirely
using remote method invocation over HTTP (“web
services”). The web service is the interface to the backend
runtime for the clients. The Runtime provides the business
logic handling event tracing, retrieval, storage, rating
Symbology-
Payload
Mapping
Input
Data
Cache
Figure 2. AURA Architecture diagram
calculations, and other tasks. The local database stores
contain user profiles, barcodes, ratings, written and speech
annotations, which are stored in a SQL2000 database.
Information on books and UPC’s are provided by multiple
remote data stores including the Amazon Web Service for
books and music, and the ServiceObjects Web Service for
UPC lookup.
MOBILE CLIENT SOFTWARE
The client is a standalone application on the Pocket PC (as
opposed to a web front-end) to support improved user
interactivity. Network connectivity is not assumed to be
continuous for the mobile client. The client application
provides queuing and retry services for the storage and
retrieval of data to and from the backend servers. These
services are not possible for a thin web based client.
Caches or local stores on the client can dramatically reduce
the demand on network access for content. In addition, a
client side application allows for a richer user interface.
This is especially true when considering delays and
intermittent network connectivity.
CLIENT INTERFACE COMPONENTS
Users can login to the system by creating a unique
username and password combination either from the mobile
device or through the web portal interface. Without an
account the device can be used to scan objects but the
device creates an Anonymous User account and all the
comments created in that context are by default public.
When a user sees an object that interests them and finds a
bar code printed or affixed to it they point the head of the
device at the bar code from a distance of about 6-12 inches
and press the scan trigger button which we mapped to the
thumb button normally used to invoke the voice recorder
feature of the Pocket PC. If the device acquires the tag’s
data, the application gives the user feedback and based on
some properties of the bar code data and sends a series of
network queries out to appropriate web services.
We have initially created or linked to services to support
three types of bar codes: tags created for a local art gallery,
UPC (Universal Product Code) codes commonly used to
tag consumer products and foods, and ISBN (International
Standard Book Number) codes for books. Any number of
additional or alternate payloads are possible within this
framework to provide services for these or other forms of
object identifiers.
These payloads are linked to the resolution service registry
which contains pairs of pattern matches and pointers to
related web resources. When a tag is scanned it is matched
to an appropriate payload on the basis of the structure of the
identifier string. For example, ISBN codes start with “978”
and have a total of 14 digits. All bar codes starting with
that series of numbers with that number of digits are
assumed to be an ISBN and are submitted to web services
that are listed in the client’s directory of resolution services
that are registered as resolving such codes. We made use of
a web service offered by Amazon.com that returns metadata
about books and music when passed an ISBN number.
Figure 4. UPC Item Display Screen.
When objects with UPC codes are scanned the system
recognizes that the code is not in other classes of codes and
submits the identifier to a UPC mapping service. We made
use of a UPC metadata service provided to the public by
Figure 3. User scenario for grocery and related
retail environments. Query highlighted the recall of
the breakfast cereal by the FDA.
ServiceObjects.Net, a commercial web service provider.
This service returns a set of meta-data about the object and
the client presents this data and creates hyperlinks to search
engines based on the results. For example, when a box of
breakfast cereal is scanned the resulting display provides
two tap access to search results, the first of which notes that
the product has been recalled due to food safety issues
related to undocumented ingredients that might cause fatal
allergic reactions for some people (figure 4 and 5).
Figure 5. Search results linked from UPC meta data
WEB PORTAL
Users can access the system through a web portal as well as
the mobile device. Users can log into the web site and view
their scan history sorted by various properties of the items.
Scans can be sorted by time, by product category (books,
food stuffs, etc.), or by the ratings or comments of other
users or data found in other systems. This creates a simple
way to assemble inventories of tagged objects, for example
a collection of books, videos or music CDs. Alternatively,
it creates a diary-like history of the series of objects
scanned while, for example, browsing through a shopping
mall or museum gallery.
CONCLUSION
A wave of annotation systems for physical objects is likely
to be about to break. Cell phones are already integrating
digital cameras and have the processing power needed to
natively decode bar codes. As pocket computers merge
with cell phones the resulting hybrids will no doubt
combine a vision system with network connectivity and
computation. The widespread distribution of such devices
is likely to have dislocating effects in many sectors of life.
Retail environments seem the most likely to change as
consumers bring the power of the Internet to bear at the
point of sale.
REFERENCES
1.Rheingold, Howard. Smart Mobs: The Next Social
Revolution, Cambridge, MA: Perseus Publishing, 2002.
2.Fiore, Lee Teirnanan and Smith, 2001
3.Service Objects Universal Product Code Web Service
http://www.serviceobjects.com/products/dots_upc.asp?bh
cp=1
4.Abowd, G. D., et al. “Cyberguide: a Mobile Context-
aware Tour guide.” Wireless Networks, vol. 3, (1997) pp.
421-433.
5.Kindberg, T., et al. “People, Places, Things: Web
Presence for the Real World.” Proceedings
WMCSA2000, (2000).
6.Ljungstrand, P. , J. Redström, and L. E. Holmquist.
“Webstickers: Using Physical Tokens to Access, Manage
and Share Bookmarks to the Web.” Proceedings of
Designing Augmented Reality Environments (DARE)
2000 (2000).
7.MIT (2002) Project Oxygen. http://oxygen.lcs.mit.edu/
8.Priyantha, N. B., A. Chakraborty, and H. Balakrishnan.
“The Cricket Location-Support System.” 6th ACM
International Conference on Mobile Computing and
Networking (2000).
9.Schilit, B., and R. Want. “Context-Aware Computing
Applications.” IEEE Workshop on Mobile Computing
Systems and Applications (1994).
10.Want, R., et al. “The PARCTAB Ubiquitous Computing
Experiment.” Technical Report CSL-95-1, Xerox Palo
Alto Research Center (1995).
11.Want, R., et al. “Bridging Physical and Virtual Worlds
with Electronic Tags.” In Proceedings CHI 1999 (1999).
... Smart products -products that share information with consumers -are designed to combine the online and offline world. They give physical store operators in today's world of intense online shopping the possibility to compete with online retailers by providing direct access to the online information (Maass and Janzen 2007, Roussos 2005, Smith et al. 2003. Smart products communicate with the consumer and also enable new ways of interaction . ...
... RFID-based application myGrocer (Kourouthanassis and Roussos 2003), the mobile device and barcode-based AURA system (Smith et al. 2003) and conceptual applications for socially-conscious practices (Novak 2005) have been investigated. Although the authors claim user benefits, it is not clear if consumers and retailers will really respond to these systems. ...
... Toward this concept, different projects aimed to identify the objects grasped to support the more reliable and precise classification of activities. First approaches, such as by Smith et al. (2003), used barcodes to tag objects. Following research used radio-frequency identification (RFID) tags instead, attached to certain objects, tools, or even the environment, e.g., at a doorknob in Berlin et al. (2010). ...
Article
Full-text available
With a smaller form factor and a larger set of applications, body-worn devices have evolved into a collection of simultaneously deployed hardware units, rather than into a single all-round wearable. The sensor data, logged by such devices across the user's body, contains a wealth of information but is often difficult to synchronize. Especially the application of machine learning techniques, e.g., for activity recognition, suffers from the inaccuracy of the devices' internal clocks. In recent years, intra-body communication emerged as a promising alternative to the traditional wired and wireless communication techniques. Distributed wearable systems will notably benefit from its advantages, such as a superior energy efficiency. However, due to the absence of commercially available platforms, applications using this innovative technique remain rare and underinvestigated. With IBSync, we present a novel concept in which artificial landmark signals are received by body-worn devices on touching, approaching, or passing certain areas, surfaces, or objects with embedded transmitter beacons. The landmark signals enable both the wearables' intentional or incidental synchronization as well as the implicit contextualization using supplementary information about the beacons' situational context. For the detection of the landmarks, we propose to repurpose the on-board ECG sensor front-end available in recent commercial wearable devices. Evaluated on a total of 215 min of recordings from two devices, we demonstrate the concept's feasibility and a promising synchronization error of 0.80±1.79 samples or 6.25±14.00 ms at a device's sampling rate of 128 Hz.
... Furthermore, we are focusing on the extension of our system to pervasive computing environment. Similar to AURA [22], we will be able to use mobile devices for being helped to travel and guide in smart space in which objects are semantically annotated. ...
Article
Full-text available
Web browsing task is based on depth-first searching scheme, so that searching relevant information from Web may be very tedious. In this paper, we propose personal browsing assistant system based on user intentions modeling. Before explicitly requested by a user, this system can analyze the prefetched resources from the hyperlinked Webpages and compare them with the estimated user intention, so that it can help him to make a better decision like which Webpage should be requested next. More important problem is the semantic heterogeneity between Web spaces. It makes the understandability of locally annotated resources more difficult. We apply semantic annotation, which is a transcoding procedure with the global ontology. Therefore, each local metadata can be semantically enriched, and efficiently comparable. As testing bed of our experiment, we organized three different online clothes stores whose images are annotated by semantically heterogeneous metadata. We simulated virtual customers navigating these cyberspaces. According to the predefined preferences of customer models, they conducted comparison-shopping. We have shown the reasonability of supporting the Web browsing, and its performance was evaluated as measuring the total size of browsed hyperspace.
... Both of these scenarios can be described as physical mobile interactions. This research field deals for instance with mobile interaction with enhanced physical objects [1] [2] [3], sensing the environment to get awareness of the context of the user [4] [5], mobile interaction with public and semi-public displays [6] [7], mobile interactions in smart environments [13] mobile annotations [8] or using the mobile device as a universal remote control [9]. Figure 1 illustrates typical examples of physical mobile interactions in which the user interacts via the mobile device with things, peoples and places [1] in the physical world. Herby we particularly focus on the interaction between mobile devices and objects in the physical world. ...
Article
Full-text available
Mobile interactions with the physical world, meaning a person uses her mobile device as mediator for the interaction with a physical object, get more and more popular in industry and academia. Typical technologies supporting this kind of interactions are Radio Frequency Identification (RFID), visual marker recognition, Near Field Communication (NFC), or Bluetooth. Currently, there only exists very little tool support for building systems that consider this kind of interactions. But this is necessary because of the complexity, variety and distribution of such systems. A framework would also support the development and the dissemination of physical mobile interactions in our everyday live. Therefore we present in this paper the requirements for such tool support, the architecture of the Physical Mobile Interaction Framework (PMIF) and a first version of the implementation.
... Webenhanced GPS [3] uses GPS to find a user's location, maps this via a database to a street address, then uses this as a search term to find relevant web pages. AURA [14] enables users to scan physical objects such as books or DVDs with their mobile devices to retrieve detailed information or commentary regarding these artifacts via the web. Two recent studies [15], [16] found that location-based services accessed by mobile devices would be well-received by users in general. ...
Article
We introduce AD LOC, a system for mobile users to col-laboratively tie persistent virtual notes to physical locations without the need for any servers embedded in the environment or accessed via the Internet. Instead, all notes are proactively cached solely on the mobile devices of passing participants and served up to others in the vicinity via ad hoc wireless protocols like WiFi. By making use of any of the increasingly ubiquitous positioning technologies, such as GPS, devices at-tempt to ensure notes remain cached at the physical locations they were published. Through simulation we demonstrate AD LOC's behaviour under different scenarios and caching strate-gies. Our results show that despite the unpredictable and in-constant storage medium, surprisingly few participants are re-quired for remarkably robust note persistence. Furthermore, the number of transmissions between devices needed to re-alise the caching strategies is relatively low and increases lin-early with the number of participants.
... runs on mobile phones and manages annotations at operator level. Recent applications also use Internet repositories to store user-provided annotations on physical objects and locations, two notable examples being Microsoft's Aura (Smith, Davenport & Hwa, 2003), based on bar codes, and Cellspotting (www.cellspotting.com), using the cell information provided by mobile phone operators. ...
Article
Full-text available
Mobile settings are not only physically and digitally mediated; they are also inhabited by people – a social space. We argue that careful design exposing the connections, gaps, overlays and mismatches within and between physical, digital and social space allow for a better understanding and thereby mastering of the resulting combined space. Two concepts are explored in MobiTip, a social mobile service for exchanging opinions among peers: intramedia seams concerning network coverage and position technology, and intermedia seams between digitally transmitted tips and the physical, social context surrounding the user. We introduce social positioning as an alternative and a complement to the current strive for seamless connectedness and exact positioning in physical space.
... This is achievable through data exchange via the mobile phone network. Many projects have investigated linking online information and services to physical media [1,3,4,8,9,11,12,13]. Whereas these projects have relied on some labelling technology, like visual markers or RFID tags, we try to achieve physical hyperlinking without any modification of the object itself and just rely on its visual appearance. ...
Article
This paper explores the use of physical signs as anchors for digital annotations and other information. In our prototype system, the user carries a camera-equipped handheld device with pen-based input to capture street signs, restaurant signs, and shop signs. Image matching is supported by interactively established point correspondences. The cap-tured image along with context information is transferred to a back-end server, which performs image matching and returns the results to the user. We present a comparison of four different algorithms for the sign matching task. We found that the SIFT algorithm performs best. More-over, we discovered that lighting conditions – especially glare – have a crucial impact on the recognition rate.
Article
We present a study of using camera-phones and visual-tags to access mobile services. Firstly, a user-experience study is described in which participants were both observed learning to interact with a prototype mobile service and interviewed about their experiences. Secondly, a pointing-device task is presented in which quantitative data was gathered regarding the speed and accuracy with which participants aimed and clicked on visual-tags using camera-phones. We found that participants’ attitudes to visual-tag-based applications were broadly positive, although they had several important reservations about camera-phone technology more generally. Data from our pointing-device task demonstrated that novice users were able to aim and click on visual-tags quickly (well under 3 s per pointing-device trial on average) and accurately (almost all meeting our defined speed/accuracy tradeoff of 6% error-rate). Based on our findings, design lessons for camera-phone and visual-tag applications are presented.
Article
Full-text available
This paper considers the possibilities and implications of using ubiquitous computing to augment retail stores with information on critical consumer-driven issues, such as health and food safety, environmental and ethical concerns. The main hypothesis is that the availability of such socially conscious consumer information at point-of-purchase is expected to modulate consumer product choice criteria towards socially responsible consumption and increase their capabilities of negotiation with retailers and producers. This would effect changes in marketing strategies and production processes of retailers and producers towards more socially-conscious practices which still have a sound commercial basis. We oppose the common assumption that such consumer-support systems would benefit only the consumers at the expense of retailers, hence making them non-realistic. To this end we present scenarios and a model for a consumer-support system that supports the interests and value propositions of different actors involved in the retail chain. The relevance of the proposed approach is supported by findings from the sociology of consumption and studies on the influence of consumer opinion forums on marketing strategies of companies. Critical issues for technological realization, such as unobtrusive interfaces and contextualised information presentation for highly-situated interaction are also discussed.
Conference Paper
Full-text available
In the WebStickers system, where barcode stickers may be attached to physical objects making them act as bookmarks to the worldwide web in a convenient way to the user. Using readily available technology, i.e., standard barcode readers and adhesive stickers, WebStickers enable users to take advantage of their physical environment when organizing and sharing bookmarks. Starting from a user-centered rather than technology-driven point of view, we discuss how the affordances of physical tokens, as well as the context they are placed in, can act as useful cues for users. Since many objects already have barcodes printed on them, they can be used with the WebStickers system without physical modification. In addition, WebStickers meets proposed design criteria for information workspaces.
Article
Full-text available
Future computing environments will free the user from the constraints of the desktop. Applications for a mobile environment should take advantage of contextual information, such as position, to offer greater services to the user. In this paper, we present the Cyberguide project, in which we are building prototypes of a mobile context-aware tour guide. Knowledge of the user's current location, as well as a history of past locations, are used to provide more of the kind of services that we come to expect from a real tour guide. We describe the architecture and features of a variety of Cyberguide prototypes developed for indoor and outdoor use on a number of different hand-held platforms. We also discuss the general research issues that have emerged in our context-aware applications development in a mobile environment.
Conference Paper
Full-text available
This paper describes systems that examine and react to an individual's changing context. Such systems can promote and mediate people's interactions with devices, computers, and other people, and they can help navigate unfamiliar places. We believe that a limited amount of information covering a person's proximate environment is most important for this form of computing since the interesting part of the world around us is what we can see, hear, and touch. In this paper we define context-aware computing, and describe four catagories of context-aware applications: proximate selection, automatic contextual reconfiguration, contextual information and commands, and contex-triggered actions. Instances of these application types have been prototyped on the PARCTAB, a wireless, palm-sized computer.
Article
Full-text available
This paper describes the UbiquitousComputing philosophy, the PARCTAB system, user-interface issues for small devices, and our experience developing and testing a variety of mobile applications.
Article
Full-text available
This paper describes systems that examine and react to an individual's changing context. Such systems can promote and mediate people's interactions with devices, computers, and other people, and they can help navigate unfamiliar places. We believe that a limited amount of information covering a person's proximate environment is most important for this form of computing since the interesting part of the world around us is what we can see, hear, and touch. In this paper we define context-aware computing, and describe four categories of context-aware applications: proximate selection, automatic contextual reconfiguration, contextual information and commands, and context-triggered actions. Instances of these application types have been prototyped on the ParcTab, a wireless, palm-sized computer. 1 Introduction Our investigation focuses on an extended form of mobile computing in which users employ many different mobile, stationary and embedded computers over the course of the day. In this ...
Article
This paper presents the design, implementation, and evaluation of Cricket, a location-support system for in-building, mobile, locationdependent applications. It allows applications running on mobile and static nodes to learn their physical location by using listeners that hear and analyze information from beacons spread throughout the building. Cricket is the result of several design goals, including user privacy, decentralized administration, network heterogeneity, and low cost. Rather than explicitly tracking user location, Cricket helps devices learn where they are and lets them decide whom to advertise this information to; it does not rely on any centralized management or control and there is no explicit coordination between beacons; it provides information to devices regardless of their type of network connectivity; and each Cricket device is made from off-the-shelf components and costs less than U.S. $10. We describe the randomized algorithm used by beacons to transmit information, the use of concurrent radio and ultrasonic signals to infer distance, the listener inference algorithms to overcome multipath and interference, and practical beacon configuration and positioning techniques that improve accuracy. Our experience with Cricket shows that several location-dependent applications such as in-building active maps and device control can be developed with little effort or manual configuration. 1