SlideShare una empresa de Scribd logo
1 de 6
Proceedings of 2004 IEEE/RSJ International Conference on
Intelligent Robots and Systems
September 28 - October 2, 2004, Sendai, Japan




    RFID in Robot-Assisted Indoor Navigation for
               the Visually Impaired
         Vladimir Kulyukin                Chaitanya Gharpure           John Nicholson               Sachin Pavithran
  Department of Computer Science              Department of Computer Science               Center for Persons with Disabilities
       Utah State University                       Utah State University                          Utah State University
            Logan, UT                                  Logan, Utah                                     Logan, UT
       vkulyukin@cs.usu.edu                cpg@cc.usu.edu, jnicholson@cc.usu.edu                  sachin@cpd2.usu.edu



   Abstract— We describe how Radio Frequency Identification            centers, are a perfect niche for robotic guides. Guide dogs
(RFID) can be used in robot-assisted indoor navigation for            and white canes are of limited use in such environments,
the visually impaired. We present a robotic guide for the             because they do not have any environment-specific topo-
visually impaired that was deployed and tested both with
and without visually impaired participants in two indoor              logical knowledge and, consequently, cannot help their
environments. We describe how we modified the standard                 users find paths to useful destinations.
potential fields algorithms to achieve navigation at moderate
walking speeds and to avoid oscillation in narrow spaces.                                II. R ELATED W ORK
The experiments illustrate that passive RFID tags deployed               The idea of robotic guides is not new. Horswill[5] used
in the environment can act as reliable stimuli that trigger local
navigation behaviors to achieve global navigation objectives.         the situated activity theory to build Polly, a robotic guide
                                                                      for the MIT AI Lab. Polly used lightweight vision routines
                     I. I NTRODUCTION                                 that depended on textures specific to the lab. Thrun et al.[6]
   For most visually impaired people the main barrier to              built Minerva, a completely autonomous tour-guide robot
improving their quality of life is the inability to navigate.         that was deployed in the National Museum of American
This inability denies the visually impaired equal access              History in Washington, D.C. Burgard et al.[7] developed
to buildings, limits their use of public transportation, and          RHINO, a close sibling of Minerva’s, which was deployed
makes the visually impaired in the United States a group              as an interactive tour guide in the Deutsches Museum in
with one of the highest unemployment rates (74%)[1].                  Bonn, Germany.
   Robot-assisted navigation can help the visually impaired              Unfortunately, these robots do not address the needs of
overcome the navigation barrier for several reasons. First,           the visually impaired. First, the robots depend on the users’
the amount of body gear required by wearable navigation               ability to maintain visual contact with them, which cannot
technologies, e.g., [2], [3], is significantly minimized, be-          be assumed for the visually impaired. The only way users
cause most of it is mounted on the robot and powered from             could interact with Polly[5] was by tapping their feet. To
on-board batteries. Consequently, the navigation-related              request a museum tour from RHINO[7], the user had to
physical load is significantly reduced. Second, the user can           identify and press a button of a specific color on the robot’s
interact with the robot in ways unimaginable with guide               panel. Second, these solutions require substantial invest-
dogs and white canes, i.e., speech, wearable keyboard,                ments in customized engineering to become operational,
audio, etc. These interaction modes make the user feel                which makes it difficult to use them as models of replicable
more at ease and reduce her navigation-related cognitive              solutions that work out of the box in a variety of environ-
load. Third, the robot can interact with other people in the          ments. The approach on which Polly is based requires that
environment, e.g., ask them to yield. Fourth, robotic guides          a robot be evolved by its designer to fit its environment not
can carry useful payloads, e.g., suitcases and grocery bags.          only in terms of software but also in terms of hardware.
Finally, the user can use robotic guides in conjunction               The probabilistic localization algorithms used in RHINO
with her conventional navigation aids, e.g., white canes              and MINERVA, require a great deal of processing power.
and guide dogs.                                                       For example, to remain operational, RHINO had to run 20
   What environments are suitable for robotic guides?                 parallel processes on 3 on-board PCs and 2 off-board SUN
There is little need for such guides in familiar environ-             workstations connected via a customized Ethernet-based
ments where conventional navigation aids are adequate.                point-to-point socket communication protocol. Even with
For example, a guide dog typically picks a route after                these high software and hardware commitments, RHINO
three to five trials. While there is a great need for assisted         reportedly experienced six collisions over a period of forty-
navigation outdoors, the robotic solutions, due to severe             seven hours, although each tour was less than ten mintues
sensor challenges, have so far been inadequate for the                long[7].
job and have not compared favorably to guide dogs[4].                    Mori and Kotani[4] developed HARUNOBU-6, a robotic
Therefore, we believe that unfamiliar indoor environments             travel aid to guide the visually impaired on the Ya-
that are dynamic and complex, e.g., airports and conference           manashi University campus. HARUNOBU-6 is a motor




0-7803-8463-6/04/$20.00 ©2004 IEEE
                                                                    1979
(a) RG                            (b) An RFID Tag                        (c) Navigation

                                               Fig. 1.    Robot-Assisted Navigation



wheel chair equipped with a vision system, sonars, a                two to three hours, and require only commercial off-the-
differential GPS, and a portable GIS. While the wheel               shelf (COTS) hardware components; 2) that sensors be
chair is superior to the guide dog in its knowledge of the          inexpensive, reliable, easy to maintain (no external power
environment, the experiments run by the HARUNOBU-6                  supply), and provide accurate localization; 3) that all com-
research team demonstrated that the wheel chair is inferior         putation run onboard the robot; 4) that robot navigation be
to the guide dog in mobility and obstacle avoidance. The            smooth (few sideways jerks and abrupt velocity changes)
major source of problems was vision-based navigation                and keep pace with a moderate walking speed; and 5) that
because the recognition of patterns and landmarks was               human-robot interaction be both reliable and intuitive from
greatly influenced by the time of day, weather, and season.          the perspective of the visually impaired users.
Additionally, HARUNOBU-6 is a higly customized piece                   The first two requirements make the systems that sat-
of equipment, which negatively affects its portability across       isfy them replicable, maintainable, and robust. The third
a broad spectrum of environments.                                   requirement eliminates the necessity of running substantial
   Several research efforts in mobile robotics are similar to       off-board computation to keep the robot operational. In
the research described in this paper in that they also use          emergency situations, e.g., computer security breaches,
RFID technology for robot navigation. Kantor and Singh              power failures, and fires, off-board computers are likely to
used RFID tags for robot localization and mapping[8].               become dysfunctional and paralyze the robot if it depends
Once the positions of the RFID tags are known, their                on them. The last two requirements explicitly consider
system uses time-of-arrival type of information to estimate         the needs of the target population and make our project
the distance from detected tags. Tsukiyama[9] developed a           different from the RFID-based robot navigation systems
navigation system for mobile robots using RFID tags. The            mentioned above.
system assumes perfect signal reception and measurement
and does not deal with uncertainty. H¨hnel et al.[10]
                                           a                        A. Hardware
developed a robotic mapping and localization system to                 RG is built on top of the Pioneer 2DX commercial
analyze whether RFID can be used to improve the lo-                 robotic platform [12] (See Figure 1(a)). The platform has
calization of mobile robots in office environments. They             three wheels, 16 ultrasonic sonars, 8 in front and 8 in the
proposed a probabilistic measurement model for RFID                 back, and is equipped with three rechargeable Power Sonic
readers that accurately localizes RFID tags in a simple             PS-1270 onboard batteries that can operate for up to two
office environment.                                                  hours at a time.
                                                                       What turns the platform into a robotic guide is a
III. A ROBOTIC G UIDE FOR THE V ISUALLY I MPAIRED                   Wayfinding Toolkit (WT) mounted on top of the platform
   In May 2003, the Department of Computer Science of               and powered from the on-board batteries. As can be seen
Utah State Univeristy (USU) and the USU Center for                  in Figure 1(a), the WT currently resides in a PVC pipe
Persons with Disabilities launched a collaborative project          structure attached to the top of the platform. The WT’s
whose objective is to build an indoor robotic guide for the         core component is a Dell laptop connected to the platform’s
visually impaired. In this paper, we describe a prototype           microcontroller. The laptop has a Pentium 4 mobile 1.6
we have built and deployed in two indoor environments. Its          GHz processor with 512 MB of RAM. Communication
name is RG, which stands for “robotic guide.” RG is shown           between the laptop and the microcontroller is done through
in Figure 1(a). We refer to our approach as unintrusive             a usb-to-serial cable. The laptop interfaces to a radio-
instrumentation of environments. Our current research ob-           frequency identification (RFID) reader through another
jective is to alleviate localization and navigation problems        usb-to-serial cable. The TI Series 2000 RFID reader is
of purely autonomous approaches by instrumenting envi-              connected to a square 200mm × 200mm antenna. The
ronments with inexpensive and reliable sensors that can be          arrow in Figure 1(b) points to a TI RFID Slim Disk tag
placed in and out of environments without disrupting any            attached to a wall. Only these RFID tags are currently
indigenous activities. Effectively, the environment becomes         used by the system. These tags can be attached to any
a distributed tracking and guidance system[11]. Additional          objects in the environment or worn on clothing. They do
requirements are: 1) that the instrumentation be fast, e.g.,        not require any external power source or direct line of sight




                                                                1980
(a) Empty Spaces                                                  (b) RG’s Grid

                                             Fig. 2.   Potential Fields and Empty Spaces.



to be detected by the RFID reader. They are activated by             is iteratively decreased by 100 mm. In Figure 2(a), laser
the spherical electromagnetic field generated by the RFID             readings R1 and R2 are the boundary readings of the
antenna with a radius of approximately 1.5 meters. Each              maximum empty space. All readings between R1 and R2
tag is programmatically assigned a unique ID.                        are greater than the threshold. The next step is to find the
   A dog leash is attached to the battery bay handle on the          target direction. For that, we find the midway point between
back of the platform. The upper end of the leash is hung             R1 and R2 and the direction to that point is the target
on a PCV pole next to the RFID antenna’s pole. As shown              direction αt .
in Figure 1(c), visually impaired individuals follow RG by              RG’s PF is a 10 × 30 egocentric grid. Each cell in the
holding onto that leash.                                             grid is 200mm × 200mm. The grid covers an area of 12
B. Navigation                                                        square meters (2 meters in front and 3 meters on each side)
                                                                     in front of RG. Each cell Cij holds a vector that contributes
   Since RG’s objective is to assist the visually impaired           in calculating the resultant PF vector. The direction to the
in navigating unknown environments, we had to pay close              cell from RG’s center is αij . There are three types of
attention to three navigational features. First, RG should           cells: 1) occupied cells, which hold the repulsive vectors
move at moderate walking speeds. For example, the robot              generated by walls and obstacles; 2) free cells, which
developed by H¨hnel et al.[10] travels at an average speed
                a                                                    hold the vector pointing in the target direction obtained
of 0.223 m/s, which is too slow for our purposes because             by finding the maximum empty space; and 3) unknown
it is slower than a moderate walking speed (0.7 m/s) by              cells, the contents of which are unknown, since they lie
almost half a meter. Second, the motion must be smooth,              beyond detected obstacles. Unknown cells do not hold any
without sideways jerks or abrupt speed changes. Third, RG            vectors. In Figure 2(b), dark gray cells are occupied, white
should be able to avoid obstacles.                                   cells are free, and light gray cells are unknown. If d(Cij )
   RG navigates in indoor environments using potential
                                                                     is the distance from the robot’s center to the cell Cij in
fields (PF) and by finding empty spaces around itself.
                                                                     the grid, L(αij ) is the laser reading in the cell’s direction,
PFs have been widely used in navigation and obstacle
                                                                     and T is a tolerance constant, then the occupation of Cij
avoidance[13]. The basic concept behind the PF approach
                                                                     is computed by the function ζ(i, j):
is to populate the robot’s sensing grid with an vector field
in which the robot is repulsed away from obstacles and                                  ⎧
                                                                                        ⎨1       if |d(Cij ) − L(αij )| < T
attracted towards the target. Thus, the walls and obstacles                   ζ(i, j) = 0        if L(αij ) − d(Cij ) > T        (1)
around RG generate a PF in which RG acts like a moving                                  ⎩
                                                                                          −1 if d(Cij ) − L(αij ) > T
particle[14]. The desired direction of travel is the direction
of the maximum empty space around RG, which, when                       In Equation 1, the constants 1, 0, -1 denote occupied,
found, becomes the target direction to guide RG through              free, and unknown, respectively. By default, all vectors in
the PF. This simple strategy takes explicit advantage of             occupied and free cells are unit vectors. However, since
the way human indoor environments are organized. For                 closer obstacles have more effect on the robot, the vector
example, if the maximum empty space is in front, the                 magnitude increases with the proximity of the cell to the
navigator can keep moving forward; if the maximum empty              robot. Therefore, the vector magnitude in the cell is a
space is on the left, a left turn can be made, etc. This             function of the cell’s row and column.
strategy allows RG to follow hallways, avoid obstacles, and             A repulsive vector in Cij is denoted as Rij (mij , −αij ),
turn without using any orientation sensors, such as digital          where mij is the vector’s magnitude and −αij is its
compasses or inertia cubes.                                          direction. The magnitude is inversely proportional to the
   To find the maximum empty space, RG uses a total of 90             distance of the occupied cell from the robot. For the left-
laser range finder readings, taken at every 2 degrees. The            side vectors, mij = M agn(i, j) ∗ P1 ; for the right-side
readings are taken every millisecond. An initial threshold of        vectors, mij = M agn(i, j) ∗ P2 , where M agn(i, j) is
3000 mm is used. If no empty space is found, this threshold          the magnitude of the corresponding vector and P1 and P2




                                                                  1981
are constants that vary the replusion vectors on the robot’s      nents: 1) a map server, 2) a path planner, and 3) a speech
left and right sides, respectively. Thus, one can adjust the      recognition and synthesis engine.
distance maintained by RG from the right or left wall,               The Map Server realizes the causal and topological
respectively. Since RG’s localization relies on RFID tags         levels of the SSH. The server’s knowledge base repre-
placed in hallways, it has to navigate closer to the right        sents a connectivity graph of the environment in which
wall, which is achieved by increasing the repulsive force         RG operates. No global map is assumed. In addition,
of the left side vectors. Repulsive vectors for occupied          the knowledge base contains tag to destination mappings
cells are summed up to get the resultant repulsive vector         and simple behavior trigger/disable scripts associated with
Rr = ij Rij .                                                     specific tags. The Map Server continuously registers the
   The target vector in a free cell Cij is denoted as             latest location of RG on the connectivity graph. The
Tij (M, αt ), where M is the vector’s magnitude. In our           location is updated as soon as RG detects a RFID tag.
implementation, all vectors in the unoccupied cells are unit      Given the connectivity graph, the Path Planner uses the
vectors. The resultant target vector is Tr =         ij Tij =     standard breadth first search algorithm to find a path from
Tr (M ∗N, αt ), where N is the number of unoccupied cells.        one location to the other. A path plan is a sequence
The resultant vector, RES, is the sum of the repulsive and        of tag numbers and behavior scripts at each tag. Thus,
target vectors: RES(mr , αr ) = Rr + Tr , where, mr is the        RG’s trips are sequences of locally triggered behaviors that
magnitude and αr is the direction.                                achieve global navigation objectives. The SSH metric level
   To ensure smooth turns and avoid abrupt speed changes,         is not implemented, because, as studies in mobile robotics
RG never stops and turns in place. Instead, RG sets the           show[16], [14], odometry, from which metric information
left (V1 ) and right (V2 ) wheel velocities, to produce a         is typically obtained, is not reliable in robotic navigation.
smooth turn. V1 and V2 are functions of mr and αr :
V1 = V2 = v + (αr ∗ S)/mr , v is the robot’s velocity,            D. Human-Robot Interaction
and S is a constant that determines the sharpness of turns;          Human-robot interaction in RG is described in detail
αr is positive for left turns and negative for right. The         elsewhere[18], [19]. Here we give a brief summary only
robot’s velocity v is a function of the front distance. Thus,     for the sake of completeness. Visually impaired users can
if mr is large, the turns are less sharp. This is precisely       interact with RG through speech and wearable keyboards.
why RG follows a smooth, straight path even in narrow             Speech is received by RG through a wireless microphone
hallways without oscillating, which has been a problem            placed on the user’s clothing. Speech is recognized and
for some PF algorithms[15]. Given this implementation,            synthesized with Microsoft Speech API (SAPI) 5.1. RG
RG maintains, at most times, a moderate walking speed of          interacts with its users and people in the environment
0.7 m/s without losing smoothness or robustness.                  through speech and audio icons, i.e., non-verbal sounds that
                                                                  are readily associated with specific objects, e.g., the sound
C. Ethology and Spatial Semantic Hierarchy                        of water bubbles associated with a water cooler. When RG
   As a software system, RG is based on Kupiers’                  is passing a water cooler, it can either say “water cooler”
Spatial Semantic Hierarchy (SSH)[16] and Tinbergen’s              or play an audio file with sounds of water bubbles. We
ethology[17]. The SSH is a framework for representing             added audio icons to the system because, as recent research
spatial knowledge. It divides spatial knowledge of au-            findings indicate [20], speech perception can be slow and
tonomous agents into four levels: control, causal, topolog-       prone to block ambient sounds from the environment. To
ical, and metric. The control level consists of low level         other people in the environment, RG is personified as
mobility laws, e.g., trajectory following and aligning with       Merlin, a Microsoft software character, always present on
a surface. The causal level represents the world in terms         the WT laptop’s screen.
of views and actions. A view is a collection of data items
that an agent gathers from its sensors. Actions move agents                           IV. E XPERIMENTS
from view to view. The topological level represents the              We deployed our system for a total of approximately
world’s connectivity, i.e., how different locations are con-      seventy hours in two indoor environments: the Assistive
nected. The metric level adds distances between locations.        Technology Laboratory (ATL) of the USU Center for
   In RG, the control level is implemented with the PF            Persons with Disabilities and the USU CS Department. The
methods described above and includes the following be-            ATL occupies part of a floor in a building on the USU
haviors: follow-wall, turn-left, turn-right, avoid-obstacles,     North Campus. The floor has an area of approximately
go-thru-doorway, pass-doorway, and make-u-turn. These             4,270 square meters. The floor contains 6 laboratories, two
behaviors are coordinated and controlled through Tinber-          bathrooms, two staircases, and an elevator. The CS Depart-
gen’s release mechanisms[17]. RFID tags are viewed as             ment occupies an entire floor in a multi-floor building. The
stimuli that trigger or disable specific behaviors. To ensure      floor’s area is 6,590 square meters. The floor contains 23
portabilty, all these behaviors are written in the behavior       offices, 7 laboratories, a conference room, a student lounge,
programming language of the ActivMedia Robotics Inter-            a tutor room, two elevators, several bathrooms, and two
face for Applications (ARIA) system from ActivMedia               staircases.
Robotics, Inc. The routines run on the WT laptop. In                 Forty RFID tags were deployed at the ATL and one hun-
addition, the WT laptop runs three other software compo-          dred tags were deployed at the CS Department. It took one




                                                                1982
(a) Narrow (1m wide) Hallway Runs       (b) Medium (1.5m wide) Hallway Runs             (c) Wide (2.5m wide) Hallway Runs

                                               Fig. 3.    Path Deviations in Hallways




      (a) Narrow (1m wide) Hallway Runs       (b) Medium (1.5m wide) Hallway Runs             (c) Wide (2.5m wide) Hallway Runs

                                              Fig. 4.    Velocity Changes in Hallways.



person 20 minutes to deploy the tags and about 10 minutes            was computed as the average of the distances taken during
to remove them at the ATL. The same measurements at the              the run. Once the ideal distances were known, we ran the
CS Department were 30 and 20 minutes, respectively. As               robot three times in each type of hallway. The hallways
Figure 1(b) indicates, the tags were placed on small pieces          in which the robot ran were different from the hallways
of cardboard to insulate them from the walls and were                in which the ideal distances were computed. Obstacles,
attached to the walls with regular scotch tape. The creation         e.g., humans walking by and open doors, were allowed
of the connectivity graphs, took one hour at the ATL and             during the test runs. Figure 3 gives the distance graphs
about 2 hours at the CS Department. One administrator                of the three runs compared in each hallway type. The
first walked around the areas with a laptop and recorded              vertical bars in each graph represent the robot’s width.
tag-destination associations and then associated behavior            As can be seen from Figure 3(a), there is almost no
scripts with tags.                                                   deviation from the ideal distance in narrow hallways. Nor
   RG was first repeatedly tested in the ATL, the smaller             is there any oscillation. Figure 3(b) and Figure 3(c) show
of the two environments, and then deployed for pilot                 some insignificant deviations from the ideal distance. The
experiments at the USU CS Department. We ran two sets                deviations were caused by people walking by and by
of pilot experiments. The first set did not involve visually          open doors. However, there is no oscillation, i.e., sharp
impaired participants. The second set did. In the first set of        movements in different directions. In both environments,
experiments, we had RG navigate three types of hallways              we observed several tag detection failures, particularly in
of the CS Department: narrow (1 m), medium (1.5 m) and               metallic door frames. However, after we insulated the tags
wide (2.5 m) and estimated its navigation in terms of two            with small pieces of cardboard (see Figure 1(b)), the tag
variables: path deviations and abrupt speed changes. We              detection failures stopped.
also wanted to test how well RG’s RFID reader detected                  Figure 4 gives the velocity graphs for each hallway type
the tags.                                                            (x-axis is time in seconds, y-axis is velocity in mm/sec).
   To estimate path deviations, in each experiment we first           The graphs show that the narrow hallways cause short
computed the ideal distance that the robot has to maintain           abrupt changes in velocity. This is because in narrow
from the right wall in a certain type of hallway (narrow,            hallways even a slight disorientation, e.g., 3 degrees, in the
medium, and wide). The ideal distance was computed by                robot causes changes in velocity because less free space is
running the robot in a hallway of that type with all doors           detected in the grid. In medium and wide hallways, the
closed and no obstacles en route. During the run, the                velocity remains mostly smooth. Several speed changes
distance read by the laser range finder between the robot             occur when the robot passes or navigates through doorways
and the right wall was recorded every 50 milliseconds. In            or avoids obstacles.
recording the distance, the robot orientation was taken into            The second set of pilot experiments involved five vi-
account from two consecutive readings. The ideal distance            sually impaired participants, one participant at a time,




                                                                 1983
over a period of two months. Three participants were             Research Initiative (CURI) grant from the State of Utah,
completely blind and two participants could perceive only        and through a New Faculty Research grant from Utah State
light. The participants had no speech impediments, hearing       University.
problems, or cognitive disabilities. Two participants were
dog users and the other three used white canes. The
participants were asked to use RG to navigate to three                                       R EFERENCES
distinct locations (an office, a lounge, and a bathroom)           [1] M. P. LaPlante and D. Carlson, Disability in the United States:
at the USU CS Department. All participants were new                   Prevalence and Causes. Washington, DC: U.S. Department of Ed-
to the environment and had to navigate approximately 40               ucation, National Institute of Disability and Rehabilitation Research,
                                                                      2000.
meters to get to all destinations. Thus, in the experiments       [2] S. Shoval, J. Borenstein, and Y. Koren, “Mobile Robot Obstacle
with visually impaired participants, the robot navigated              Avoidance in a Computerized Travel for the Blind,” in IEEE
approximately 200 meters. All participants reached their              International Conference on Robotics and Automation, San Diego,
                                                                      CA, 1994.
destinations without a problem. In their exit interviews,         [3] D. Ross and B. Blasch, “Development of a Wearable Computer
the participants complained mostly about the human-robot              Orientation System,” IEEE Personal and Ubiquitous Computing,
interaction aspects of the system. For example, all of them           vol. 6, pp. 49–63, 2002.
                                                                  [4] H. Mori and S. Kotani, “Robotic Travel Aid for the Blind:
had problems with the speech recognition system[21], [19].            HARUNOBU-6,” in Second European Conference on Disability,
The participants especially liked the fact that they did not                                                        o
                                                                      Virtual Reality, and Assistive Technology, S¨vde, Sweden, 1998.
have to give up their white canes and guide dogs to use           [5] I. Horswill, “Polly: A Vision-Based Artificial Agent,” in Pro-
                                                                      ceedings of the 11th Conference of the American Association for
RG.                                                                   Artificial Intelligence (AAAI-93), Washington, DC, July 1993.
                                                                  [6] S. Thrun, M. Bennewitz, W. Burgard, A. B. Cremers, F. Dellaert,
                     V. L IMITATIONS                                                    a
                                                                      D. Fox, D. H¨hnel, C. Rosenberg, N. Roby, J. Schutle, and
                                                                      D. Schultz, “Minerva: A Second Generation Mobile Tour-Guide
   In addition to velocity changes in narrow hallways, RG             Robot,” in Proceedings of the IEEE International Conference on
has three other limitations. First, the robot cannot create           Robotics and Automation (ICRA-99), Antwerp, Belgium, June 1999.
a connectivity graph for a given environment once the                                                                a
                                                                  [7] W. Burgard, A. Cremers, D. Fox, D. H¨hnel, G. Lakemeyer,
                                                                      D. Schulz, W. Steiner, and S. Thrun, “Experiences with an Inter-
RFID tags are deployed. We are currently working on                   active Museum Tour-Guide Robot,” Artificial Intelligence, no. 114,
creating connectivity graphs and behavior scripts in a semi-          pp. 3–55, 1999.
automatic fashion. Second, the robot cannot detect route          [8] G. Kantor and S. Singh, “Preliminary Results in Range-Only Lo-
                                                                      calization and Mapping,” in Proceedings of the IEEE Conference
blockages. If the route is blocked, the robot first slows              on Robotics and Automation, Washington, DC, May 2002.
down to a stop and then starts turning in order to find            [9] T. Tsukiyama, “Navigation System for the Mobile Robots using
some free space. In this fashion, RG makes a gradual u-               RFID Tags,” in Proceedings of the IEEE Conference on Advanced
                                                                      Robotics, Coimbra, Portugal, June-July 2003.
turn by looking for the maximum free space around itself.                   a
                                                                 [10] D. H¨hnel, W. Burgard, D. Fox, K. Fishkin, and M. Philipose,
Since RG has no orientation sensor, currently the only way            “Mapping and localization with rfid technology,” Intel Research
it can detect a detour is by detecting an RFID tag that is            Institute, Seattle, WA, Tech. Rep. IRS-TR-03-014, December 2003.
                                                                 [11] V. Kulyukin and M. Blair, “Distributed Tracking and Guidance in
not on the path to the current destination. Finally, while            Indoor Environments,” in Conference of the Rehabilitation Engi-
several visually impaired participants told us that it would          neering and Assistive Technology Society of North America (RESNA-
be helpful if RG could guide them in and out of elevators,            2003), Atlanta, GA, June 2003.
                                                                 [12] http://www.activmedia.com, ActivMedia Robotic Platforms. Activ-
RG cannot negotiate elevators yet.                                    Media Robotics, Inc.
                                                                 [13] J. H. Chuang and N. Ahuja, “An Analytically Tractable Potential
                    VI. C ONCLUSION                                   Field Model of Free Space and its Application in Obstacle Avoid-
                                                                      ance,” IEEE Trans. Sys. Man, Cyb., vol. 5, no. 28, pp. 729–736,
   In this paper, we showed how Radio Frequency Iden-                 1998.
tification (RFID) can be used in robot-assisted indoor            [14] R. Murphy, Introduction to AI Robotics. Cambridge, MA: The MIT
navigation for the visually impaired. We presented a robotic          Press, 2000.
                                                                 [15] Y. Koren and J. Borenstein, “Potential Field Methods and their
guide for the visually impaired that was deployed and                 Inherent Limitations for Mobile Robot Navigation,” in Proceedings
tested both with and without visually impaired participants           of the IEEE Conference on Robotics and Automation, Sacramento,
in two indoor environments. The experiments illustrate that           CA, April 1991.
                                                                 [16] B. Kupiers, “The Spatial Semantic Hierarchy,” Artificial Intelligence,
passive RFID tags can act as reliable stimuli that trigger            no. 119, pp. 191–233, 2000.
local navigation behaviors to achieve global navigation          [17] N. Tinbergen, Animal in its World: Laboratory Experiments and
objectives.                                                           General Papers. Cambridge, MA: Harvard University Press, 1976.
                                                                 [18] V. Kulyukin, “Towards Hands-Free Human-Robot Interaction
                                                                      through Spoken Dialog,” in AAAI Spring Symposium on Human
                   ACKNOWLEDGMENT                                     Interaction with Autonomous Systems in Complex Environments,
   The authors would like to thank the visually impaired              Palo Alto, CA, March 2003.
                                                                 [19] ——, “Human-Robot Interaction through Gesture-Free Spoken Di-
participants for generously volunteering their time for the           alogue,” Autonomous Robots,, vol. 16, no. 3, 2004.
pilot experiments. The authors would like to thank Marty         [20] T. V. Tran, T. Letowski, and K. S. Abouchacra, “Evaluation of
Blair, Director of the Utah Assistive Technology Program              Acoustic Beacon Characteristics for Navigation Tasks,” Ergonomics,
                                                                      vol. 43, no. 6, pp. 807–827, 2000.
for his administrative assistance and support. The first          [21] V. Kulyukin, C. Gharpure, and N. De Graw, “Human-Robot In-
author would like to acknowledge that this research has               teraction in a Robotic Guide for the Visually Impaired,” in AAAI
been supported, in part, through the NSF Universal Ac-                Spring Symposium on Interaction between Humans and Autonomous
                                                                      Systems over Extended Operation, Palo Alto, CA, March 2004.
cess Career Grant (IIS-0346880), a Community University




                                                               1984

Más contenido relacionado

Destacado

Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...
Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...
Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...Vladimir Kulyukin
 
Presentatie Gastcollege Gamification voor Hogeschool Utrecht - ICT opleiding
Presentatie Gastcollege Gamification voor Hogeschool Utrecht - ICT opleidingPresentatie Gastcollege Gamification voor Hogeschool Utrecht - ICT opleiding
Presentatie Gastcollege Gamification voor Hogeschool Utrecht - ICT opleidingSasja Beerendonk
 
Tatacarashalatrasulutklelakiperempuan 1219158319928446-8
Tatacarashalatrasulutklelakiperempuan 1219158319928446-8Tatacarashalatrasulutklelakiperempuan 1219158319928446-8
Tatacarashalatrasulutklelakiperempuan 1219158319928446-8YISC Al-Azhar
 
Robot-Assisted Shopping for the Blind: Issues in Spatial Cognition and Produc...
Robot-Assisted Shopping for the Blind: Issues in Spatial Cognition and Produc...Robot-Assisted Shopping for the Blind: Issues in Spatial Cognition and Produc...
Robot-Assisted Shopping for the Blind: Issues in Spatial Cognition and Produc...Vladimir Kulyukin
 
Jaslin's 21 day OBS journey
Jaslin's 21 day OBS journeyJaslin's 21 day OBS journey
Jaslin's 21 day OBS journeyTerence Lee
 
Rijksmonumenten.info
Rijksmonumenten.infoRijksmonumenten.info
Rijksmonumenten.infoArchief 2.0
 
PermaMatrix is cheaper than dirt!
PermaMatrix is cheaper than dirt!PermaMatrix is cheaper than dirt!
PermaMatrix is cheaper than dirt!Sarah Probst
 
Surface-Embedded Passive RF Exteroception: Kepler, Greed, and Buffon’s Needle
Surface-Embedded Passive RF Exteroception: Kepler, Greed, and Buffon’s NeedleSurface-Embedded Passive RF Exteroception: Kepler, Greed, and Buffon’s Needle
Surface-Embedded Passive RF Exteroception: Kepler, Greed, and Buffon’s NeedleVladimir Kulyukin
 
Stadsarchief Ieper zkt. publiek
Stadsarchief Ieper zkt. publiekStadsarchief Ieper zkt. publiek
Stadsarchief Ieper zkt. publiekArchief 2.0
 
The Blind Leading the Blind: Toward Collaborative Online Route Information Ma...
The Blind Leading the Blind: Toward Collaborative Online Route Information Ma...The Blind Leading the Blind: Toward Collaborative Online Route Information Ma...
The Blind Leading the Blind: Toward Collaborative Online Route Information Ma...Vladimir Kulyukin
 
Silver heroes roundtable maart2015
Silver heroes roundtable maart2015Silver heroes roundtable maart2015
Silver heroes roundtable maart2015Sasja Beerendonk
 
An Algorithm for In-Place Vision-Based Skewed 1D Barcode Scanning in the Cloud
An Algorithm for In-Place Vision-Based Skewed 1D Barcode Scanning in the CloudAn Algorithm for In-Place Vision-Based Skewed 1D Barcode Scanning in the Cloud
An Algorithm for In-Place Vision-Based Skewed 1D Barcode Scanning in the CloudVladimir Kulyukin
 
Narrative Map Augmentation with Automated Landmark Extraction and Path Inference
Narrative Map Augmentation with Automated Landmark Extraction and Path InferenceNarrative Map Augmentation with Automated Landmark Extraction and Path Inference
Narrative Map Augmentation with Automated Landmark Extraction and Path InferenceVladimir Kulyukin
 
Negara yang bebas korupsi
Negara yang bebas korupsiNegara yang bebas korupsi
Negara yang bebas korupsiFelice Luce
 

Destacado (20)

Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...
Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...
Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...
 
Presentatie Gastcollege Gamification voor Hogeschool Utrecht - ICT opleiding
Presentatie Gastcollege Gamification voor Hogeschool Utrecht - ICT opleidingPresentatie Gastcollege Gamification voor Hogeschool Utrecht - ICT opleiding
Presentatie Gastcollege Gamification voor Hogeschool Utrecht - ICT opleiding
 
Mengenal sunnah
Mengenal sunnahMengenal sunnah
Mengenal sunnah
 
Tatacarashalatrasulutklelakiperempuan 1219158319928446-8
Tatacarashalatrasulutklelakiperempuan 1219158319928446-8Tatacarashalatrasulutklelakiperempuan 1219158319928446-8
Tatacarashalatrasulutklelakiperempuan 1219158319928446-8
 
Mengenal sunnah
Mengenal sunnahMengenal sunnah
Mengenal sunnah
 
PDF Test
PDF TestPDF Test
PDF Test
 
Robot-Assisted Shopping for the Blind: Issues in Spatial Cognition and Produc...
Robot-Assisted Shopping for the Blind: Issues in Spatial Cognition and Produc...Robot-Assisted Shopping for the Blind: Issues in Spatial Cognition and Produc...
Robot-Assisted Shopping for the Blind: Issues in Spatial Cognition and Produc...
 
Jaslin's 21 day OBS journey
Jaslin's 21 day OBS journeyJaslin's 21 day OBS journey
Jaslin's 21 day OBS journey
 
Rijksmonumenten.info
Rijksmonumenten.infoRijksmonumenten.info
Rijksmonumenten.info
 
PermaMatrix is cheaper than dirt!
PermaMatrix is cheaper than dirt!PermaMatrix is cheaper than dirt!
PermaMatrix is cheaper than dirt!
 
EUscreen
EUscreenEUscreen
EUscreen
 
Surface-Embedded Passive RF Exteroception: Kepler, Greed, and Buffon’s Needle
Surface-Embedded Passive RF Exteroception: Kepler, Greed, and Buffon’s NeedleSurface-Embedded Passive RF Exteroception: Kepler, Greed, and Buffon’s Needle
Surface-Embedded Passive RF Exteroception: Kepler, Greed, and Buffon’s Needle
 
Mengenal sunnah sii
Mengenal sunnah siiMengenal sunnah sii
Mengenal sunnah sii
 
Stadsarchief Ieper zkt. publiek
Stadsarchief Ieper zkt. publiekStadsarchief Ieper zkt. publiek
Stadsarchief Ieper zkt. publiek
 
The Blind Leading the Blind: Toward Collaborative Online Route Information Ma...
The Blind Leading the Blind: Toward Collaborative Online Route Information Ma...The Blind Leading the Blind: Toward Collaborative Online Route Information Ma...
The Blind Leading the Blind: Toward Collaborative Online Route Information Ma...
 
Silver heroes roundtable maart2015
Silver heroes roundtable maart2015Silver heroes roundtable maart2015
Silver heroes roundtable maart2015
 
An Algorithm for In-Place Vision-Based Skewed 1D Barcode Scanning in the Cloud
An Algorithm for In-Place Vision-Based Skewed 1D Barcode Scanning in the CloudAn Algorithm for In-Place Vision-Based Skewed 1D Barcode Scanning in the Cloud
An Algorithm for In-Place Vision-Based Skewed 1D Barcode Scanning in the Cloud
 
PizzaRaptor
PizzaRaptorPizzaRaptor
PizzaRaptor
 
Narrative Map Augmentation with Automated Landmark Extraction and Path Inference
Narrative Map Augmentation with Automated Landmark Extraction and Path InferenceNarrative Map Augmentation with Automated Landmark Extraction and Path Inference
Narrative Map Augmentation with Automated Landmark Extraction and Path Inference
 
Negara yang bebas korupsi
Negara yang bebas korupsiNegara yang bebas korupsi
Negara yang bebas korupsi
 

Similar a RFID in Robot-Assisted Indoor Navigation for the Visually Impaired

RoboCart: Toward Robot-Assisted Navigation of Grocery Stores by the Visually ...
RoboCart: Toward Robot-Assisted Navigation of Grocery Stores by the Visually ...RoboCart: Toward Robot-Assisted Navigation of Grocery Stores by the Visually ...
RoboCart: Toward Robot-Assisted Navigation of Grocery Stores by the Visually ...Vladimir Kulyukin
 
Passive Radio Frequency Exteroception in Robot-Assisted Shopping for the Blind
Passive Radio Frequency Exteroception in Robot-Assisted Shopping for the BlindPassive Radio Frequency Exteroception in Robot-Assisted Shopping for the Blind
Passive Radio Frequency Exteroception in Robot-Assisted Shopping for the BlindVladimir Kulyukin
 
Robot-Assisted Wayfinding for the Visually Impaired in Structured Indoor Envi...
Robot-Assisted Wayfinding for the Visually Impaired in Structured Indoor Envi...Robot-Assisted Wayfinding for the Visually Impaired in Structured Indoor Envi...
Robot-Assisted Wayfinding for the Visually Impaired in Structured Indoor Envi...Vladimir Kulyukin
 
Robot operating system based autonomous navigation platform with human robot ...
Robot operating system based autonomous navigation platform with human robot ...Robot operating system based autonomous navigation platform with human robot ...
Robot operating system based autonomous navigation platform with human robot ...TELKOMNIKA JOURNAL
 
military surveillance robot
military surveillance robot  military surveillance robot
military surveillance robot KrishGupta94
 
Paper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdf
Paper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdfPaper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdf
Paper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdfSathishRadcliffe
 
Design & Development of Vision Controlled Snake Robot
Design & Development of Vision Controlled Snake RobotDesign & Development of Vision Controlled Snake Robot
Design & Development of Vision Controlled Snake Robotvivatechijri
 
High-Speed Neural Network Controller for Autonomous Robot Navigation using FPGA
High-Speed Neural Network Controller for Autonomous Robot Navigation using FPGAHigh-Speed Neural Network Controller for Autonomous Robot Navigation using FPGA
High-Speed Neural Network Controller for Autonomous Robot Navigation using FPGAiosrjce
 
Figure 1
Figure 1Figure 1
Figure 1butest
 
Figure 1
Figure 1Figure 1
Figure 1butest
 
Figure 1
Figure 1Figure 1
Figure 1butest
 
AN APPROACH OF IR-BASED SHORT-RANGE CORRESPONDENCE SYSTEMS FOR SWARM ROBOT BA...
AN APPROACH OF IR-BASED SHORT-RANGE CORRESPONDENCE SYSTEMS FOR SWARM ROBOT BA...AN APPROACH OF IR-BASED SHORT-RANGE CORRESPONDENCE SYSTEMS FOR SWARM ROBOT BA...
AN APPROACH OF IR-BASED SHORT-RANGE CORRESPONDENCE SYSTEMS FOR SWARM ROBOT BA...ijaia
 
Wireless Indoor Localization with Dempster-Shafer Simple Support Functions
Wireless Indoor Localization with Dempster-Shafer Simple Support FunctionsWireless Indoor Localization with Dempster-Shafer Simple Support Functions
Wireless Indoor Localization with Dempster-Shafer Simple Support FunctionsVladimir Kulyukin
 
IRJET- Path Finder with Obstacle Avoidance Robot
IRJET-  	  Path Finder with Obstacle Avoidance RobotIRJET-  	  Path Finder with Obstacle Avoidance Robot
IRJET- Path Finder with Obstacle Avoidance RobotIRJET Journal
 
introduction to Robotics (the role of computer science)
introduction to Robotics (the role of computer science)introduction to Robotics (the role of computer science)
introduction to Robotics (the role of computer science)RajeshSharma1041
 

Similar a RFID in Robot-Assisted Indoor Navigation for the Visually Impaired (20)

RoboCart: Toward Robot-Assisted Navigation of Grocery Stores by the Visually ...
RoboCart: Toward Robot-Assisted Navigation of Grocery Stores by the Visually ...RoboCart: Toward Robot-Assisted Navigation of Grocery Stores by the Visually ...
RoboCart: Toward Robot-Assisted Navigation of Grocery Stores by the Visually ...
 
Passive Radio Frequency Exteroception in Robot-Assisted Shopping for the Blind
Passive Radio Frequency Exteroception in Robot-Assisted Shopping for the BlindPassive Radio Frequency Exteroception in Robot-Assisted Shopping for the Blind
Passive Radio Frequency Exteroception in Robot-Assisted Shopping for the Blind
 
Robot-Assisted Wayfinding for the Visually Impaired in Structured Indoor Envi...
Robot-Assisted Wayfinding for the Visually Impaired in Structured Indoor Envi...Robot-Assisted Wayfinding for the Visually Impaired in Structured Indoor Envi...
Robot-Assisted Wayfinding for the Visually Impaired in Structured Indoor Envi...
 
Ts 2 b topic
Ts 2 b topicTs 2 b topic
Ts 2 b topic
 
Robot operating system based autonomous navigation platform with human robot ...
Robot operating system based autonomous navigation platform with human robot ...Robot operating system based autonomous navigation platform with human robot ...
Robot operating system based autonomous navigation platform with human robot ...
 
military surveillance robot
military surveillance robot  military surveillance robot
military surveillance robot
 
Paper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdf
Paper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdfPaper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdf
Paper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdf
 
Design & Development of Vision Controlled Snake Robot
Design & Development of Vision Controlled Snake RobotDesign & Development of Vision Controlled Snake Robot
Design & Development of Vision Controlled Snake Robot
 
H011114758
H011114758H011114758
H011114758
 
High-Speed Neural Network Controller for Autonomous Robot Navigation using FPGA
High-Speed Neural Network Controller for Autonomous Robot Navigation using FPGAHigh-Speed Neural Network Controller for Autonomous Robot Navigation using FPGA
High-Speed Neural Network Controller for Autonomous Robot Navigation using FPGA
 
Figure 1
Figure 1Figure 1
Figure 1
 
Figure 1
Figure 1Figure 1
Figure 1
 
Figure 1
Figure 1Figure 1
Figure 1
 
Shamira-IIECE.pptx
Shamira-IIECE.pptxShamira-IIECE.pptx
Shamira-IIECE.pptx
 
AN APPROACH OF IR-BASED SHORT-RANGE CORRESPONDENCE SYSTEMS FOR SWARM ROBOT BA...
AN APPROACH OF IR-BASED SHORT-RANGE CORRESPONDENCE SYSTEMS FOR SWARM ROBOT BA...AN APPROACH OF IR-BASED SHORT-RANGE CORRESPONDENCE SYSTEMS FOR SWARM ROBOT BA...
AN APPROACH OF IR-BASED SHORT-RANGE CORRESPONDENCE SYSTEMS FOR SWARM ROBOT BA...
 
Wireless Indoor Localization with Dempster-Shafer Simple Support Functions
Wireless Indoor Localization with Dempster-Shafer Simple Support FunctionsWireless Indoor Localization with Dempster-Shafer Simple Support Functions
Wireless Indoor Localization with Dempster-Shafer Simple Support Functions
 
IRJET- Path Finder with Obstacle Avoidance Robot
IRJET-  	  Path Finder with Obstacle Avoidance RobotIRJET-  	  Path Finder with Obstacle Avoidance Robot
IRJET- Path Finder with Obstacle Avoidance Robot
 
Robocup2006
Robocup2006Robocup2006
Robocup2006
 
FIANL REPORT
FIANL REPORTFIANL REPORT
FIANL REPORT
 
introduction to Robotics (the role of computer science)
introduction to Robotics (the role of computer science)introduction to Robotics (the role of computer science)
introduction to Robotics (the role of computer science)
 

Más de Vladimir Kulyukin

Toward Sustainable Electronic Beehive Monitoring: Algorithms for Omnidirectio...
Toward Sustainable Electronic Beehive Monitoring: Algorithms for Omnidirectio...Toward Sustainable Electronic Beehive Monitoring: Algorithms for Omnidirectio...
Toward Sustainable Electronic Beehive Monitoring: Algorithms for Omnidirectio...Vladimir Kulyukin
 
Digitizing Buzzing Signals into A440 Piano Note Sequences and Estimating Fora...
Digitizing Buzzing Signals into A440 Piano Note Sequences and Estimating Fora...Digitizing Buzzing Signals into A440 Piano Note Sequences and Estimating Fora...
Digitizing Buzzing Signals into A440 Piano Note Sequences and Estimating Fora...Vladimir Kulyukin
 
Generalized Hamming Distance
Generalized Hamming DistanceGeneralized Hamming Distance
Generalized Hamming DistanceVladimir Kulyukin
 
Adapting Measures of Clumping Strength to Assess Term-Term Similarity
Adapting Measures of Clumping Strength to Assess Term-Term SimilarityAdapting Measures of Clumping Strength to Assess Term-Term Similarity
Adapting Measures of Clumping Strength to Assess Term-Term SimilarityVladimir Kulyukin
 
A Cloud-Based Infrastructure for Caloric Intake Estimation from Pre-Meal Vide...
A Cloud-Based Infrastructure for Caloric Intake Estimation from Pre-Meal Vide...A Cloud-Based Infrastructure for Caloric Intake Estimation from Pre-Meal Vide...
A Cloud-Based Infrastructure for Caloric Intake Estimation from Pre-Meal Vide...Vladimir Kulyukin
 
Exploring Finite State Automata with Junun Robots: A Case Study in Computabil...
Exploring Finite State Automata with Junun Robots: A Case Study in Computabil...Exploring Finite State Automata with Junun Robots: A Case Study in Computabil...
Exploring Finite State Automata with Junun Robots: A Case Study in Computabil...Vladimir Kulyukin
 
Image Blur Detection with 2D Haar Wavelet Transform and Its Effect on Skewed ...
Image Blur Detection with 2D Haar Wavelet Transform and Its Effect on Skewed ...Image Blur Detection with 2D Haar Wavelet Transform and Its Effect on Skewed ...
Image Blur Detection with 2D Haar Wavelet Transform and Its Effect on Skewed ...Vladimir Kulyukin
 
Text Skew Angle Detection in Vision-Based Scanning of Nutrition Labels
Text Skew Angle Detection in Vision-Based Scanning of Nutrition LabelsText Skew Angle Detection in Vision-Based Scanning of Nutrition Labels
Text Skew Angle Detection in Vision-Based Scanning of Nutrition LabelsVladimir Kulyukin
 
Vision-Based Localization and Scanning of 1D UPC and EAN Barcodes with Relaxe...
Vision-Based Localization and Scanning of 1D UPC and EAN Barcodes with Relaxe...Vision-Based Localization and Scanning of 1D UPC and EAN Barcodes with Relaxe...
Vision-Based Localization and Scanning of 1D UPC and EAN Barcodes with Relaxe...Vladimir Kulyukin
 
Effective Nutrition Label Use on Smartphones
Effective Nutrition Label Use on SmartphonesEffective Nutrition Label Use on Smartphones
Effective Nutrition Label Use on SmartphonesVladimir Kulyukin
 
An Algorithm for Mobile Vision-Based Localization of Skewed Nutrition Labels ...
An Algorithm for Mobile Vision-Based Localization of Skewed Nutrition Labels ...An Algorithm for Mobile Vision-Based Localization of Skewed Nutrition Labels ...
An Algorithm for Mobile Vision-Based Localization of Skewed Nutrition Labels ...Vladimir Kulyukin
 
Skip Trie Matching: A Greedy Algorithm for Real-Time OCR Error Correction on ...
Skip Trie Matching: A Greedy Algorithm for Real-Time OCR Error Correction on ...Skip Trie Matching: A Greedy Algorithm for Real-Time OCR Error Correction on ...
Skip Trie Matching: A Greedy Algorithm for Real-Time OCR Error Correction on ...Vladimir Kulyukin
 
Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...
Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...
Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...Vladimir Kulyukin
 
Skip Trie Matching for Real Time OCR Output Error Correction on Android Smart...
Skip Trie Matching for Real Time OCR Output Error Correction on Android Smart...Skip Trie Matching for Real Time OCR Output Error Correction on Android Smart...
Skip Trie Matching for Real Time OCR Output Error Correction on Android Smart...Vladimir Kulyukin
 
Toward Blind Travel Support through Verbal Route Directions: A Path Inference...
Toward Blind Travel Support through Verbal Route Directions: A Path Inference...Toward Blind Travel Support through Verbal Route Directions: A Path Inference...
Toward Blind Travel Support through Verbal Route Directions: A Path Inference...Vladimir Kulyukin
 
Eye-Free Barcode Detection on Smartphones with Niblack's Binarization and Sup...
Eye-Free Barcode Detection on Smartphones with Niblack's Binarization and Sup...Eye-Free Barcode Detection on Smartphones with Niblack's Binarization and Sup...
Eye-Free Barcode Detection on Smartphones with Niblack's Binarization and Sup...Vladimir Kulyukin
 
Eyesight Sharing in Blind Grocery Shopping: Remote P2P Caregiving through Clo...
Eyesight Sharing in Blind Grocery Shopping: Remote P2P Caregiving through Clo...Eyesight Sharing in Blind Grocery Shopping: Remote P2P Caregiving through Clo...
Eyesight Sharing in Blind Grocery Shopping: Remote P2P Caregiving through Clo...Vladimir Kulyukin
 
On Natural Language Dialogue with Assistive Robots
On Natural Language Dialogue with Assistive RobotsOn Natural Language Dialogue with Assistive Robots
On Natural Language Dialogue with Assistive RobotsVladimir Kulyukin
 
Ergonomics-for-One in a Robotic Shopping Cart for the Blind
Ergonomics-for-One in a Robotic Shopping Cart for the BlindErgonomics-for-One in a Robotic Shopping Cart for the Blind
Ergonomics-for-One in a Robotic Shopping Cart for the BlindVladimir Kulyukin
 
A Wearable Two-Sensor O&M Device for Blind College Students
A Wearable Two-Sensor O&M Device for Blind College StudentsA Wearable Two-Sensor O&M Device for Blind College Students
A Wearable Two-Sensor O&M Device for Blind College StudentsVladimir Kulyukin
 

Más de Vladimir Kulyukin (20)

Toward Sustainable Electronic Beehive Monitoring: Algorithms for Omnidirectio...
Toward Sustainable Electronic Beehive Monitoring: Algorithms for Omnidirectio...Toward Sustainable Electronic Beehive Monitoring: Algorithms for Omnidirectio...
Toward Sustainable Electronic Beehive Monitoring: Algorithms for Omnidirectio...
 
Digitizing Buzzing Signals into A440 Piano Note Sequences and Estimating Fora...
Digitizing Buzzing Signals into A440 Piano Note Sequences and Estimating Fora...Digitizing Buzzing Signals into A440 Piano Note Sequences and Estimating Fora...
Digitizing Buzzing Signals into A440 Piano Note Sequences and Estimating Fora...
 
Generalized Hamming Distance
Generalized Hamming DistanceGeneralized Hamming Distance
Generalized Hamming Distance
 
Adapting Measures of Clumping Strength to Assess Term-Term Similarity
Adapting Measures of Clumping Strength to Assess Term-Term SimilarityAdapting Measures of Clumping Strength to Assess Term-Term Similarity
Adapting Measures of Clumping Strength to Assess Term-Term Similarity
 
A Cloud-Based Infrastructure for Caloric Intake Estimation from Pre-Meal Vide...
A Cloud-Based Infrastructure for Caloric Intake Estimation from Pre-Meal Vide...A Cloud-Based Infrastructure for Caloric Intake Estimation from Pre-Meal Vide...
A Cloud-Based Infrastructure for Caloric Intake Estimation from Pre-Meal Vide...
 
Exploring Finite State Automata with Junun Robots: A Case Study in Computabil...
Exploring Finite State Automata with Junun Robots: A Case Study in Computabil...Exploring Finite State Automata with Junun Robots: A Case Study in Computabil...
Exploring Finite State Automata with Junun Robots: A Case Study in Computabil...
 
Image Blur Detection with 2D Haar Wavelet Transform and Its Effect on Skewed ...
Image Blur Detection with 2D Haar Wavelet Transform and Its Effect on Skewed ...Image Blur Detection with 2D Haar Wavelet Transform and Its Effect on Skewed ...
Image Blur Detection with 2D Haar Wavelet Transform and Its Effect on Skewed ...
 
Text Skew Angle Detection in Vision-Based Scanning of Nutrition Labels
Text Skew Angle Detection in Vision-Based Scanning of Nutrition LabelsText Skew Angle Detection in Vision-Based Scanning of Nutrition Labels
Text Skew Angle Detection in Vision-Based Scanning of Nutrition Labels
 
Vision-Based Localization and Scanning of 1D UPC and EAN Barcodes with Relaxe...
Vision-Based Localization and Scanning of 1D UPC and EAN Barcodes with Relaxe...Vision-Based Localization and Scanning of 1D UPC and EAN Barcodes with Relaxe...
Vision-Based Localization and Scanning of 1D UPC and EAN Barcodes with Relaxe...
 
Effective Nutrition Label Use on Smartphones
Effective Nutrition Label Use on SmartphonesEffective Nutrition Label Use on Smartphones
Effective Nutrition Label Use on Smartphones
 
An Algorithm for Mobile Vision-Based Localization of Skewed Nutrition Labels ...
An Algorithm for Mobile Vision-Based Localization of Skewed Nutrition Labels ...An Algorithm for Mobile Vision-Based Localization of Skewed Nutrition Labels ...
An Algorithm for Mobile Vision-Based Localization of Skewed Nutrition Labels ...
 
Skip Trie Matching: A Greedy Algorithm for Real-Time OCR Error Correction on ...
Skip Trie Matching: A Greedy Algorithm for Real-Time OCR Error Correction on ...Skip Trie Matching: A Greedy Algorithm for Real-Time OCR Error Correction on ...
Skip Trie Matching: A Greedy Algorithm for Real-Time OCR Error Correction on ...
 
Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...
Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...
Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...
 
Skip Trie Matching for Real Time OCR Output Error Correction on Android Smart...
Skip Trie Matching for Real Time OCR Output Error Correction on Android Smart...Skip Trie Matching for Real Time OCR Output Error Correction on Android Smart...
Skip Trie Matching for Real Time OCR Output Error Correction on Android Smart...
 
Toward Blind Travel Support through Verbal Route Directions: A Path Inference...
Toward Blind Travel Support through Verbal Route Directions: A Path Inference...Toward Blind Travel Support through Verbal Route Directions: A Path Inference...
Toward Blind Travel Support through Verbal Route Directions: A Path Inference...
 
Eye-Free Barcode Detection on Smartphones with Niblack's Binarization and Sup...
Eye-Free Barcode Detection on Smartphones with Niblack's Binarization and Sup...Eye-Free Barcode Detection on Smartphones with Niblack's Binarization and Sup...
Eye-Free Barcode Detection on Smartphones with Niblack's Binarization and Sup...
 
Eyesight Sharing in Blind Grocery Shopping: Remote P2P Caregiving through Clo...
Eyesight Sharing in Blind Grocery Shopping: Remote P2P Caregiving through Clo...Eyesight Sharing in Blind Grocery Shopping: Remote P2P Caregiving through Clo...
Eyesight Sharing in Blind Grocery Shopping: Remote P2P Caregiving through Clo...
 
On Natural Language Dialogue with Assistive Robots
On Natural Language Dialogue with Assistive RobotsOn Natural Language Dialogue with Assistive Robots
On Natural Language Dialogue with Assistive Robots
 
Ergonomics-for-One in a Robotic Shopping Cart for the Blind
Ergonomics-for-One in a Robotic Shopping Cart for the BlindErgonomics-for-One in a Robotic Shopping Cart for the Blind
Ergonomics-for-One in a Robotic Shopping Cart for the Blind
 
A Wearable Two-Sensor O&M Device for Blind College Students
A Wearable Two-Sensor O&M Device for Blind College StudentsA Wearable Two-Sensor O&M Device for Blind College Students
A Wearable Two-Sensor O&M Device for Blind College Students
 

Último

[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdfhans926745
 
CNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of ServiceCNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of Servicegiselly40
 
Developing An App To Navigate The Roads of Brazil
Developing An App To Navigate The Roads of BrazilDeveloping An App To Navigate The Roads of Brazil
Developing An App To Navigate The Roads of BrazilV3cube
 
Factors to Consider When Choosing Accounts Payable Services Providers.pptx
Factors to Consider When Choosing Accounts Payable Services Providers.pptxFactors to Consider When Choosing Accounts Payable Services Providers.pptx
Factors to Consider When Choosing Accounts Payable Services Providers.pptxKatpro Technologies
 
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...Neo4j
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024Rafal Los
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonAnna Loughnan Colquhoun
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024The Digital Insurer
 
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking MenDelhi Call girls
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Miguel Araújo
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountPuma Security, LLC
 
Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slidevu2urc
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonetsnaman860154
 
Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...Enterprise Knowledge
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsMaria Levchenko
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)Gabriella Davis
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationSafe Software
 
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...apidays
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024The Digital Insurer
 
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...Neo4j
 

Último (20)

[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf
 
CNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of ServiceCNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of Service
 
Developing An App To Navigate The Roads of Brazil
Developing An App To Navigate The Roads of BrazilDeveloping An App To Navigate The Roads of Brazil
Developing An App To Navigate The Roads of Brazil
 
Factors to Consider When Choosing Accounts Payable Services Providers.pptx
Factors to Consider When Choosing Accounts Payable Services Providers.pptxFactors to Consider When Choosing Accounts Payable Services Providers.pptx
Factors to Consider When Choosing Accounts Payable Services Providers.pptx
 
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt Robison
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024
 
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path Mount
 
Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slide
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonets
 
Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed texts
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
 
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024
 
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
 

RFID in Robot-Assisted Indoor Navigation for the Visually Impaired

  • 1. Proceedings of 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems September 28 - October 2, 2004, Sendai, Japan RFID in Robot-Assisted Indoor Navigation for the Visually Impaired Vladimir Kulyukin Chaitanya Gharpure John Nicholson Sachin Pavithran Department of Computer Science Department of Computer Science Center for Persons with Disabilities Utah State University Utah State University Utah State University Logan, UT Logan, Utah Logan, UT vkulyukin@cs.usu.edu cpg@cc.usu.edu, jnicholson@cc.usu.edu sachin@cpd2.usu.edu Abstract— We describe how Radio Frequency Identification centers, are a perfect niche for robotic guides. Guide dogs (RFID) can be used in robot-assisted indoor navigation for and white canes are of limited use in such environments, the visually impaired. We present a robotic guide for the because they do not have any environment-specific topo- visually impaired that was deployed and tested both with and without visually impaired participants in two indoor logical knowledge and, consequently, cannot help their environments. We describe how we modified the standard users find paths to useful destinations. potential fields algorithms to achieve navigation at moderate walking speeds and to avoid oscillation in narrow spaces. II. R ELATED W ORK The experiments illustrate that passive RFID tags deployed The idea of robotic guides is not new. Horswill[5] used in the environment can act as reliable stimuli that trigger local navigation behaviors to achieve global navigation objectives. the situated activity theory to build Polly, a robotic guide for the MIT AI Lab. Polly used lightweight vision routines I. I NTRODUCTION that depended on textures specific to the lab. Thrun et al.[6] For most visually impaired people the main barrier to built Minerva, a completely autonomous tour-guide robot improving their quality of life is the inability to navigate. that was deployed in the National Museum of American This inability denies the visually impaired equal access History in Washington, D.C. Burgard et al.[7] developed to buildings, limits their use of public transportation, and RHINO, a close sibling of Minerva’s, which was deployed makes the visually impaired in the United States a group as an interactive tour guide in the Deutsches Museum in with one of the highest unemployment rates (74%)[1]. Bonn, Germany. Robot-assisted navigation can help the visually impaired Unfortunately, these robots do not address the needs of overcome the navigation barrier for several reasons. First, the visually impaired. First, the robots depend on the users’ the amount of body gear required by wearable navigation ability to maintain visual contact with them, which cannot technologies, e.g., [2], [3], is significantly minimized, be- be assumed for the visually impaired. The only way users cause most of it is mounted on the robot and powered from could interact with Polly[5] was by tapping their feet. To on-board batteries. Consequently, the navigation-related request a museum tour from RHINO[7], the user had to physical load is significantly reduced. Second, the user can identify and press a button of a specific color on the robot’s interact with the robot in ways unimaginable with guide panel. Second, these solutions require substantial invest- dogs and white canes, i.e., speech, wearable keyboard, ments in customized engineering to become operational, audio, etc. These interaction modes make the user feel which makes it difficult to use them as models of replicable more at ease and reduce her navigation-related cognitive solutions that work out of the box in a variety of environ- load. Third, the robot can interact with other people in the ments. The approach on which Polly is based requires that environment, e.g., ask them to yield. Fourth, robotic guides a robot be evolved by its designer to fit its environment not can carry useful payloads, e.g., suitcases and grocery bags. only in terms of software but also in terms of hardware. Finally, the user can use robotic guides in conjunction The probabilistic localization algorithms used in RHINO with her conventional navigation aids, e.g., white canes and MINERVA, require a great deal of processing power. and guide dogs. For example, to remain operational, RHINO had to run 20 What environments are suitable for robotic guides? parallel processes on 3 on-board PCs and 2 off-board SUN There is little need for such guides in familiar environ- workstations connected via a customized Ethernet-based ments where conventional navigation aids are adequate. point-to-point socket communication protocol. Even with For example, a guide dog typically picks a route after these high software and hardware commitments, RHINO three to five trials. While there is a great need for assisted reportedly experienced six collisions over a period of forty- navigation outdoors, the robotic solutions, due to severe seven hours, although each tour was less than ten mintues sensor challenges, have so far been inadequate for the long[7]. job and have not compared favorably to guide dogs[4]. Mori and Kotani[4] developed HARUNOBU-6, a robotic Therefore, we believe that unfamiliar indoor environments travel aid to guide the visually impaired on the Ya- that are dynamic and complex, e.g., airports and conference manashi University campus. HARUNOBU-6 is a motor 0-7803-8463-6/04/$20.00 ©2004 IEEE 1979
  • 2. (a) RG (b) An RFID Tag (c) Navigation Fig. 1. Robot-Assisted Navigation wheel chair equipped with a vision system, sonars, a two to three hours, and require only commercial off-the- differential GPS, and a portable GIS. While the wheel shelf (COTS) hardware components; 2) that sensors be chair is superior to the guide dog in its knowledge of the inexpensive, reliable, easy to maintain (no external power environment, the experiments run by the HARUNOBU-6 supply), and provide accurate localization; 3) that all com- research team demonstrated that the wheel chair is inferior putation run onboard the robot; 4) that robot navigation be to the guide dog in mobility and obstacle avoidance. The smooth (few sideways jerks and abrupt velocity changes) major source of problems was vision-based navigation and keep pace with a moderate walking speed; and 5) that because the recognition of patterns and landmarks was human-robot interaction be both reliable and intuitive from greatly influenced by the time of day, weather, and season. the perspective of the visually impaired users. Additionally, HARUNOBU-6 is a higly customized piece The first two requirements make the systems that sat- of equipment, which negatively affects its portability across isfy them replicable, maintainable, and robust. The third a broad spectrum of environments. requirement eliminates the necessity of running substantial Several research efforts in mobile robotics are similar to off-board computation to keep the robot operational. In the research described in this paper in that they also use emergency situations, e.g., computer security breaches, RFID technology for robot navigation. Kantor and Singh power failures, and fires, off-board computers are likely to used RFID tags for robot localization and mapping[8]. become dysfunctional and paralyze the robot if it depends Once the positions of the RFID tags are known, their on them. The last two requirements explicitly consider system uses time-of-arrival type of information to estimate the needs of the target population and make our project the distance from detected tags. Tsukiyama[9] developed a different from the RFID-based robot navigation systems navigation system for mobile robots using RFID tags. The mentioned above. system assumes perfect signal reception and measurement and does not deal with uncertainty. H¨hnel et al.[10] a A. Hardware developed a robotic mapping and localization system to RG is built on top of the Pioneer 2DX commercial analyze whether RFID can be used to improve the lo- robotic platform [12] (See Figure 1(a)). The platform has calization of mobile robots in office environments. They three wheels, 16 ultrasonic sonars, 8 in front and 8 in the proposed a probabilistic measurement model for RFID back, and is equipped with three rechargeable Power Sonic readers that accurately localizes RFID tags in a simple PS-1270 onboard batteries that can operate for up to two office environment. hours at a time. What turns the platform into a robotic guide is a III. A ROBOTIC G UIDE FOR THE V ISUALLY I MPAIRED Wayfinding Toolkit (WT) mounted on top of the platform In May 2003, the Department of Computer Science of and powered from the on-board batteries. As can be seen Utah State Univeristy (USU) and the USU Center for in Figure 1(a), the WT currently resides in a PVC pipe Persons with Disabilities launched a collaborative project structure attached to the top of the platform. The WT’s whose objective is to build an indoor robotic guide for the core component is a Dell laptop connected to the platform’s visually impaired. In this paper, we describe a prototype microcontroller. The laptop has a Pentium 4 mobile 1.6 we have built and deployed in two indoor environments. Its GHz processor with 512 MB of RAM. Communication name is RG, which stands for “robotic guide.” RG is shown between the laptop and the microcontroller is done through in Figure 1(a). We refer to our approach as unintrusive a usb-to-serial cable. The laptop interfaces to a radio- instrumentation of environments. Our current research ob- frequency identification (RFID) reader through another jective is to alleviate localization and navigation problems usb-to-serial cable. The TI Series 2000 RFID reader is of purely autonomous approaches by instrumenting envi- connected to a square 200mm × 200mm antenna. The ronments with inexpensive and reliable sensors that can be arrow in Figure 1(b) points to a TI RFID Slim Disk tag placed in and out of environments without disrupting any attached to a wall. Only these RFID tags are currently indigenous activities. Effectively, the environment becomes used by the system. These tags can be attached to any a distributed tracking and guidance system[11]. Additional objects in the environment or worn on clothing. They do requirements are: 1) that the instrumentation be fast, e.g., not require any external power source or direct line of sight 1980
  • 3. (a) Empty Spaces (b) RG’s Grid Fig. 2. Potential Fields and Empty Spaces. to be detected by the RFID reader. They are activated by is iteratively decreased by 100 mm. In Figure 2(a), laser the spherical electromagnetic field generated by the RFID readings R1 and R2 are the boundary readings of the antenna with a radius of approximately 1.5 meters. Each maximum empty space. All readings between R1 and R2 tag is programmatically assigned a unique ID. are greater than the threshold. The next step is to find the A dog leash is attached to the battery bay handle on the target direction. For that, we find the midway point between back of the platform. The upper end of the leash is hung R1 and R2 and the direction to that point is the target on a PCV pole next to the RFID antenna’s pole. As shown direction αt . in Figure 1(c), visually impaired individuals follow RG by RG’s PF is a 10 × 30 egocentric grid. Each cell in the holding onto that leash. grid is 200mm × 200mm. The grid covers an area of 12 B. Navigation square meters (2 meters in front and 3 meters on each side) in front of RG. Each cell Cij holds a vector that contributes Since RG’s objective is to assist the visually impaired in calculating the resultant PF vector. The direction to the in navigating unknown environments, we had to pay close cell from RG’s center is αij . There are three types of attention to three navigational features. First, RG should cells: 1) occupied cells, which hold the repulsive vectors move at moderate walking speeds. For example, the robot generated by walls and obstacles; 2) free cells, which developed by H¨hnel et al.[10] travels at an average speed a hold the vector pointing in the target direction obtained of 0.223 m/s, which is too slow for our purposes because by finding the maximum empty space; and 3) unknown it is slower than a moderate walking speed (0.7 m/s) by cells, the contents of which are unknown, since they lie almost half a meter. Second, the motion must be smooth, beyond detected obstacles. Unknown cells do not hold any without sideways jerks or abrupt speed changes. Third, RG vectors. In Figure 2(b), dark gray cells are occupied, white should be able to avoid obstacles. cells are free, and light gray cells are unknown. If d(Cij ) RG navigates in indoor environments using potential is the distance from the robot’s center to the cell Cij in fields (PF) and by finding empty spaces around itself. the grid, L(αij ) is the laser reading in the cell’s direction, PFs have been widely used in navigation and obstacle and T is a tolerance constant, then the occupation of Cij avoidance[13]. The basic concept behind the PF approach is computed by the function ζ(i, j): is to populate the robot’s sensing grid with an vector field in which the robot is repulsed away from obstacles and ⎧ ⎨1 if |d(Cij ) − L(αij )| < T attracted towards the target. Thus, the walls and obstacles ζ(i, j) = 0 if L(αij ) − d(Cij ) > T (1) around RG generate a PF in which RG acts like a moving ⎩ −1 if d(Cij ) − L(αij ) > T particle[14]. The desired direction of travel is the direction of the maximum empty space around RG, which, when In Equation 1, the constants 1, 0, -1 denote occupied, found, becomes the target direction to guide RG through free, and unknown, respectively. By default, all vectors in the PF. This simple strategy takes explicit advantage of occupied and free cells are unit vectors. However, since the way human indoor environments are organized. For closer obstacles have more effect on the robot, the vector example, if the maximum empty space is in front, the magnitude increases with the proximity of the cell to the navigator can keep moving forward; if the maximum empty robot. Therefore, the vector magnitude in the cell is a space is on the left, a left turn can be made, etc. This function of the cell’s row and column. strategy allows RG to follow hallways, avoid obstacles, and A repulsive vector in Cij is denoted as Rij (mij , −αij ), turn without using any orientation sensors, such as digital where mij is the vector’s magnitude and −αij is its compasses or inertia cubes. direction. The magnitude is inversely proportional to the To find the maximum empty space, RG uses a total of 90 distance of the occupied cell from the robot. For the left- laser range finder readings, taken at every 2 degrees. The side vectors, mij = M agn(i, j) ∗ P1 ; for the right-side readings are taken every millisecond. An initial threshold of vectors, mij = M agn(i, j) ∗ P2 , where M agn(i, j) is 3000 mm is used. If no empty space is found, this threshold the magnitude of the corresponding vector and P1 and P2 1981
  • 4. are constants that vary the replusion vectors on the robot’s nents: 1) a map server, 2) a path planner, and 3) a speech left and right sides, respectively. Thus, one can adjust the recognition and synthesis engine. distance maintained by RG from the right or left wall, The Map Server realizes the causal and topological respectively. Since RG’s localization relies on RFID tags levels of the SSH. The server’s knowledge base repre- placed in hallways, it has to navigate closer to the right sents a connectivity graph of the environment in which wall, which is achieved by increasing the repulsive force RG operates. No global map is assumed. In addition, of the left side vectors. Repulsive vectors for occupied the knowledge base contains tag to destination mappings cells are summed up to get the resultant repulsive vector and simple behavior trigger/disable scripts associated with Rr = ij Rij . specific tags. The Map Server continuously registers the The target vector in a free cell Cij is denoted as latest location of RG on the connectivity graph. The Tij (M, αt ), where M is the vector’s magnitude. In our location is updated as soon as RG detects a RFID tag. implementation, all vectors in the unoccupied cells are unit Given the connectivity graph, the Path Planner uses the vectors. The resultant target vector is Tr = ij Tij = standard breadth first search algorithm to find a path from Tr (M ∗N, αt ), where N is the number of unoccupied cells. one location to the other. A path plan is a sequence The resultant vector, RES, is the sum of the repulsive and of tag numbers and behavior scripts at each tag. Thus, target vectors: RES(mr , αr ) = Rr + Tr , where, mr is the RG’s trips are sequences of locally triggered behaviors that magnitude and αr is the direction. achieve global navigation objectives. The SSH metric level To ensure smooth turns and avoid abrupt speed changes, is not implemented, because, as studies in mobile robotics RG never stops and turns in place. Instead, RG sets the show[16], [14], odometry, from which metric information left (V1 ) and right (V2 ) wheel velocities, to produce a is typically obtained, is not reliable in robotic navigation. smooth turn. V1 and V2 are functions of mr and αr : V1 = V2 = v + (αr ∗ S)/mr , v is the robot’s velocity, D. Human-Robot Interaction and S is a constant that determines the sharpness of turns; Human-robot interaction in RG is described in detail αr is positive for left turns and negative for right. The elsewhere[18], [19]. Here we give a brief summary only robot’s velocity v is a function of the front distance. Thus, for the sake of completeness. Visually impaired users can if mr is large, the turns are less sharp. This is precisely interact with RG through speech and wearable keyboards. why RG follows a smooth, straight path even in narrow Speech is received by RG through a wireless microphone hallways without oscillating, which has been a problem placed on the user’s clothing. Speech is recognized and for some PF algorithms[15]. Given this implementation, synthesized with Microsoft Speech API (SAPI) 5.1. RG RG maintains, at most times, a moderate walking speed of interacts with its users and people in the environment 0.7 m/s without losing smoothness or robustness. through speech and audio icons, i.e., non-verbal sounds that are readily associated with specific objects, e.g., the sound C. Ethology and Spatial Semantic Hierarchy of water bubbles associated with a water cooler. When RG As a software system, RG is based on Kupiers’ is passing a water cooler, it can either say “water cooler” Spatial Semantic Hierarchy (SSH)[16] and Tinbergen’s or play an audio file with sounds of water bubbles. We ethology[17]. The SSH is a framework for representing added audio icons to the system because, as recent research spatial knowledge. It divides spatial knowledge of au- findings indicate [20], speech perception can be slow and tonomous agents into four levels: control, causal, topolog- prone to block ambient sounds from the environment. To ical, and metric. The control level consists of low level other people in the environment, RG is personified as mobility laws, e.g., trajectory following and aligning with Merlin, a Microsoft software character, always present on a surface. The causal level represents the world in terms the WT laptop’s screen. of views and actions. A view is a collection of data items that an agent gathers from its sensors. Actions move agents IV. E XPERIMENTS from view to view. The topological level represents the We deployed our system for a total of approximately world’s connectivity, i.e., how different locations are con- seventy hours in two indoor environments: the Assistive nected. The metric level adds distances between locations. Technology Laboratory (ATL) of the USU Center for In RG, the control level is implemented with the PF Persons with Disabilities and the USU CS Department. The methods described above and includes the following be- ATL occupies part of a floor in a building on the USU haviors: follow-wall, turn-left, turn-right, avoid-obstacles, North Campus. The floor has an area of approximately go-thru-doorway, pass-doorway, and make-u-turn. These 4,270 square meters. The floor contains 6 laboratories, two behaviors are coordinated and controlled through Tinber- bathrooms, two staircases, and an elevator. The CS Depart- gen’s release mechanisms[17]. RFID tags are viewed as ment occupies an entire floor in a multi-floor building. The stimuli that trigger or disable specific behaviors. To ensure floor’s area is 6,590 square meters. The floor contains 23 portabilty, all these behaviors are written in the behavior offices, 7 laboratories, a conference room, a student lounge, programming language of the ActivMedia Robotics Inter- a tutor room, two elevators, several bathrooms, and two face for Applications (ARIA) system from ActivMedia staircases. Robotics, Inc. The routines run on the WT laptop. In Forty RFID tags were deployed at the ATL and one hun- addition, the WT laptop runs three other software compo- dred tags were deployed at the CS Department. It took one 1982
  • 5. (a) Narrow (1m wide) Hallway Runs (b) Medium (1.5m wide) Hallway Runs (c) Wide (2.5m wide) Hallway Runs Fig. 3. Path Deviations in Hallways (a) Narrow (1m wide) Hallway Runs (b) Medium (1.5m wide) Hallway Runs (c) Wide (2.5m wide) Hallway Runs Fig. 4. Velocity Changes in Hallways. person 20 minutes to deploy the tags and about 10 minutes was computed as the average of the distances taken during to remove them at the ATL. The same measurements at the the run. Once the ideal distances were known, we ran the CS Department were 30 and 20 minutes, respectively. As robot three times in each type of hallway. The hallways Figure 1(b) indicates, the tags were placed on small pieces in which the robot ran were different from the hallways of cardboard to insulate them from the walls and were in which the ideal distances were computed. Obstacles, attached to the walls with regular scotch tape. The creation e.g., humans walking by and open doors, were allowed of the connectivity graphs, took one hour at the ATL and during the test runs. Figure 3 gives the distance graphs about 2 hours at the CS Department. One administrator of the three runs compared in each hallway type. The first walked around the areas with a laptop and recorded vertical bars in each graph represent the robot’s width. tag-destination associations and then associated behavior As can be seen from Figure 3(a), there is almost no scripts with tags. deviation from the ideal distance in narrow hallways. Nor RG was first repeatedly tested in the ATL, the smaller is there any oscillation. Figure 3(b) and Figure 3(c) show of the two environments, and then deployed for pilot some insignificant deviations from the ideal distance. The experiments at the USU CS Department. We ran two sets deviations were caused by people walking by and by of pilot experiments. The first set did not involve visually open doors. However, there is no oscillation, i.e., sharp impaired participants. The second set did. In the first set of movements in different directions. In both environments, experiments, we had RG navigate three types of hallways we observed several tag detection failures, particularly in of the CS Department: narrow (1 m), medium (1.5 m) and metallic door frames. However, after we insulated the tags wide (2.5 m) and estimated its navigation in terms of two with small pieces of cardboard (see Figure 1(b)), the tag variables: path deviations and abrupt speed changes. We detection failures stopped. also wanted to test how well RG’s RFID reader detected Figure 4 gives the velocity graphs for each hallway type the tags. (x-axis is time in seconds, y-axis is velocity in mm/sec). To estimate path deviations, in each experiment we first The graphs show that the narrow hallways cause short computed the ideal distance that the robot has to maintain abrupt changes in velocity. This is because in narrow from the right wall in a certain type of hallway (narrow, hallways even a slight disorientation, e.g., 3 degrees, in the medium, and wide). The ideal distance was computed by robot causes changes in velocity because less free space is running the robot in a hallway of that type with all doors detected in the grid. In medium and wide hallways, the closed and no obstacles en route. During the run, the velocity remains mostly smooth. Several speed changes distance read by the laser range finder between the robot occur when the robot passes or navigates through doorways and the right wall was recorded every 50 milliseconds. In or avoids obstacles. recording the distance, the robot orientation was taken into The second set of pilot experiments involved five vi- account from two consecutive readings. The ideal distance sually impaired participants, one participant at a time, 1983
  • 6. over a period of two months. Three participants were Research Initiative (CURI) grant from the State of Utah, completely blind and two participants could perceive only and through a New Faculty Research grant from Utah State light. The participants had no speech impediments, hearing University. problems, or cognitive disabilities. Two participants were dog users and the other three used white canes. The participants were asked to use RG to navigate to three R EFERENCES distinct locations (an office, a lounge, and a bathroom) [1] M. P. LaPlante and D. Carlson, Disability in the United States: at the USU CS Department. All participants were new Prevalence and Causes. Washington, DC: U.S. Department of Ed- to the environment and had to navigate approximately 40 ucation, National Institute of Disability and Rehabilitation Research, 2000. meters to get to all destinations. Thus, in the experiments [2] S. Shoval, J. Borenstein, and Y. Koren, “Mobile Robot Obstacle with visually impaired participants, the robot navigated Avoidance in a Computerized Travel for the Blind,” in IEEE approximately 200 meters. All participants reached their International Conference on Robotics and Automation, San Diego, CA, 1994. destinations without a problem. In their exit interviews, [3] D. Ross and B. Blasch, “Development of a Wearable Computer the participants complained mostly about the human-robot Orientation System,” IEEE Personal and Ubiquitous Computing, interaction aspects of the system. For example, all of them vol. 6, pp. 49–63, 2002. [4] H. Mori and S. Kotani, “Robotic Travel Aid for the Blind: had problems with the speech recognition system[21], [19]. HARUNOBU-6,” in Second European Conference on Disability, The participants especially liked the fact that they did not o Virtual Reality, and Assistive Technology, S¨vde, Sweden, 1998. have to give up their white canes and guide dogs to use [5] I. Horswill, “Polly: A Vision-Based Artificial Agent,” in Pro- ceedings of the 11th Conference of the American Association for RG. Artificial Intelligence (AAAI-93), Washington, DC, July 1993. [6] S. Thrun, M. Bennewitz, W. Burgard, A. B. Cremers, F. Dellaert, V. L IMITATIONS a D. Fox, D. H¨hnel, C. Rosenberg, N. Roby, J. Schutle, and D. Schultz, “Minerva: A Second Generation Mobile Tour-Guide In addition to velocity changes in narrow hallways, RG Robot,” in Proceedings of the IEEE International Conference on has three other limitations. First, the robot cannot create Robotics and Automation (ICRA-99), Antwerp, Belgium, June 1999. a connectivity graph for a given environment once the a [7] W. Burgard, A. Cremers, D. Fox, D. H¨hnel, G. Lakemeyer, D. Schulz, W. Steiner, and S. Thrun, “Experiences with an Inter- RFID tags are deployed. We are currently working on active Museum Tour-Guide Robot,” Artificial Intelligence, no. 114, creating connectivity graphs and behavior scripts in a semi- pp. 3–55, 1999. automatic fashion. Second, the robot cannot detect route [8] G. Kantor and S. Singh, “Preliminary Results in Range-Only Lo- calization and Mapping,” in Proceedings of the IEEE Conference blockages. If the route is blocked, the robot first slows on Robotics and Automation, Washington, DC, May 2002. down to a stop and then starts turning in order to find [9] T. Tsukiyama, “Navigation System for the Mobile Robots using some free space. In this fashion, RG makes a gradual u- RFID Tags,” in Proceedings of the IEEE Conference on Advanced Robotics, Coimbra, Portugal, June-July 2003. turn by looking for the maximum free space around itself. a [10] D. H¨hnel, W. Burgard, D. Fox, K. Fishkin, and M. Philipose, Since RG has no orientation sensor, currently the only way “Mapping and localization with rfid technology,” Intel Research it can detect a detour is by detecting an RFID tag that is Institute, Seattle, WA, Tech. Rep. IRS-TR-03-014, December 2003. [11] V. Kulyukin and M. Blair, “Distributed Tracking and Guidance in not on the path to the current destination. Finally, while Indoor Environments,” in Conference of the Rehabilitation Engi- several visually impaired participants told us that it would neering and Assistive Technology Society of North America (RESNA- be helpful if RG could guide them in and out of elevators, 2003), Atlanta, GA, June 2003. [12] http://www.activmedia.com, ActivMedia Robotic Platforms. Activ- RG cannot negotiate elevators yet. Media Robotics, Inc. [13] J. H. Chuang and N. Ahuja, “An Analytically Tractable Potential VI. C ONCLUSION Field Model of Free Space and its Application in Obstacle Avoid- ance,” IEEE Trans. Sys. Man, Cyb., vol. 5, no. 28, pp. 729–736, In this paper, we showed how Radio Frequency Iden- 1998. tification (RFID) can be used in robot-assisted indoor [14] R. Murphy, Introduction to AI Robotics. Cambridge, MA: The MIT navigation for the visually impaired. We presented a robotic Press, 2000. [15] Y. Koren and J. Borenstein, “Potential Field Methods and their guide for the visually impaired that was deployed and Inherent Limitations for Mobile Robot Navigation,” in Proceedings tested both with and without visually impaired participants of the IEEE Conference on Robotics and Automation, Sacramento, in two indoor environments. The experiments illustrate that CA, April 1991. [16] B. Kupiers, “The Spatial Semantic Hierarchy,” Artificial Intelligence, passive RFID tags can act as reliable stimuli that trigger no. 119, pp. 191–233, 2000. local navigation behaviors to achieve global navigation [17] N. Tinbergen, Animal in its World: Laboratory Experiments and objectives. General Papers. Cambridge, MA: Harvard University Press, 1976. [18] V. Kulyukin, “Towards Hands-Free Human-Robot Interaction through Spoken Dialog,” in AAAI Spring Symposium on Human ACKNOWLEDGMENT Interaction with Autonomous Systems in Complex Environments, The authors would like to thank the visually impaired Palo Alto, CA, March 2003. [19] ——, “Human-Robot Interaction through Gesture-Free Spoken Di- participants for generously volunteering their time for the alogue,” Autonomous Robots,, vol. 16, no. 3, 2004. pilot experiments. The authors would like to thank Marty [20] T. V. Tran, T. Letowski, and K. S. Abouchacra, “Evaluation of Blair, Director of the Utah Assistive Technology Program Acoustic Beacon Characteristics for Navigation Tasks,” Ergonomics, vol. 43, no. 6, pp. 807–827, 2000. for his administrative assistance and support. The first [21] V. Kulyukin, C. Gharpure, and N. De Graw, “Human-Robot In- author would like to acknowledge that this research has teraction in a Robotic Guide for the Visually Impaired,” in AAAI been supported, in part, through the NSF Universal Ac- Spring Symposium on Interaction between Humans and Autonomous Systems over Extended Operation, Palo Alto, CA, March 2004. cess Career Grant (IIS-0346880), a Community University 1984