SlideShare a Scribd company logo
1 of 6
Download to read offline
International Journal of Artificial Intelligence & Applications (IJAIA), Vol. 5, No. 1, January 2014

Voice-Controlled Artificial Handspeak System
Jonathan Gatti, Carlo Fonda, Livio Tenze and Enrique Canessa
Science Dissemination Unit (SDU),
The Abdus Salam International Centre for Theoretical Physics (ICTP), Trieste, ITALY

Abstract
A man-machine interaction project is described which aims to establish an automated voice to sign
language translator for communication with the deaf using integrated open technologies. The first
prototype consists of a robotic hand designed with OpenSCAD and manufactured with a low-cost 3D
printer ─which smoothly reproduces the alphabet of the sign language controlled by voice only. The core
automation comprises an Arduino UNO controller used to activate a set of servo motors that follow
instructions from a Raspberry Pi mini-computer having installed the open source speech recognition engine
Julius. We discuss its features, limitations and possible future developments.

Keywords
Artificial Intelligence, Man-Machine Interaction, Open Technologies, 3D Printing

1. Motivation
Science Dissemination for the disabled is a new research topic with the noble ideal of facilitating
the access to science and education to a minority of scholars by trying to offer them equal
opportunities. There are a few interesting recent initiatives worth to mention including the project
"A Touch of the Universe" of the Astronomical Observatory of Valencia for blind students [1], and
the large action named "Astro vers Tous", which promotes astronomy in different directions for
deaf people in hospitals and jails [2]. According to the World Health Organization [3], a 5% of the
world population has disabling hearing loss (with more than 40 million people in Europe), often
preventing their proper psychological and social development. To be affected by deafness
generally implies the risk of exclusion and illiteracy, thus compromising communication skills
and integration.
There are some rehabilitation techniques aimed to helping individuals affected by deafness,
especially young people and children. These techniques are often expensive and not accessible to
all. This situation becomes more critical in the developing countries [3]. In less severe cases, such
methods involve the use of expensive hearing aids, or in the most extreme cases, the sign
language (SL) is taught. However, this last solution may become inefficient when a person
suffering from this handicap needs, for example, to communicate to a person without the
knowledge of the SL, or need to travel abroad. There is not a universal SL yet [4]. There are
variants such as the popular American (A)SL that depends on its society, culture, idiom, etc.
Therefore, providing the opportunity to have an alternative, automated and configurable system
capable of receiving information from anyone and anywhere would allow deaf people to explore
new spaces of inclusion and participation.
Motivated by these needs and facts, we investigated the possibility for having an automated
translator for communication with the deaf to help reduce language barriers. We have developed a
DOI : 10.5121/ijaia.2014.5108

107
International Journal of Artificial Intelligence & Applications (IJAIA), Vol. 5, No. 1, January 2014

man-machine interaction project which aims to establish an automated voice to sign language
translator for communication with the deaf using integrated open technologies. Our prototype
consists of a robotic hand ─designed with OpenSCAD [5] and manufactured with a low-cost 3D
printer [6], which smoothly reproduces the alphabet of the sign language controlled by voice only.
The core automation comprises an Arduino UNO micro-controller [7] used to activate a set of
servo motors that follow instructions from a Raspberry Pi mini-computer [8] having installed the
open source speech recognition engine Julius [9]. In this paper, we discuss its features, limitations
and development as a mean to support deaf individuals. The hope is to create an innovative
channel of communication based on the integration of different open and affordable technologies.

2. Automated Prototype
Our Voice to SL Translator consists of a robotic hand aiming to satisfy an important fundamental
human need such as face to face communication. As shown in Fig. 1, the robotic hand has been
designed using the OpenSCAD software [5] and is made out of PLA bioplastic by means of a
low-cost 3D printer [6]. This 3D hand is printed with the FDM technology and the fingers are
printed in one shot as single objects, without the need of any screws to ensemble its parts.

Figure 1: Modellng process: OpenSCAD design and 3D printout. Example of modular finger is on the
right.

In this initial study the rotational movements needed to mimic the SL have been obtained using
six servo motors (torque 9 kg*cm) for a precise control of angular motion and velocity of the
fingers, as well as of the whole hand. The moving parts were linked to the servo motors by tiny
nylon fish-lines, hidden along the arm. Instructions were sent to the servo motors using an
internal Arduino UNO controller board and drivers [7]. Thanks to the open source Julius voice
recognition framework [9], installed on a Raspberry Pi lightweight desktop computer [8], we
were able to convert the voice input (received via a USB microphone) into text. And precise
instructions were given (via the second USB port of Raspberry Pi) to the Arduino controlling the
six servo motors, to position them at specific angles so that the fingers could move in
correspondence to the single number or letter being pronounced (see our YouTube video example
in [10]). The integrated prototype we envisioned is schematically shown in Fig. 2.
In this way we translated the human voice into actual signs using a 3D printed hand integrated
with an electronic board able to command its fingers by voice inputs. In fact, thanks to the open
108
International Journal of Artificial Intelligence & Applications (IJAIA), Vol. 5, No. 1, January 2014

source software Julius (easily implemented into a Raspberry Pi [11]), we are able to translate
reasonably well the human voice into text to a high precision for isolated letters and numbers
(more than 90% of success). Due to its small size and low power requirements, the Raspberry Pi
platform is adequate for the Julius open source software to execute. However, its small processor
in some cases can slow down the speech recognition. Julius is a Large-Vocabulary Speech
Recognition engine (LVCSR) which needs a language (vocabulary) and an acoustic model
(grammars), which have to be loaded manually into the system.

Figure 2: Diagram of the handspeak system based on open technologies (see video example in [10]).

An Arduino UNO controller is set to command six servo motors following instructions received
by an USB connected Raspberry Pi mini-computer with the speech recognition decoder Julius.
Peripherals also include a USB microphone and a RCA composite video.
Since Julius is not able to translate voice directly into a visual language, we wrote two Clanguage programs: julius4hand.c to drive the servo motors attached to the Arduino UNO
controller and julius4screen.c to display a still image on a small video screen.
The Julius application provides speech recognition results via the standard output (STDO). The
julius4hand.c code reads from the standard input (STDIN) and reacts to the recognized letters and
numbers by sending the proper command to the Arduino ONE board. To achieve this, we created
a generic transmission protocol to interact with Arduino. The connection between the Raspberry
PI and the Arduino board consisted in a serial channel which can be easily driven by using the
standard Linux system call. Julius4hand.c reads an external configuration file where the mapping
between the recognized letters and numbers and their associated positions of servo motors are
stored.
On the other hand, our julius4screen.c code waits on the STDIN for a recognized character, sent
via the STDO by Julius, and calls the external tool fbi (which allows to display images directly on
109
International Journal of Artificial Intelligence & Applications (IJAIA), Vol. 5, No. 1, January 2014

the Linux frame buffer) to show up the correspondent image in SL. In this case the mapping
configuration file contains the recognized characters and their associated images (.jpg) to be
shown.
Finally, we developed the C code to manage the six attached servo motors, connected to the
Arduino UNO board by exploiting the Servo.h library [7], provided by the standard Arduino
framework. The code reads inputs from the USB serial connection and sets the required servos
angles to move the hand to the required position. Servo motors are driven by using a PWM (Pulse
Width Modulation) protocol.

3. Protoype Features and Development
To some extent, our simple, low-cost system is capable of reproducing most of the SL alphabet
following the voice of an interlocutor as shown in Fig. 3 or see our video in [10]. It does not need
an internet connection, neither the use of smartphones or fast computers, which makes it
completely independent and easy to use especially in rural areas where connectivity is poor or
absent. The plastic robotic hand in Figs. 1 and 3 is easily replicated with low-cost 3D printers.
The whole integrated system consumes very little energy.
Our open device is different from alternative approaches designed for those people with hearing
problems. In fact, among the available applications, there are those which can translate voice into
text like Telecommunications Device for the Deaf (TDD). Alternative smarthphone Apps include,
e.g., the finger App, which can play SL and can recreate gestures of one hand by typed text.
Important to notice is that smartphones Apps can make massive use of the battery for long talks.
A non-hearing person may also communicate by cell phones via SMS texting. There are also
special systems, called Auditory Science Interact (AS), that transmit written text in streaming to
the interlocutor (for a recent review see, e.g,. [12]). However, so far as we know, there does not
exist an application that can translate the voice directly into SL as done with our first prototype.
Vocally controlled robot arms, e.g, the earlier VoiceBot, have been proposed to follow voice
commands limited in their movements and actions, such as opening and closing fingers, holding
objects or moving them from an initial position to a target [11,13-17]. Such robots do not have
any specific purpose for the deaf people.

Figure 3: SL sequence for numbers 1,2,3 and 5 following the voice of an interlocutor (see video in [10]).

Robotic devices in general can also be useful for teaching purposes. By constructing a prototype,
110
International Journal of Artificial Intelligence & Applications (IJAIA), Vol. 5, No. 1, January 2014

many non-experts could use it to communicate with deaf people. Our device, in principle, could
be able to reproduce any spoken language by properly configuring Julius. It could help to
illustrate the alphabet and all the movements which are necessary to learn the SL, especially to
children in schools, hospitals and within specialized associations. Our prototype may offer an
interactive and fun way to learn SL by cognitive means, especially in the developing world where
there are not enough facilities dedicated to the disabled.
The rapid expansion and the exponential improvement of the quality of manufacturing guarantee
the possibility to offer customized systems on-demand in line with the modern "Long Tail"
distribution model of, e.g., Amazon.com. The plastic provides long durability and, if a piece of
the printed hand eventually breaks, it can be easily reproduced by a low-cost 3D printer [6]. The
authors intend the designs and codes of this research to be the first in a series of automated voicecontrolled handspeak prototypes, released on the public domain [10].

4. Limitations
For single letters and numbers, the speed reaction of the fingers to the input voice can be
increased with future improvements in hardware and software, and the servo motor activation
sequences can be better implemented by a fine tuning configuration of the rotation angles (from 0
to 180 degree). The reproduction of more complex signs, such as the letters Z, J, requires larger
and complex rotational and translational movements. The use of a USB directional microphone
with an on/off switch can help to reduce any external noise. One major limitation of the current
prototype is its inability to reproduce a speech (to express an ongoing action or describe a specific
meaning). Still, a set of hand signs can spell out a word. The problem of the voice-controlled
representation of concepts by a 3D robotic hand can, in principle, be solved by using an
alternative PC/TV screen as also represented in Fig. 2 to display still images or even short videos
associated to the speaker input.
We believe that with further research into voice-controlled robotic hands that mimics and
reproduce the sign language (let it be 3D printed or, even displayed by sequences of images or
videos on a small screen), a more advance communication tool could be created to allow deaf
individuals to listen to anyone, anywhere, and thus facilitate their integration into society. The
community working on open technologies is making fast progress to improve the performance,
efficiency and durability of the compact Arduino UNO [7] and Raspberry PI [8] hardware.
Our first prototype device, even if incomplete and still in its infancy, sets for itself the future goal
of following the voice of any speaker/lecturer and translating its speech into a particular SL
alphabet configured for a specific country. Looking forward, we hope that one day an automated
voice-controlled handspeak system will help to improve the quality of life and demolish the walls
which separate the main population to those who have hearing problems and are isolated from
access to, e.g., scientific knowledge.

Conclusions
In this work we have introduced, and evaluated the feasibility of a voice to SL translator for
letters and numbers with affordable open technologies (i.e., 3D printing with OpenSCAD
modelling, Arduino ONE controller and Raspberry Pi mini-computer). As far as we know, our
results are the first instance of a voice-controlled robotic hand specially built with the aim to help
deaf individuals. Our initial results demonstrate that such a possibility seems to be reasonable and
feasible. To translate automatically a spoken language into a specific SL is certainly the most
challenging target since SL is mainly a visual language also incorporating complex gestures,
facial expressions, head movements and body language -work along these lines is in progress to
111
International Journal of Artificial Intelligence & Applications (IJAIA), Vol. 5, No. 1, January 2014

convey a lot of information, e.g., within a few video frames.

Acknowledgements
Sincere thanks go to B. Nava and M. Zennaro (ICTP-ICT/4D) for their important contributions
during the development of this project. Prof. F. Quevedo (ICTP Director) is also acknowledged
for continuous support. One of us, Jonathan Gatti, participated in this work during his Internship
at ICTP-SDU, in the Summer of 2013, assisted by Gaia Fior, collaborator of ICTP-SDU.

References
[1]
[2]
[3]
[4]
[5]
[6]
[7]
[8]
[9]
[10]
[11]
[12]
[13]
[14]
[15]

[16]
[17]

UNAWE non-profit project "A Touch of the Universe" at http://astrokit.uv.es
D. Proust in "Astronomy in Sign Languge" at http://www.youtube.com/watch?v=ksFhan9IKKU
World Health Organization (WHO), 2013 Deafness and Hearing Loss Statistics, at
http://www.who.int/mediacentre/factsheets/fs300/en/
Wikipedia
on
"Sign
Languages
relation
with
Spoken
Lnguages"
at
http://en.wikipedia.org/wiki/Sign_language
OpenSCAD – The Programmers Solid 3D CAD Modeller at http://www.openscad.org
E. Canessa, C. Fonda and M. Zennaro Eds., open book in “Low-cost 3D Printing for Science,
Education and Sustainable Development”, Trieste, ICTP 2013, ISBN 9295003489
The Arduino Uno microcontroller board a http://www.arduino.cc
Servo.h library available at
http://arduino.cc/en/reference/servo
Raspberry Pi credit-card sized computer at http://www.raspberrypi.org
Julius speech recognition decoder developed by CSRC, Japan at http://julius.sourceforge.jp
ICTP-SDU voice2hand project at http://sdu.ictp.it/3D/voice2hand.html @sciencedissemination
YouTube video "Voice to Sign Language System" at http://www.youtube.com/watch?v=J5whsEsGr4s
YouTube
video:
“Speech
Recognition
using
the
Raspberry
Pi”
at
http://www.youtube.com/watch?v=_U_v9tRD68k
Top 10 iPhone Apps for People Who Are Deaf or Hard of Hearing at http://atcoalition.org/news/top10-iphone-apps-people-who-are-deaf-or-hard-hearing
Free, open-source 3D printing robotic hands are available at http://www.thingiverse.com
H. Dang and P. Allen, "Stable Grasping under Pose Uncertainty using Tactile Feedback", Autonomous
Robots, August 2013.
R.R. Ma, L.U. Odhner and A.M. Dollar, "A Modular, Open-Source 3D Printed Underactuated Hand",
Proceedings IEEE International Conference on Robotics and Automation (ICRA) Karlsruhe,
Germany, May 6-10, 2013, pp. 2722-2728.
B. Leon, A. Morales and J. Sancho-Bru, "The Model of the Human Hand", Cognitive System
Monographs 19 (2014) 123-173.
P.S. Ramaiah , M. Venkateswara Rao , G.V. Satyanarayana, "A Microcontroller Based Four Fingered
Robotic Hand", Int. J. Art. Int. & App. (IJAIA) 2 (2011) 90-102.

112

More Related Content

What's hot

Simputer new ppt
Simputer new pptSimputer new ppt
Simputer new pptRama B
 
Sign language translator ieee power point
Sign language translator ieee power pointSign language translator ieee power point
Sign language translator ieee power pointMadhuri Yellapu
 
Enable talk project
Enable talk projectEnable talk project
Enable talk projectslmnsvn
 
Silent Sound Technology
Silent Sound TechnologySilent Sound Technology
Silent Sound TechnologyHafiz Sanni
 
Smart glove hand gesture vocalizer final year report
Smart glove hand gesture vocalizer final year reportSmart glove hand gesture vocalizer final year report
Smart glove hand gesture vocalizer final year reportMUHAMMAD NOMAN KHAN
 
Sign Language Translator
Sign Language TranslatorSign Language Translator
Sign Language TranslatorManjari Raj
 
Alter ego - PPT
Alter ego - PPT Alter ego - PPT
Alter ego - PPT Shirazjaf
 
Silent sound interface
Silent sound interfaceSilent sound interface
Silent sound interfaceJeevitha Reddy
 
Silentsound documentation
Silentsound documentationSilentsound documentation
Silentsound documentationRaj Niranjan
 
Conversion of sign language to speech using kinect
Conversion of sign language to speech using kinectConversion of sign language to speech using kinect
Conversion of sign language to speech using kinectrajaganapathy28091100
 
silent sound technology pdf
silent sound technology pdfsilent sound technology pdf
silent sound technology pdfrahul mishra
 
Gesture Based Communication System
Gesture Based Communication SystemGesture Based Communication System
Gesture Based Communication Systemijtsrd
 
IRJET- Virtual Vision for Blinds
IRJET- Virtual Vision for BlindsIRJET- Virtual Vision for Blinds
IRJET- Virtual Vision for BlindsIRJET Journal
 
Silent sound technology
Silent sound technologySilent sound technology
Silent sound technologyMaria Dominica
 
Deaf and Dump Gesture Recognition System
Deaf and Dump Gesture Recognition SystemDeaf and Dump Gesture Recognition System
Deaf and Dump Gesture Recognition SystemPraveena T
 
IRJET - Sign Language Recognition System
IRJET -  	  Sign Language Recognition SystemIRJET -  	  Sign Language Recognition System
IRJET - Sign Language Recognition SystemIRJET Journal
 

What's hot (20)

Simputer new ppt
Simputer new pptSimputer new ppt
Simputer new ppt
 
Smart glove
Smart gloveSmart glove
Smart glove
 
Sign language translator ieee power point
Sign language translator ieee power pointSign language translator ieee power point
Sign language translator ieee power point
 
Enable talk project
Enable talk projectEnable talk project
Enable talk project
 
AlterEgo Device PPT
AlterEgo Device PPTAlterEgo Device PPT
AlterEgo Device PPT
 
Silent Sound Technology
Silent Sound TechnologySilent Sound Technology
Silent Sound Technology
 
Smart glove hand gesture vocalizer final year report
Smart glove hand gesture vocalizer final year reportSmart glove hand gesture vocalizer final year report
Smart glove hand gesture vocalizer final year report
 
Sign Language Translator
Sign Language TranslatorSign Language Translator
Sign Language Translator
 
Alter ego - PPT
Alter ego - PPT Alter ego - PPT
Alter ego - PPT
 
Silent sound interface
Silent sound interfaceSilent sound interface
Silent sound interface
 
Silentsound documentation
Silentsound documentationSilentsound documentation
Silentsound documentation
 
Conversion of sign language to speech using kinect
Conversion of sign language to speech using kinectConversion of sign language to speech using kinect
Conversion of sign language to speech using kinect
 
silent sound technology pdf
silent sound technology pdfsilent sound technology pdf
silent sound technology pdf
 
Gesture Based Communication System
Gesture Based Communication SystemGesture Based Communication System
Gesture Based Communication System
 
Human computer interaction
Human computer interactionHuman computer interaction
Human computer interaction
 
IRJET- Virtual Vision for Blinds
IRJET- Virtual Vision for BlindsIRJET- Virtual Vision for Blinds
IRJET- Virtual Vision for Blinds
 
Silent sound technology
Silent sound technologySilent sound technology
Silent sound technology
 
Deaf and Dump Gesture Recognition System
Deaf and Dump Gesture Recognition SystemDeaf and Dump Gesture Recognition System
Deaf and Dump Gesture Recognition System
 
5 pen technology documentation
5 pen technology documentation5 pen technology documentation
5 pen technology documentation
 
IRJET - Sign Language Recognition System
IRJET -  	  Sign Language Recognition SystemIRJET -  	  Sign Language Recognition System
IRJET - Sign Language Recognition System
 

Viewers also liked

speech recognition calculater
speech recognition calculaterspeech recognition calculater
speech recognition calculateranil u
 
Roboticsprogramming101
Roboticsprogramming101Roboticsprogramming101
Roboticsprogramming101OSCON Byrum
 
Final report on gui based device control using matlab
Final  report  on  gui based device control using matlabFinal  report  on  gui based device control using matlab
Final report on gui based device control using matlabPalak Sinha
 
gui based device controller using matlab Major presentation
gui based device controller using matlab Major presentationgui based device controller using matlab Major presentation
gui based device controller using matlab Major presentationPalak Sinha
 
The Near Future of CSS
The Near Future of CSSThe Near Future of CSS
The Near Future of CSSRachel Andrew
 
Classroom Management Tips for Kids and Adolescents
Classroom Management Tips for Kids and AdolescentsClassroom Management Tips for Kids and Adolescents
Classroom Management Tips for Kids and AdolescentsShelly Sanchez Terrell
 
The Buyer's Journey - by Chris Lema
The Buyer's Journey - by Chris LemaThe Buyer's Journey - by Chris Lema
The Buyer's Journey - by Chris LemaChris Lema
 
The Presentation Come-Back Kid
The Presentation Come-Back KidThe Presentation Come-Back Kid
The Presentation Come-Back KidEthos3
 

Viewers also liked (8)

speech recognition calculater
speech recognition calculaterspeech recognition calculater
speech recognition calculater
 
Roboticsprogramming101
Roboticsprogramming101Roboticsprogramming101
Roboticsprogramming101
 
Final report on gui based device control using matlab
Final  report  on  gui based device control using matlabFinal  report  on  gui based device control using matlab
Final report on gui based device control using matlab
 
gui based device controller using matlab Major presentation
gui based device controller using matlab Major presentationgui based device controller using matlab Major presentation
gui based device controller using matlab Major presentation
 
The Near Future of CSS
The Near Future of CSSThe Near Future of CSS
The Near Future of CSS
 
Classroom Management Tips for Kids and Adolescents
Classroom Management Tips for Kids and AdolescentsClassroom Management Tips for Kids and Adolescents
Classroom Management Tips for Kids and Adolescents
 
The Buyer's Journey - by Chris Lema
The Buyer's Journey - by Chris LemaThe Buyer's Journey - by Chris Lema
The Buyer's Journey - by Chris Lema
 
The Presentation Come-Back Kid
The Presentation Come-Back KidThe Presentation Come-Back Kid
The Presentation Come-Back Kid
 

Similar to Voice controlled artificial handspeak system

Human Computer Interface Glove for Sign Language Translation
Human Computer Interface Glove for Sign Language TranslationHuman Computer Interface Glove for Sign Language Translation
Human Computer Interface Glove for Sign Language TranslationPARNIKA GUPTA
 
An Embedded Module as “VIRTUAL TONGUE” for Voiceless
An Embedded Module as “VIRTUAL TONGUE” for VoicelessAn Embedded Module as “VIRTUAL TONGUE” for Voiceless
An Embedded Module as “VIRTUAL TONGUE” for Voicelessijistjournal
 
Wake-up-word speech recognition using GPS on smart phone
Wake-up-word speech recognition using GPS on smart phoneWake-up-word speech recognition using GPS on smart phone
Wake-up-word speech recognition using GPS on smart phoneIJERA Editor
 
A Translation Device for the Vision Based Sign Language
A Translation Device for the Vision Based Sign LanguageA Translation Device for the Vision Based Sign Language
A Translation Device for the Vision Based Sign Languageijsrd.com
 
DESIGN OF LOW COST AND EFFICIENT SIGN LANGUAGE INTERPRETER FOR THE SPEECH AND...
DESIGN OF LOW COST AND EFFICIENT SIGN LANGUAGE INTERPRETER FOR THE SPEECH AND...DESIGN OF LOW COST AND EFFICIENT SIGN LANGUAGE INTERPRETER FOR THE SPEECH AND...
DESIGN OF LOW COST AND EFFICIENT SIGN LANGUAGE INTERPRETER FOR THE SPEECH AND...Dr.SHANTHI K.G
 
Survey Paper on Raspberry pi based Assistive Device for Communication between...
Survey Paper on Raspberry pi based Assistive Device for Communication between...Survey Paper on Raspberry pi based Assistive Device for Communication between...
Survey Paper on Raspberry pi based Assistive Device for Communication between...IRJET Journal
 
Human computerinterface
Human computerinterfaceHuman computerinterface
Human computerinterfaceKumar Aryan
 
5.smart multilingual sign boards
5.smart multilingual sign boards5.smart multilingual sign boards
5.smart multilingual sign boardsEditorJST
 
Phase1 presentation1FINAL.pptx
Phase1 presentation1FINAL.pptxPhase1 presentation1FINAL.pptx
Phase1 presentation1FINAL.pptxPoornimak52
 
Design of a Communication System using Sign Language aid for Differently Able...
Design of a Communication System using Sign Language aid for Differently Able...Design of a Communication System using Sign Language aid for Differently Able...
Design of a Communication System using Sign Language aid for Differently Able...IRJET Journal
 
THE THIRD EYE-Presentation
THE THIRD EYE-PresentationTHE THIRD EYE-Presentation
THE THIRD EYE-PresentationRomil Shah
 
AlterEgo: A Personalized Wearable Silent Speech Interface
AlterEgo: A Personalized Wearable Silent Speech InterfaceAlterEgo: A Personalized Wearable Silent Speech Interface
AlterEgo: A Personalized Wearable Silent Speech Interfaceeraser Juan José Calderón
 
Virtual Personal Assistant
Virtual Personal AssistantVirtual Personal Assistant
Virtual Personal AssistantIRJET Journal
 
International Journal of Engineering Research and Development
International Journal of Engineering Research and DevelopmentInternational Journal of Engineering Research and Development
International Journal of Engineering Research and DevelopmentIJERD Editor
 
IRJET- IoT based Portable Hand Gesture Recognition System
IRJET-  	  IoT based Portable Hand Gesture Recognition SystemIRJET-  	  IoT based Portable Hand Gesture Recognition System
IRJET- IoT based Portable Hand Gesture Recognition SystemIRJET Journal
 
IRJET - Sign Language to Speech Conversion Gloves using Arduino and Flex Sens...
IRJET - Sign Language to Speech Conversion Gloves using Arduino and Flex Sens...IRJET - Sign Language to Speech Conversion Gloves using Arduino and Flex Sens...
IRJET - Sign Language to Speech Conversion Gloves using Arduino and Flex Sens...IRJET Journal
 

Similar to Voice controlled artificial handspeak system (20)

Human Computer Interface Glove for Sign Language Translation
Human Computer Interface Glove for Sign Language TranslationHuman Computer Interface Glove for Sign Language Translation
Human Computer Interface Glove for Sign Language Translation
 
An Embedded Module as “VIRTUAL TONGUE” for Voiceless
An Embedded Module as “VIRTUAL TONGUE” for VoicelessAn Embedded Module as “VIRTUAL TONGUE” for Voiceless
An Embedded Module as “VIRTUAL TONGUE” for Voiceless
 
Wake-up-word speech recognition using GPS on smart phone
Wake-up-word speech recognition using GPS on smart phoneWake-up-word speech recognition using GPS on smart phone
Wake-up-word speech recognition using GPS on smart phone
 
A Translation Device for the Vision Based Sign Language
A Translation Device for the Vision Based Sign LanguageA Translation Device for the Vision Based Sign Language
A Translation Device for the Vision Based Sign Language
 
DESIGN OF LOW COST AND EFFICIENT SIGN LANGUAGE INTERPRETER FOR THE SPEECH AND...
DESIGN OF LOW COST AND EFFICIENT SIGN LANGUAGE INTERPRETER FOR THE SPEECH AND...DESIGN OF LOW COST AND EFFICIENT SIGN LANGUAGE INTERPRETER FOR THE SPEECH AND...
DESIGN OF LOW COST AND EFFICIENT SIGN LANGUAGE INTERPRETER FOR THE SPEECH AND...
 
Survey Paper on Raspberry pi based Assistive Device for Communication between...
Survey Paper on Raspberry pi based Assistive Device for Communication between...Survey Paper on Raspberry pi based Assistive Device for Communication between...
Survey Paper on Raspberry pi based Assistive Device for Communication between...
 
delna's journal
delna's journaldelna's journal
delna's journal
 
Human computerinterface
Human computerinterfaceHuman computerinterface
Human computerinterface
 
5.smart multilingual sign boards
5.smart multilingual sign boards5.smart multilingual sign boards
5.smart multilingual sign boards
 
Phase1 presentation1FINAL.pptx
Phase1 presentation1FINAL.pptxPhase1 presentation1FINAL.pptx
Phase1 presentation1FINAL.pptx
 
Bt35408413
Bt35408413Bt35408413
Bt35408413
 
Design of a Communication System using Sign Language aid for Differently Able...
Design of a Communication System using Sign Language aid for Differently Able...Design of a Communication System using Sign Language aid for Differently Able...
Design of a Communication System using Sign Language aid for Differently Able...
 
THE THIRD EYE-Presentation
THE THIRD EYE-PresentationTHE THIRD EYE-Presentation
THE THIRD EYE-Presentation
 
AlterEgo: A Personalized Wearable Silent Speech Interface
AlterEgo: A Personalized Wearable Silent Speech InterfaceAlterEgo: A Personalized Wearable Silent Speech Interface
AlterEgo: A Personalized Wearable Silent Speech Interface
 
Virtual Personal Assistant
Virtual Personal AssistantVirtual Personal Assistant
Virtual Personal Assistant
 
International Journal of Engineering Research and Development
International Journal of Engineering Research and DevelopmentInternational Journal of Engineering Research and Development
International Journal of Engineering Research and Development
 
ch03.ppt
ch03.pptch03.ppt
ch03.ppt
 
IRJET- IoT based Portable Hand Gesture Recognition System
IRJET-  	  IoT based Portable Hand Gesture Recognition SystemIRJET-  	  IoT based Portable Hand Gesture Recognition System
IRJET- IoT based Portable Hand Gesture Recognition System
 
IRJET - Sign Language to Speech Conversion Gloves using Arduino and Flex Sens...
IRJET - Sign Language to Speech Conversion Gloves using Arduino and Flex Sens...IRJET - Sign Language to Speech Conversion Gloves using Arduino and Flex Sens...
IRJET - Sign Language to Speech Conversion Gloves using Arduino and Flex Sens...
 
ppt project pk.pptx
ppt project pk.pptxppt project pk.pptx
ppt project pk.pptx
 

Recently uploaded

Digital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxDigital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxLoriGlavin3
 
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024BookNet Canada
 
Testing tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examplesTesting tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examplesKari Kakkonen
 
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentEmixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentPim van der Noll
 
Generative Artificial Intelligence: How generative AI works.pdf
Generative Artificial Intelligence: How generative AI works.pdfGenerative Artificial Intelligence: How generative AI works.pdf
Generative Artificial Intelligence: How generative AI works.pdfIngrid Airi González
 
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxA Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxLoriGlavin3
 
Generative AI - Gitex v1Generative AI - Gitex v1.pptx
Generative AI - Gitex v1Generative AI - Gitex v1.pptxGenerative AI - Gitex v1Generative AI - Gitex v1.pptx
Generative AI - Gitex v1Generative AI - Gitex v1.pptxfnnc6jmgwh
 
Decarbonising Buildings: Making a net-zero built environment a reality
Decarbonising Buildings: Making a net-zero built environment a realityDecarbonising Buildings: Making a net-zero built environment a reality
Decarbonising Buildings: Making a net-zero built environment a realityIES VE
 
TrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data PrivacyTrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data PrivacyTrustArc
 
The State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxThe State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxLoriGlavin3
 
TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024Lonnie McRorey
 
[Webinar] SpiraTest - Setting New Standards in Quality Assurance
[Webinar] SpiraTest - Setting New Standards in Quality Assurance[Webinar] SpiraTest - Setting New Standards in Quality Assurance
[Webinar] SpiraTest - Setting New Standards in Quality AssuranceInflectra
 
2024 April Patch Tuesday
2024 April Patch Tuesday2024 April Patch Tuesday
2024 April Patch TuesdayIvanti
 
Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...
Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...
Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...Nikki Chapple
 
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxPasskey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxLoriGlavin3
 
Glenn Lazarus- Why Your Observability Strategy Needs Security Observability
Glenn Lazarus- Why Your Observability Strategy Needs Security ObservabilityGlenn Lazarus- Why Your Observability Strategy Needs Security Observability
Glenn Lazarus- Why Your Observability Strategy Needs Security Observabilityitnewsafrica
 
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptxUse of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptxLoriGlavin3
 
React Native vs Ionic - The Best Mobile App Framework
React Native vs Ionic - The Best Mobile App FrameworkReact Native vs Ionic - The Best Mobile App Framework
React Native vs Ionic - The Best Mobile App FrameworkPixlogix Infotech
 
Bridging Between CAD & GIS: 6 Ways to Automate Your Data Integration
Bridging Between CAD & GIS:  6 Ways to Automate Your Data IntegrationBridging Between CAD & GIS:  6 Ways to Automate Your Data Integration
Bridging Between CAD & GIS: 6 Ways to Automate Your Data Integrationmarketing932765
 
So einfach geht modernes Roaming fuer Notes und Nomad.pdf
So einfach geht modernes Roaming fuer Notes und Nomad.pdfSo einfach geht modernes Roaming fuer Notes und Nomad.pdf
So einfach geht modernes Roaming fuer Notes und Nomad.pdfpanagenda
 

Recently uploaded (20)

Digital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxDigital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
 
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
 
Testing tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examplesTesting tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examples
 
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentEmixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
 
Generative Artificial Intelligence: How generative AI works.pdf
Generative Artificial Intelligence: How generative AI works.pdfGenerative Artificial Intelligence: How generative AI works.pdf
Generative Artificial Intelligence: How generative AI works.pdf
 
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxA Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
 
Generative AI - Gitex v1Generative AI - Gitex v1.pptx
Generative AI - Gitex v1Generative AI - Gitex v1.pptxGenerative AI - Gitex v1Generative AI - Gitex v1.pptx
Generative AI - Gitex v1Generative AI - Gitex v1.pptx
 
Decarbonising Buildings: Making a net-zero built environment a reality
Decarbonising Buildings: Making a net-zero built environment a realityDecarbonising Buildings: Making a net-zero built environment a reality
Decarbonising Buildings: Making a net-zero built environment a reality
 
TrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data PrivacyTrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data Privacy
 
The State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxThe State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptx
 
TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024
 
[Webinar] SpiraTest - Setting New Standards in Quality Assurance
[Webinar] SpiraTest - Setting New Standards in Quality Assurance[Webinar] SpiraTest - Setting New Standards in Quality Assurance
[Webinar] SpiraTest - Setting New Standards in Quality Assurance
 
2024 April Patch Tuesday
2024 April Patch Tuesday2024 April Patch Tuesday
2024 April Patch Tuesday
 
Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...
Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...
Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...
 
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxPasskey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
 
Glenn Lazarus- Why Your Observability Strategy Needs Security Observability
Glenn Lazarus- Why Your Observability Strategy Needs Security ObservabilityGlenn Lazarus- Why Your Observability Strategy Needs Security Observability
Glenn Lazarus- Why Your Observability Strategy Needs Security Observability
 
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptxUse of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
 
React Native vs Ionic - The Best Mobile App Framework
React Native vs Ionic - The Best Mobile App FrameworkReact Native vs Ionic - The Best Mobile App Framework
React Native vs Ionic - The Best Mobile App Framework
 
Bridging Between CAD & GIS: 6 Ways to Automate Your Data Integration
Bridging Between CAD & GIS:  6 Ways to Automate Your Data IntegrationBridging Between CAD & GIS:  6 Ways to Automate Your Data Integration
Bridging Between CAD & GIS: 6 Ways to Automate Your Data Integration
 
So einfach geht modernes Roaming fuer Notes und Nomad.pdf
So einfach geht modernes Roaming fuer Notes und Nomad.pdfSo einfach geht modernes Roaming fuer Notes und Nomad.pdf
So einfach geht modernes Roaming fuer Notes und Nomad.pdf
 

Voice controlled artificial handspeak system

  • 1. International Journal of Artificial Intelligence & Applications (IJAIA), Vol. 5, No. 1, January 2014 Voice-Controlled Artificial Handspeak System Jonathan Gatti, Carlo Fonda, Livio Tenze and Enrique Canessa Science Dissemination Unit (SDU), The Abdus Salam International Centre for Theoretical Physics (ICTP), Trieste, ITALY Abstract A man-machine interaction project is described which aims to establish an automated voice to sign language translator for communication with the deaf using integrated open technologies. The first prototype consists of a robotic hand designed with OpenSCAD and manufactured with a low-cost 3D printer ─which smoothly reproduces the alphabet of the sign language controlled by voice only. The core automation comprises an Arduino UNO controller used to activate a set of servo motors that follow instructions from a Raspberry Pi mini-computer having installed the open source speech recognition engine Julius. We discuss its features, limitations and possible future developments. Keywords Artificial Intelligence, Man-Machine Interaction, Open Technologies, 3D Printing 1. Motivation Science Dissemination for the disabled is a new research topic with the noble ideal of facilitating the access to science and education to a minority of scholars by trying to offer them equal opportunities. There are a few interesting recent initiatives worth to mention including the project "A Touch of the Universe" of the Astronomical Observatory of Valencia for blind students [1], and the large action named "Astro vers Tous", which promotes astronomy in different directions for deaf people in hospitals and jails [2]. According to the World Health Organization [3], a 5% of the world population has disabling hearing loss (with more than 40 million people in Europe), often preventing their proper psychological and social development. To be affected by deafness generally implies the risk of exclusion and illiteracy, thus compromising communication skills and integration. There are some rehabilitation techniques aimed to helping individuals affected by deafness, especially young people and children. These techniques are often expensive and not accessible to all. This situation becomes more critical in the developing countries [3]. In less severe cases, such methods involve the use of expensive hearing aids, or in the most extreme cases, the sign language (SL) is taught. However, this last solution may become inefficient when a person suffering from this handicap needs, for example, to communicate to a person without the knowledge of the SL, or need to travel abroad. There is not a universal SL yet [4]. There are variants such as the popular American (A)SL that depends on its society, culture, idiom, etc. Therefore, providing the opportunity to have an alternative, automated and configurable system capable of receiving information from anyone and anywhere would allow deaf people to explore new spaces of inclusion and participation. Motivated by these needs and facts, we investigated the possibility for having an automated translator for communication with the deaf to help reduce language barriers. We have developed a DOI : 10.5121/ijaia.2014.5108 107
  • 2. International Journal of Artificial Intelligence & Applications (IJAIA), Vol. 5, No. 1, January 2014 man-machine interaction project which aims to establish an automated voice to sign language translator for communication with the deaf using integrated open technologies. Our prototype consists of a robotic hand ─designed with OpenSCAD [5] and manufactured with a low-cost 3D printer [6], which smoothly reproduces the alphabet of the sign language controlled by voice only. The core automation comprises an Arduino UNO micro-controller [7] used to activate a set of servo motors that follow instructions from a Raspberry Pi mini-computer [8] having installed the open source speech recognition engine Julius [9]. In this paper, we discuss its features, limitations and development as a mean to support deaf individuals. The hope is to create an innovative channel of communication based on the integration of different open and affordable technologies. 2. Automated Prototype Our Voice to SL Translator consists of a robotic hand aiming to satisfy an important fundamental human need such as face to face communication. As shown in Fig. 1, the robotic hand has been designed using the OpenSCAD software [5] and is made out of PLA bioplastic by means of a low-cost 3D printer [6]. This 3D hand is printed with the FDM technology and the fingers are printed in one shot as single objects, without the need of any screws to ensemble its parts. Figure 1: Modellng process: OpenSCAD design and 3D printout. Example of modular finger is on the right. In this initial study the rotational movements needed to mimic the SL have been obtained using six servo motors (torque 9 kg*cm) for a precise control of angular motion and velocity of the fingers, as well as of the whole hand. The moving parts were linked to the servo motors by tiny nylon fish-lines, hidden along the arm. Instructions were sent to the servo motors using an internal Arduino UNO controller board and drivers [7]. Thanks to the open source Julius voice recognition framework [9], installed on a Raspberry Pi lightweight desktop computer [8], we were able to convert the voice input (received via a USB microphone) into text. And precise instructions were given (via the second USB port of Raspberry Pi) to the Arduino controlling the six servo motors, to position them at specific angles so that the fingers could move in correspondence to the single number or letter being pronounced (see our YouTube video example in [10]). The integrated prototype we envisioned is schematically shown in Fig. 2. In this way we translated the human voice into actual signs using a 3D printed hand integrated with an electronic board able to command its fingers by voice inputs. In fact, thanks to the open 108
  • 3. International Journal of Artificial Intelligence & Applications (IJAIA), Vol. 5, No. 1, January 2014 source software Julius (easily implemented into a Raspberry Pi [11]), we are able to translate reasonably well the human voice into text to a high precision for isolated letters and numbers (more than 90% of success). Due to its small size and low power requirements, the Raspberry Pi platform is adequate for the Julius open source software to execute. However, its small processor in some cases can slow down the speech recognition. Julius is a Large-Vocabulary Speech Recognition engine (LVCSR) which needs a language (vocabulary) and an acoustic model (grammars), which have to be loaded manually into the system. Figure 2: Diagram of the handspeak system based on open technologies (see video example in [10]). An Arduino UNO controller is set to command six servo motors following instructions received by an USB connected Raspberry Pi mini-computer with the speech recognition decoder Julius. Peripherals also include a USB microphone and a RCA composite video. Since Julius is not able to translate voice directly into a visual language, we wrote two Clanguage programs: julius4hand.c to drive the servo motors attached to the Arduino UNO controller and julius4screen.c to display a still image on a small video screen. The Julius application provides speech recognition results via the standard output (STDO). The julius4hand.c code reads from the standard input (STDIN) and reacts to the recognized letters and numbers by sending the proper command to the Arduino ONE board. To achieve this, we created a generic transmission protocol to interact with Arduino. The connection between the Raspberry PI and the Arduino board consisted in a serial channel which can be easily driven by using the standard Linux system call. Julius4hand.c reads an external configuration file where the mapping between the recognized letters and numbers and their associated positions of servo motors are stored. On the other hand, our julius4screen.c code waits on the STDIN for a recognized character, sent via the STDO by Julius, and calls the external tool fbi (which allows to display images directly on 109
  • 4. International Journal of Artificial Intelligence & Applications (IJAIA), Vol. 5, No. 1, January 2014 the Linux frame buffer) to show up the correspondent image in SL. In this case the mapping configuration file contains the recognized characters and their associated images (.jpg) to be shown. Finally, we developed the C code to manage the six attached servo motors, connected to the Arduino UNO board by exploiting the Servo.h library [7], provided by the standard Arduino framework. The code reads inputs from the USB serial connection and sets the required servos angles to move the hand to the required position. Servo motors are driven by using a PWM (Pulse Width Modulation) protocol. 3. Protoype Features and Development To some extent, our simple, low-cost system is capable of reproducing most of the SL alphabet following the voice of an interlocutor as shown in Fig. 3 or see our video in [10]. It does not need an internet connection, neither the use of smartphones or fast computers, which makes it completely independent and easy to use especially in rural areas where connectivity is poor or absent. The plastic robotic hand in Figs. 1 and 3 is easily replicated with low-cost 3D printers. The whole integrated system consumes very little energy. Our open device is different from alternative approaches designed for those people with hearing problems. In fact, among the available applications, there are those which can translate voice into text like Telecommunications Device for the Deaf (TDD). Alternative smarthphone Apps include, e.g., the finger App, which can play SL and can recreate gestures of one hand by typed text. Important to notice is that smartphones Apps can make massive use of the battery for long talks. A non-hearing person may also communicate by cell phones via SMS texting. There are also special systems, called Auditory Science Interact (AS), that transmit written text in streaming to the interlocutor (for a recent review see, e.g,. [12]). However, so far as we know, there does not exist an application that can translate the voice directly into SL as done with our first prototype. Vocally controlled robot arms, e.g, the earlier VoiceBot, have been proposed to follow voice commands limited in their movements and actions, such as opening and closing fingers, holding objects or moving them from an initial position to a target [11,13-17]. Such robots do not have any specific purpose for the deaf people. Figure 3: SL sequence for numbers 1,2,3 and 5 following the voice of an interlocutor (see video in [10]). Robotic devices in general can also be useful for teaching purposes. By constructing a prototype, 110
  • 5. International Journal of Artificial Intelligence & Applications (IJAIA), Vol. 5, No. 1, January 2014 many non-experts could use it to communicate with deaf people. Our device, in principle, could be able to reproduce any spoken language by properly configuring Julius. It could help to illustrate the alphabet and all the movements which are necessary to learn the SL, especially to children in schools, hospitals and within specialized associations. Our prototype may offer an interactive and fun way to learn SL by cognitive means, especially in the developing world where there are not enough facilities dedicated to the disabled. The rapid expansion and the exponential improvement of the quality of manufacturing guarantee the possibility to offer customized systems on-demand in line with the modern "Long Tail" distribution model of, e.g., Amazon.com. The plastic provides long durability and, if a piece of the printed hand eventually breaks, it can be easily reproduced by a low-cost 3D printer [6]. The authors intend the designs and codes of this research to be the first in a series of automated voicecontrolled handspeak prototypes, released on the public domain [10]. 4. Limitations For single letters and numbers, the speed reaction of the fingers to the input voice can be increased with future improvements in hardware and software, and the servo motor activation sequences can be better implemented by a fine tuning configuration of the rotation angles (from 0 to 180 degree). The reproduction of more complex signs, such as the letters Z, J, requires larger and complex rotational and translational movements. The use of a USB directional microphone with an on/off switch can help to reduce any external noise. One major limitation of the current prototype is its inability to reproduce a speech (to express an ongoing action or describe a specific meaning). Still, a set of hand signs can spell out a word. The problem of the voice-controlled representation of concepts by a 3D robotic hand can, in principle, be solved by using an alternative PC/TV screen as also represented in Fig. 2 to display still images or even short videos associated to the speaker input. We believe that with further research into voice-controlled robotic hands that mimics and reproduce the sign language (let it be 3D printed or, even displayed by sequences of images or videos on a small screen), a more advance communication tool could be created to allow deaf individuals to listen to anyone, anywhere, and thus facilitate their integration into society. The community working on open technologies is making fast progress to improve the performance, efficiency and durability of the compact Arduino UNO [7] and Raspberry PI [8] hardware. Our first prototype device, even if incomplete and still in its infancy, sets for itself the future goal of following the voice of any speaker/lecturer and translating its speech into a particular SL alphabet configured for a specific country. Looking forward, we hope that one day an automated voice-controlled handspeak system will help to improve the quality of life and demolish the walls which separate the main population to those who have hearing problems and are isolated from access to, e.g., scientific knowledge. Conclusions In this work we have introduced, and evaluated the feasibility of a voice to SL translator for letters and numbers with affordable open technologies (i.e., 3D printing with OpenSCAD modelling, Arduino ONE controller and Raspberry Pi mini-computer). As far as we know, our results are the first instance of a voice-controlled robotic hand specially built with the aim to help deaf individuals. Our initial results demonstrate that such a possibility seems to be reasonable and feasible. To translate automatically a spoken language into a specific SL is certainly the most challenging target since SL is mainly a visual language also incorporating complex gestures, facial expressions, head movements and body language -work along these lines is in progress to 111
  • 6. International Journal of Artificial Intelligence & Applications (IJAIA), Vol. 5, No. 1, January 2014 convey a lot of information, e.g., within a few video frames. Acknowledgements Sincere thanks go to B. Nava and M. Zennaro (ICTP-ICT/4D) for their important contributions during the development of this project. Prof. F. Quevedo (ICTP Director) is also acknowledged for continuous support. One of us, Jonathan Gatti, participated in this work during his Internship at ICTP-SDU, in the Summer of 2013, assisted by Gaia Fior, collaborator of ICTP-SDU. References [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12] [13] [14] [15] [16] [17] UNAWE non-profit project "A Touch of the Universe" at http://astrokit.uv.es D. Proust in "Astronomy in Sign Languge" at http://www.youtube.com/watch?v=ksFhan9IKKU World Health Organization (WHO), 2013 Deafness and Hearing Loss Statistics, at http://www.who.int/mediacentre/factsheets/fs300/en/ Wikipedia on "Sign Languages relation with Spoken Lnguages" at http://en.wikipedia.org/wiki/Sign_language OpenSCAD – The Programmers Solid 3D CAD Modeller at http://www.openscad.org E. Canessa, C. Fonda and M. Zennaro Eds., open book in “Low-cost 3D Printing for Science, Education and Sustainable Development”, Trieste, ICTP 2013, ISBN 9295003489 The Arduino Uno microcontroller board a http://www.arduino.cc Servo.h library available at http://arduino.cc/en/reference/servo Raspberry Pi credit-card sized computer at http://www.raspberrypi.org Julius speech recognition decoder developed by CSRC, Japan at http://julius.sourceforge.jp ICTP-SDU voice2hand project at http://sdu.ictp.it/3D/voice2hand.html @sciencedissemination YouTube video "Voice to Sign Language System" at http://www.youtube.com/watch?v=J5whsEsGr4s YouTube video: “Speech Recognition using the Raspberry Pi” at http://www.youtube.com/watch?v=_U_v9tRD68k Top 10 iPhone Apps for People Who Are Deaf or Hard of Hearing at http://atcoalition.org/news/top10-iphone-apps-people-who-are-deaf-or-hard-hearing Free, open-source 3D printing robotic hands are available at http://www.thingiverse.com H. Dang and P. Allen, "Stable Grasping under Pose Uncertainty using Tactile Feedback", Autonomous Robots, August 2013. R.R. Ma, L.U. Odhner and A.M. Dollar, "A Modular, Open-Source 3D Printed Underactuated Hand", Proceedings IEEE International Conference on Robotics and Automation (ICRA) Karlsruhe, Germany, May 6-10, 2013, pp. 2722-2728. B. Leon, A. Morales and J. Sancho-Bru, "The Model of the Human Hand", Cognitive System Monographs 19 (2014) 123-173. P.S. Ramaiah , M. Venkateswara Rao , G.V. Satyanarayana, "A Microcontroller Based Four Fingered Robotic Hand", Int. J. Art. Int. & App. (IJAIA) 2 (2011) 90-102. 112