mobile sign language learning outside the classroom · production. in addition to the three...

4
Mobile Sign Language Learning Outside the Classroom Kimberly A. Weaver School of Interactive Computing Georgia Institute of Technology Atlanta, GA 30332 USA [email protected] Thad Starner School of Interactive Computing Georgia Institute of Technology Atlanta, GA 30332 USA [email protected] Copyright is held by the author/owner(s). CHI 2012 EIST Workshop ACM. Abstract The majority of deaf children in the United States are born to hearing parents with limited prior exposure to American Sign Language (ASL). Our research involves creating and validating a mobile language tool called SMARTSign. The goal is to help hearing parents learn ASL in a way that fits seamlessly into their daily routine. Author Keywords sign language, informal learning, mobile language learning ACM Classification Keywords K.3.1 [Computers and Education]: Computer Uses in Education]: Computer-assisted instruction. Introduction Ninety-five percent of deaf children are born to hearing parents who have little prior experience with American Sign Language (ASL), the most accessible language to the deaf in North America [4]. When faced with the difficulties of communicating with their deaf child, most parents make the decision to learn ASL. They achieve varying degrees of aptitude. One factor impacting parents’ ability to learn, is that ASL can be as difficult for an English speaker to learn as Japanese or Chinese due to the use of simultaneous morphological inflections and non-manual grammar in ASL [3]. A second factor is the

Upload: others

Post on 15-Mar-2020

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Mobile Sign Language Learning Outside the Classroom · production. In addition to the three components, parents have the ability to set noti cations to remind them to study. They

Mobile Sign Language LearningOutside the Classroom

Kimberly A. Weaver

School of Interactive Computing

Georgia Institute of Technology

Atlanta, GA 30332 USA

[email protected]

Thad Starner

School of Interactive Computing

Georgia Institute of Technology

Atlanta, GA 30332 USA

[email protected]

Copyright is held by the author/owner(s).CHI 2012 EIST WorkshopACM.

AbstractThe majority of deaf children in the United States areborn to hearing parents with limited prior exposure toAmerican Sign Language (ASL). Our research involvescreating and validating a mobile language tool calledSMARTSign. The goal is to help hearing parents learnASL in a way that fits seamlessly into their daily routine.

Author Keywordssign language, informal learning, mobile language learning

ACM Classification KeywordsK.3.1 [Computers and Education]: Computer Uses inEducation]: Computer-assisted instruction.

IntroductionNinety-five percent of deaf children are born to hearingparents who have little prior experience with AmericanSign Language (ASL), the most accessible language to thedeaf in North America [4]. When faced with thedifficulties of communicating with their deaf child, mostparents make the decision to learn ASL. They achievevarying degrees of aptitude. One factor impacting parents’ability to learn, is that ASL can be as difficult for anEnglish speaker to learn as Japanese or Chinese due to theuse of simultaneous morphological inflections andnon-manual grammar in ASL [3]. A second factor is the

Page 2: Mobile Sign Language Learning Outside the Classroom · production. In addition to the three components, parents have the ability to set noti cations to remind them to study. They

limited availability of classes. Classes are taught in manyareas of the country but may not be accessible toindividuals living in more rural areas. There are also notmany classes that advance beyond introductory material.

Figure 1: The search interface

Figure 2: The quiz interface

Figure 3: The practice interface

Parents’ difficulties mastering ASL have a large impact ontheir child. Deaf children of hearing parents developlanguage at a much slower rate than hearing children ofhearing parents or deaf children of deaf parents. Thisslower development has been attributed both toincomplete language models and less parent-childinteraction [1, 5].

Application designSMARTSign is an application designed to help parentslearn and practice ASL on their mobile phones. Byproviding parents with better tools for learning thelanguage, we hope to eliminate the language deficitexperienced by deaf children of hearing parents. Thecurrent focus of SMARTSign is on vocabulary acquisition.Future work will involve incorporating ASL grammar andfluency practice into the application.

SMARTSign is an Android application designed forsmartphones with front-facing cameras. It currently hasthree components for learning: search, quiz, and practice.Each component fulfills a different function. The searchcomponent is intended to be used when parents haveimmediate communication needs with their child. Theycan search for an ASL video either by typing the Englishequivalent on the keyboard or by speaking using speechrecognition. If the parent chooses to use the keyboard forinput, the search interface suggests terms from thedictionary as soon as characters are entered to aid infaster retrieval. The search interface is shown in Figure 1.The search component aids learning because parents are

retrieving signs so they can immediately use them incontext.

The quiz interface is shown in Figure 2. This componentis useful when parents have a free moment during the dayto study new vocabulary. First, a video with a sign isshown. Once the video has finished playing, parents arepresented with five choices: four vocabulary words and oneoption when the parent just wishes to see the answer. Ifthe parent chooses the incorrect response, the target videowill be replayed and the incorrect option will be grayedout when the choices are repeated. If the parent chooses“Show me the answer,” the video will be replayed and theanswer will be displayed. Parents can continue watchingvideos and responding for as long as they have time.

In interviews, parents reported feeling uncomfortablesigning. We created the practice component so they couldrecord themselves performing a sign and compare it to theexample video. This interface is shown in Figure 3.Parents can continue record themselves until they feelcomfortable with their performance. Previous researchwith just a quiz interface showed that it could helpindividuals learn how to recognize ASL videos, but theywere not able to recall how to produce the signs [2].Including the practice component changes theapplication’s learning focus from sign recognition to signproduction.

In addition to the three components, parents have theability to set notifications to remind them to study. Theycan specify both the time of the reminder and the days ofthe week they would like to be notified. This way parentscan set reminders for times when they are least likely tohave other responsibilities. Parents can also specify howmany new signs they would like to learn daily in the quizcomponent. Once the parent has studied the new words

Page 3: Mobile Sign Language Learning Outside the Classroom · production. In addition to the three components, parents have the ability to set noti cations to remind them to study. They

for the day, any further use will be review.

Parents are also able to track their progress using theapplication. Progress is determined by the number ofsigns they have gotten correct when using the quizcomponent and signed when using the practicecomponent. They can view general progress learning all ofthe available vocabulary as well as progress learningdifferent word categories. If a parent selects a category intheir progress report, they can also see specific details onthe individual words, how many times they have viewed asign versus how many times they have responded correctlyin the quiz, their latest response, and when a word asbeen signed using practice. Initial deployments ofSMARTSign to hearing parents with deaf children havebeen promising. Parents have reported enjoying theapplication and sharing it with others in their family.

Making time for learningSMARTSign is designed for parents who do not haveaccess to ASL classes. We plan to conduct a study withthe goal of understanding how parents decide when it istime to learn and how these decisions impact their abilityto learn to recognize and produce ASL vocabulary. All ofthe parents will be given access to a smartphone runningthe SMARTSign application for a month. To investigatehow to motivate use of the system, half of the parents willbe told the application is designed for learning commonwords that should be used and recognized by theirchildren. The other half will be told the system is forlearning vocabulary for common children’s stories such asGo Dog, Go!.

We will collect a variety of data over the course of thestudy. A pre- and post-test of vocabulary knowledge willbe administered to all parents. This data will give

information about how well parents are able to learn signlanguage using the SMARTSign application. We will alsorecord detailed logs of application use. This data willinclude: when people access the application, how longpeople access the different components, and whatvocabulary is accessed. Parents will also be asked toprovide information about where they are, who they areusing the system with, and reasons for accessing thesystem throughout the course of the study. Thisinformation will help create a better understanding of thelevel of parental self motivation. It could be that parentsare most interested in a tool for immediatecommunication provided by a dictionary search ofSMARTSign express and will ignore study notifications.Another possible situation is that parents will takeadvantage of the mobile affordances of the device and findsmall moments throughout the day to study more often.

Deploying technologyWe have performed two formative studies with our targetpopulation: an interview study and a four-day deploymentfor usability issues. In interviews with hearing parents ofdeaf children, parents provided feedback about what wasimportant to them in a phone. Two mothers were mostinterested in the devices with front-facing cameras. Onemother said that since her son is getting older (he’s eightnow), he will start going out alone to play with friends. Ifshe wants to be able to communicate with him, signlanguage would be the most convenient. The existence oftools like FaceTime make the iPhone desirable. Phonesalso serve an added bonus as entertainment for theirchildren while waiting. A parent described how doctors’offices do not usually activate captioning on their TV. Sheliked being able to give her phone to her child to playgames as entertainment. Parents are also excited aboutpotential educational opportunities for their child that can

Page 4: Mobile Sign Language Learning Outside the Classroom · production. In addition to the three components, parents have the ability to set noti cations to remind them to study. They

be provided by smartphones. One mother said her homeprovider has been encouraging her to purchase an iPhonefor this reason.

The difficulty with designing a learning tool for parents ishow broad should we reach and how quickly. Currentphone ownership is far from standardized. Four out ofeleven parents reported owning a smartphone. Twoparents owned iPhones, one parent owned an Androiddevice, and one parent owned an older Palm device butwas intending to buy a new Android phone in the nearfuture. Some mothers reported that their husbands ownedsmartphones, but they had to make decisions based onfinances as to who in the family had smartphones and whohad “dumb” phones. Data plans were also not common.Only half of the families already possessed data plans.

Developers have to make a decision on the trade-offsbetween developing for multiple platforms for widestaccess to potential users or developing for a singleplatform for better experimental control. We areattempting to find a middle path with SMARTSign. Ourinitial development and deployment is taking place on theAndroid platform. We will provide the same type of phoneas well as a data plan to all participants so that everyonewill have access to a full implementation of theapplication. This will allow us to validate the tools wehave created. Once the tool has been validated, we canthen focus on a broader deployment of the system forgreater impact.

ConclusionIn this paper we have explored what it means to create amobile informal learning tool for hearing parents of deafchildren attempting to learn sign language. We describedthe application and outlined a study which will investigate

how using this application will impact parents’ ability tolearn vocabulary.

The final contribution of this paper is a reflection on thehow device platform decisions may impact thedevelopmental cycle and the evaluation of mobileapplications. In the case of SMARTSign we strive first forexperimental validity and will then expand development toprovide wider access to the target population.

References[1] H. Hamilton and D. Lillo-Martin. Imitative production

of ASL verbs of movement and location: Acomparative study. Sign Language Studies, 1986.

[2] V. Henderson-Summet. Facilitating Communicationfor Deaf Individuals with Mobile Technologies.Doctoral thesis, Georgia Institute of Technology,Atlanta, GA, USA, 2010.

[3] R. Jacobs. Just how hard is it to learn ASL? the casefor ASL as a truly foreign language. In MulticulturalAspects of Sociolinguistics in Deaf CommunitiesSeries, volume 2, pages 183–226. Gallaudet UniversityPress, Washington, DC, 1996.

[4] R. E. Mitchell and M. A. Karchmer. Chasing themythical ten percent: Parental hearing status of deafand hard of hearing students in the united states. SignLanguage Studies, 4(2):138–163, 2004.

[5] P. Spencer and A. Lederberg. Different modes,different models: Communication and language ofyoung deaf children and their mothers. In L. Adamsonand M. Rornski, editors, Research on communicationand language disorders: Contributions to theories oflanguage development. Brooks, Baltimore, 1997.