Show simple item record

dc.contributor.advisorMurphy, Robin R
dc.creatorSrinivasan, Vasant
dc.date.accessioned2016-09-22T19:39:27Z
dc.date.available2016-09-22T19:39:27Z
dc.date.created2014-08
dc.date.issued2014-08-12
dc.date.submittedAugust 2014
dc.identifier.urihttps://hdl.handle.net/1969.1/157923
dc.description.abstractThis research demonstrates that robots can achieve socially acceptable interactions, using loosely synchronized head gaze-speech, without understanding the semantics of the dialog. Prior approaches used tightly synchronized head gaze-speech, which requires significant human effort and time to manually annotate synchronization events in advance, restricting interactive dialog, and requiring the operator to act as a puppeteer. This approach has two novel aspects. First, it uses affordances in the sentence structure, time delays, and typing to achieve autonomous synchronization of head gaze-speech. Second, it is implemented within a behavioral robotics framework derived from 32 previous implementations. The efficacy of the loosely synchronized approach was validated through a 93-participant 1 x 3 (loosely synchronized head gaze-speech, tightly synchronized head gaze-speech, no-head gazespeech) between-subjects experiment using the “Survivor Buddy” rescue robot in a victim management scenario. The results indicated that the social acceptance of loosely synchronized head gaze-speech is similar to tightly synchronized head gazespeech (manual annotation), and preferred to the no head gaze-speech case. These findings contribute to the study of social robotics in three ways. First, the research overall contributes to a fundamental understanding of the role of social head gaze in social acceptance, and the production of social head gaze. Second, it shows that autonomously generated head gaze-speech coordination is both possible and acceptable. Third, the behavioral robotics framework simplifies creation, analysis, and comparison of implementations.en
dc.format.mimetypeapplication/pdf
dc.language.isoen
dc.subjectSocial Head Gazeen
dc.subjectAutonomous Annotationen
dc.subjectInteractive Conversationen
dc.subjectUser Study and Evaluationen
dc.subjectHuman-Robot Interactionen
dc.titleHigh Social Acceptance of Head Gaze Loosely Synchronized with Speech for Social Robotsen
dc.typeThesisen
thesis.degree.departmentComputer Science and Engineeringen
thesis.degree.disciplineComputer Engineeringen
thesis.degree.grantorTexas A & M Universityen
thesis.degree.nameDoctor of Philosophyen
thesis.degree.levelDoctoralen
dc.contributor.committeeMemberSong, Dezhen
dc.contributor.committeeMemberShell, Dylan A
dc.contributor.committeeMemberFerris, Thomas K
dc.contributor.committeeMemberBethel, Cindy L
dc.type.materialtexten
dc.date.updated2016-09-22T19:39:27Z
local.etdauthor.orcid0000-0003-1931-9882


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record