Repository logo

Multi-Modal Technology for User Interface Analysis including Mental State Detection and Eye Tracking Analysis

dc.contributor.authorHusseini Orabi, Ahmed
dc.contributor.supervisorLethbridge, Timothy
dc.date.accessioned2017-07-31T13:25:44Z
dc.date.available2017-07-31T13:25:44Z
dc.date.issued2017
dc.description.abstractWe present a set of easy-to-use methods and tools to analyze human attention, behaviour, and physiological responses. A potential application of our work is evaluating user interfaces being used in a natural manner. Our approach is designed to be scalable and to work remotely on regular personal computers using expensive and noninvasive equipment. The data sources our tool processes are nonintrusive, and captured from video; i.e. eye tracking, and facial expressions. For video data retrieval, we use a basic webcam. We investigate combinations of observation modalities to detect and extract affective and mental states. Our tool provides a pipeline-based approach that 1) collects observational, data 2) incorporates and synchronizes the signal modality mentioned above, 3) detects users' affective and mental state, 4) records user interaction with applications and pinpoints the parts of the screen users are looking at, 5) analyzes and visualizes results. We describe the design, implementation, and validation of a novel multimodal signal fusion engine, Deep Temporal Credence Network (DTCN). The engine uses Deep Neural Networks to provide 1) a generative and probabilistic inference model, and 2) to handle multimodal data such that its performance does not degrade due to the absence of some modalities. We report on the recognition accuracy of basic emotions for each modality. Then, we evaluate our engine in terms of effectiveness of recognizing basic six emotions and six mental states, which are agreeing, concentrating, disagreeing, interested, thinking, and unsure. Our principal contributions include the implementation of a 1) multimodal signal fusion engine, 2) real time recognition of affective and primary mental states from nonintrusive and inexpensive modality, 3) novel mental state-based visualization techniques, 3D heatmaps, 3D scanpaths, and widget heatmaps that find parts of the user interface where users are perhaps unsure, annoyed, frustrated, or satisfied.en
dc.identifier.urihttp://hdl.handle.net/10393/36451
dc.identifier.urihttp://dx.doi.org/10.20381/ruor-20731
dc.language.isoenen
dc.publisherUniversité d'Ottawa / University of Ottawaen
dc.subjectComputer visionen
dc.subjectConvolutional Neural Networken
dc.subjectMultimodal Deep learningen
dc.subjectMachine learningen
dc.subjectHCIen
dc.subjectUsabilityen
dc.subjectAffective computingen
dc.subjectMental statesen
dc.subjectEye trackingen
dc.subjectSoftware engineeringen
dc.subjectPsychophysiologyen
dc.subjectPipelineen
dc.subjectUmpleen
dc.subjectComponent-based Developmenten
dc.subjectEmotionen
dc.subjectFacial Action Coding Systemen
dc.subjectFacial Expressionen
dc.titleMulti-Modal Technology for User Interface Analysis including Mental State Detection and Eye Tracking Analysisen
dc.typeThesisen
thesis.degree.disciplineGénie / Engineeringen
thesis.degree.levelDoctoralen
thesis.degree.namePhDen
uottawa.departmentScience informatique et génie électrique / Electrical Engineering and Computer Scienceen

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail ImageThumbnail Image
Name:
Husseini_Orabi_Ahmed_2017_thesis.pdf
Size:
5.19 MB
Format:
Adobe Portable Document Format
Description:
PhD Thesis

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail ImageThumbnail Image
Name:
license.txt
Size:
6.65 KB
Format:
Item-specific license agreed upon to submission
Description: