Contact Us

a11y+


Send email

Research

Our research seeks to improve the accessibility of interactive technologies. Accessibility does not only benefit people with disability but also empowers everyone. Recent advances in AI have enabled new forms of human-computer interaction characterized by greater adaptability and better human-machine symbiosis. Our research uses AI to accelerate the development of assistive technology and enhance the accessibility of user interfaces. We are looking for graduate students and undergraduate students to participant in these areas:

  • Disability, Usability, and Digital Accessibility
  • Adaptive human-technology interaction
  • Explainable AI and Visualization
  • Computer Science (STEAM) Education

You are welcome to visit CSUN Human-computing Lab @ JD2222 and our design space @ADC410.

Research Projects

A phone with zoom-in screen

By:

  • Jaime Garcia, Arts & Visual Design
  • Summayah Waseem, Computer Science

People with disabilities (PWDs) use video-sharing platforms such as YouTube to share experiences and information, which contains the challenges and difficulties they have met with social circles, technologies, and the environment. Human Computer-Interaction (HCI) researchers studied online videos to understand PWDs’ particular needs and challenges when doing daily activities, barriers with the video interactions, and ways to have playful experiences. This work presents a preliminary analysis of 257 video clips that mention disability challenges to understand how and why PWDs use YouTube to discourse disabilities.

Publication

Shuo Niu, Jaime Garcia, Summayah Waseem, and Li Liu. 2022. Investigating How People with Disabilities Disclose Difficulties on YouTube. In Proceedings of the 24th International ACM SIGACCESS Conference on Computers and Accessibility (ASSETS ’22). Association for Computing Machinery, New York, NY, USA, Article 58, 1–5. https://doi.org/10.1145/3517428.3550383

An Apple Watch

By: Megan Ngo

This case study explores the accessibility features of the Apple Watch, evaluating its usability for individuals with varying abilities. The study investigates features like AssistiveTouch, VoiceOver, haptic feedback, larger text options, and customized display settings. Through testing and user feedback, we examine how these features enable more inclusive interactions, assess their effectiveness, and identify areas for potential improvement to make wearable technology more accessible for everyone.

  • AI Visualization for multi-agent robotics by Gage Aschenbrenner and Ramin Roufeh
  • Automated Speech Recognition for Instructional Contents by Timothy Spengler
  • Visual Biofeedback in Speech Rehabilitation by Luan Ta
  • Behavioral Biometrics Classification by Shen Huang
  • Web Accessibility Suggestions through Short-text Classification by Gerardo Rodriguez
  • Deep Convolutional GANs in Procedural Terrain Generation Systems by Edgar Lopez-Garcia
  • Procedural Terrain Generation in Virtual Reality by Ryan Vitacion

Contact Us

a11y+


Send email

Scroll back to the top of the page