Prof. Dr. Stefan Schneegass
- SM 208
- (+49) 201/183 - 4251
- stefan.schneegass (at) uni-due.de
- University of Duisburg-Essen
Institute for Computer Science and Business Information Systems (ICB)
Stefan Schneegass is an assistant professor for computer science at the University of Duisburg-Essen. He received his Ph.D. from the University of Stuttgart in 2016. Since the start of his Ph.D. in 2012, he worked on several different national and international research projects and published at the major venues in human-computer interaction. Stefan holds a B.Sc. and M.Sc. from the University of Duisburg-Essen. During his studies, he worked as student assistant at the Pervasive Computing and User Interface Engineering group and for the DFKI Saarbrucken. His current research interest is in the area of human-computer interaction, in particular mobile, wearable, and ubiquitous interaction.
- Borsum, Florian; Pascher, Max; Auda, Jonas; Schneegass, Stefan; Lux, Gregor; Gerken, Jens: Stay on Course in VR: Comparing the Precision of Movement between Gamepad, Armswinger, and Treadmill. In: MuC '21. ACM, 2021. CitationAbstractDetails
In diesem Beitrag wird untersucht, inwieweit verschiedene Formen von Fortbewegungstechniken in Virtual Reality UmgebungenEinfluss auf die Präzision bei der Interaktion haben. Dabei wurden insgesamt drei Techniken untersucht: Zwei der Techniken integrierendabei eine körperliche Aktivität, um einen hohen Grad an Realismus in der Bewegung zu erzeugen (Armswinger, Laufstall). Alsdritte Technik wurde ein Gamepad als Baseline herangezogen. In einer Studie mit 18 Proband:innen wurde die Präzision dieser dreiFortbewegungstechniken über sechs unterschiedliche Hindernisse in einem VR-Parcours untersucht. Die Ergebnisse zeigen, dassfür einzelne Hindernisse, die zum einen eine Kombination aus Vorwärts- und Seitwärtsbewegung erfordern (Slalom, Klippe) sowieauf Geschwindigkeit abzielen (Schiene), der Laufstall eine signifikant präzisere Steuerung ermöglicht als der Armswinger. Auf dengesamten Parcours gesehen ist jedoch kein Eingabegerät signifikant präziser als ein anderes. Die Benutzung des Laufstalls benötigtzudem signifikant mehr Zeit als Gamepad und Armswinger. Ebenso zeigte sich, dass das Ziel, eine reale Laufbewegung 1:1 abzubilden,auch mit einem Laufstall nach wie vor nicht erreicht wird, die Bewegung aber dennoch als intuitiv und immersiv wahrgenommenwird.
- Arboleda, S. A.; Pascher, Max; Lakhnati, Y.; Gerken, Jens: Understanding Human-Robot Collaboration for People with Mobility Impairments at the Workplace, a Thematic Analysis. In: 29th IEEE International Conference on Robot and Human Interactive Communication. ACM, 2021. doi:10.1109/RO-MAN47096.2020.9223489.CitationAbstractDetails
Assistive technologies such as human-robot collaboration, have the potential to ease the life of people with physical mobility impairments in social and economic activities. Currently, this group of people has lower rates of economic participation, due to the lack of adequate environments adapted to their capabilities. We take a closer look at the needs and preferences of people with physical mobility impairments in a human-robot cooperative environment at the workplace. Specifically, we aim to design how to control a robotic arm in manufacturing tasks for people with physical mobility impairments. We present a case study of a sheltered-workshop as a prototype for an institution that employs people with disabilities in manufacturing jobs. Here, we collected data of potential end-users with physical mobility impairments, social workers, and supervisors using a participatory design technique (Future-Workshop). These stakeholders were divided into two groups, primary (end-users) and secondary users (social workers, supervisors), which were run across two separate sessions. The gathered information was analyzed using thematic analysis to reveal underlying themes across stakeholders. We identified concepts that highlight underlying concerns related to the robot fitting in the social and organizational structure, human-robot synergy, and human-robot problem management. In this paper, we present our findings and discuss the implications of each theme when shaping an inclusive human-robot cooperative workstation for people with physical mobility impairments.
- Arboleda, Stephanie Arevalo; Pascher, Max; Baumeister, Annalies; Klein, Barbara; Gerken, Jens: Reflecting upon Participatory Design in Human-Robot Collaboration for People with Motor Disabilities: Challenges and Lessons Learned from Three Multiyear Projects. ACM, New York, NY, USA 2021. doi:https://doi.org/10.1145/3453892.3458044CitationAbstractDetails
Human-robot technology has the potential to positively impact the lives of people with motor disabilities. However, current efforts have mostly been oriented towards technology (sensors, devices, modalities, interaction techniques), thus relegating the user and their valuable input to the wayside. In this paper, we aim to present a holistic perspective of the role of participatory design in Human-Robot Collaboration (HRC) for People with Motor Disabilities (PWMD). We have been involved in several multiyear projects related to HRC for PWMD, where we encountered different challenges related to planning and participation, preferences of stakeholders, using certain participatory design techniques, technology exposure, as well as ethical, legal, and social implications. These challenges helped us provide five lessons learned that could serve as a guideline to researchers when using participatory design with vulnerable groups. In particular, young researchers who are starting to explore HRC research for people with disabilities.
- Liebers, Jonathan; Abdelaziz, Mark; Mecke, Lukas; Saad, Alia; Auda, Jonas; Alt, Florian; Schneegaß, Stefan: Understanding User Identification in Virtual Reality Through Behavioral Biometrics and the Effect of Body Normalization. In: CHI Conference on Human Factors in Computing Systems 2021. ResearchGate, Yokohama, Japan 2021. doi:10.1145/3411764.3445528CitationDetails
- Detjen, Henrik; Geisler, Stefan; Schneegass, Stefan: "Help, Accident Ahead!": Using Mixed Reality Environments in Automated Vehicles to Support Occupants After Passive Accident Experiences. In: Acm (Ed.): AutomotiveUI (adjunct) 2020. 2020. doi:10.1145/3409251.3411723CitationDetails
- Detjen, Henrik; Pfleging, Bastian; Schneegass, Stefan: A Wizard of Oz Field Study to Understand Non-Driving-Related Activities, Trust, and Acceptance of Automated Vehicles. In: AutomotiveUI 2020. ACM, 2020. doi:10.1145/3409120.3410662CitationDetails
- Schneegaß, Stefan; Auda, Jonas; Heger, Roman; Grünefeld, Uwe; Kosch, Thomas: EasyEG: A 3D-printable Brain-Computer Interface. In: UIST 2020. Minnesota, USA 2020. doi:https://doi.org/10.1145/3379350.3416189CitationDetails
- Poguntke, Romina; Schneegass, Christina; van der Vekens, Lucas; Rzayev, Rufat; Auda, Jonas; Schneegass, Stefan; Schmidt:, Albrecht: NotiModes: an investigation of notification delay modes and their effects on smartphone usersy. In: MuC '20: Proceedings of the Conference on Mensch und Computer. ACM, 2020. doi:https://doi.org/10.1145/3404983.3410006CitationDetails
- Agarwal, Shivam; Auda, Jonas; Schneegaß, Stefan; Beck:, Fabian: A Design and Application Space for Visualizing User Sessions of Virtual and Mixed Reality Environments. In: VMV2020. ACM, 2020. doi:10.2312/vmv.20201194CitationDetails
- Poguntke, Romina; Schneegass, Christina; van der Vekens, Lucas; Rzayev, Rufat; Auda, Jonas; Schneegass, Stefan; Schmidt, Albrecht: NotiModes: an investigation of notification delay modes and their effects on smartphone users. In: MuC '20: Proceedings of the Conference on Mensch und Computer. ACM, Magdebug, Germany 2020. doi:10.1145/3404983.3410006CitationDetails
- Saad, Alia; Elkafrawy, Dina Hisham; Abdennadher, Slim; Schneegass, Stefan: Are They Actually Looking? Identifying Smartphones Shoulder Surfing Through Gaze Estimation. In: ETRA. ACM, Stuttgart, Germany 2020. doi:10.1145/3379157.3391422CitationDetails
- Jonathan Liebers, Stefan Schneegass: Gaze-based Authentication in Virtual Reality. In: ETRA. ACM, 2020. doi:10.1145/3379157.3391421CitationDetails
- Safwat, Sherine Ashraf; Bolock, Alia El; Alaa, Mostafa; Faltaous, Sarah; Schneegass, Stefan; Abdennadher, Slim: The Effect of Student-Lecturer Cultural Differences on Engagement in Learning Environments - A Pilot Study. In: Communications in Computer and Information Science. Springer, 2020. doi:10.1007/978-3-030-51999-5\_10CitationDetails
- Schneegass, Stefan; Sasse, Angela; Alt, Florian; Vogel, Daniel: Authentication Beyond Desktops and Smartphones: Novel Approaches for Smart Devices and Environments. In: CHI'20 Proceedings. ACM, Honolulu, HI, USA 2020. doi:10.1145/3334480.3375144CitationDetails
- Ranasinghe, Champika; Holländer, Kai; Currano, Rebecca; Sirkin, David; Moore, Dylan; Schneegass, Stefan; Ju, Wendy: Autonomous Vehicle-Pedestrian Interaction Across Cultures: Towards Designing Better External Human Machine Interfaces (eHMIs). In: CHI'20 Proceedings. ACM, Honolulu, HI, USA 2020. doi:10.1145/3334480.3382957CitationDetails
- Jonathan Liebers, Stefan Schneegass: Introducing Functional Biometrics: Using Body-Reflections as a Novel Class of Biometric Authentication Systems. In: CHI Extended Abstracts 2020. ACM, Honolulu, HI, USA 2020. doi:10.1145/3334480.3383059CitationDetails
- Pascher, Max; Schneegass, Stefan; Gerken, Jens; Lamas, David; Loizides, Fernando; Nacke, Lennart; Petrie, Helen; Winckler, Marco; Zaphiris, Panayiotis: SwipeBuddy: A Teleoperated Tablet and Ebook-Reader Holder for a Hands-Free Interaction. In: INTERACT 2019. Springer, 2019. CitationAbstractDetails
Mobile devices are the core computing platform we use in our everyday life to communicate with friends, watch movies, or read books. For people with severe physical disabilities, such as tetraplegics, who cannot use their hands to operate such devices, these devices are barely usable. Tackling this challenge, we propose SwipeBuddy, a teleoperated robot allowing for touch interaction with a smartphone, tablet, or ebook-reader. The mobile device is mounted on top of the robot and can be teleoperated by a user through head motions and gestures controlling a stylus simulating touch input. Further, the user can control the position and orientation of the mobile device. We demonstrate the SwipeBuddy robot device and its different interaction capabilities.
- Detjen, Henrik; Faltaous, Sarah; Geisler, Stefan; Schneegass, Stefan: User-Defined Voice and Mid-Air Gesture Commands - for Maneuver-based Interventions in Automated Vehicles. In: MuC'19: Proceedings of Mensch und Computer 2019. ACM, New York, USA 2019. doi: DOI: https://doi.org/10.1145/3340764.3340798 PDFCitationDetails
- Poguntke, Romina; Mantz, Tamara; Hassib, Mariam; Schmidt, Albrecht; Schneegass, Stefan: Smile to Me - Investigating Emotions and their Representation in Text-based Messaging in the Wild. In: MuC'19: Proceedings of Mensch und Computer 2019. ACM, New York, USA 2019. doi:https://doi.org/10.1145/3340764.3340795PDFCitationDetails
- Pascher, Max; Schneegass, Stefan; Gerken:, Jens: SwipeBuddy A Teleoperated Tablet and Ebook-Reader Holder for a Hands-Free Interaction. In: Human-Computer Interaction – INTERACT 2019. Springer, Paphos, Cyprus 2019. doi:10.1007/978-3-030-29390-1_39CitationDetails
- Pfeiffer, Max; Medrano, Samuel Navas; Auda, Jonas; Schneegass, Stefan: STOP! Enhancing Drone Gesture Interaction with Force Feedback. In: CHI'19 Proceedings. HAL, Glasgow, UK 2019. doi:https://hal.archives-ouvertes.fr/hal-02128395/documentPDFFull textCitationDetails
- Jonas Auda, Max Pascher; Schneegass, Stefan: Around the (Virtual) World - Infinite Walking in Virtual Reality Using Electrical Muscle Stimulation. In: Acm (Ed.): CHI'19 Proceedings. Glasgow 2019. doi:https://doi.org/10.1145/3290605.3300661PDFCitationAbstractDetails
Virtual worlds are infinite environments in which the user can move around freely. When shifting from controller-based movement to regular walking as an input, the limitation of the real world also limits the virtual world. Tackling this challenge, we propose the use of electrical muscle stimulation to limit the necessary real-world space to create an unlimited walking experience. We thereby actuate the users‘ legs in a way that they deviate from their straight route and thus, walk in circles in the real world while still walking straight in the virtual world. We report on a study comparing this approach to vision shift – the state-of-the-art approach – as well as combining both approaches. The results show that particularly combining both approaches yield high potential to create an infinite walking experience.
- Faltaous, Sarah; Haas, Gabriel; Barrios, Liliana; Seiderer, Andreas; Rauh, Sebastian Felix; Chae, Han Joo; Schneegass, Stefan; Alt, Florian: BrainShare: A Glimpse of Social Interaction for Locked-in Syndrome Patients. In: CHI EA '19: Extended Abstracts of the 2019 CHI Conference on Human Factors in Computing Systems. ACM, New York, NY, USA 2019. PDFCitationDetails
- Schneegass, Stefan; Poguntke, Romina; Machulla, Tonja Katrin: Understanding the Impact of Information Representation on Willingness to Share Information. In: CHI '19: Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. ACM, New YorkNYUnited States 2019. doi:https://doi.org/10.1145/3290605.3300753PDFCitationDetails
- Sarah Faltaous, Jonathan Liebers; Yomna Abdelrahman, Florian Alt; Schneegass, Stefan: VPID: Towards Vein Pattern Identification Using Thermal Imaging. In: i-com (2019) No 18 (3), p. 259-270. doi:10.1515/icom-2019-0009PDFCitationAbstractDetails
Biometric authentication received considerable attention lately. The vein pattern on the back of the hand is a unique biometric that can be measured through thermal imaging. Detecting this pattern provides an implicit approach that can authenticate users while interacting. In this paper, we present the Vein-Identification system, called VPID. It consists of a vein pattern recognition pipeline and an authentication part. We implemented six different vein-based authentication approaches by combining thermal imaging and computer vision algorithms. Through a study, we show that the approaches achieve a low false-acceptance rate (“FAR”) and a low false-rejection rate (“FRR”). Our findings show that the best approach is the Hausdorff distance-difference applied in combination with a Convolutional Neural Networks (CNN) classification of stacked images.
- Hoppe, Matthias; Knierim, Pascal; Kosch, Thomas; Funk, Markus; Futami, Lauren; Schneegass, Stefan; Henze, Niels; Schmidt, Albrecht; Machulla, Tonja: VRHapticDrones - Providing Haptics in Virtual Reality through Quadcopters. In: MUM 2018: Proceedings of the 17th International Conference on Mobile and Ubiquitous Multimedia. ACM, Cairo, Egypt 2018. doi:https://doi.org/10.1145/3282894.3282898PDFCitationDetails
- Schneegass, Christina; Terzimehić, Nađa; Nettah, Mariam; Schneegass, Stefan: Informing the Design of User - adaptive Mobile Language Learning Applications. In: MUM 2018: Proceedings of the 17th International Conference on Mobile and Ubiquitous Multimedia. ACM, Cairo, Egypt 2018. doi:https://doi.org/10.1145/3282894.3282926PDFCitationDetails
- Faltaous, Sarah; Elbolock, Alia; Talaat, Mostafa; Abdennadher, Slim; Schneegass, Stefan: Virtual Reality for Cultural Competences. In: MUM 2018: Proceedings of the 17th International Conference on Mobile and Ubiquitous Multimedia. ACM, Cairo, Egypt 2018. doi:https://doi.org/10.1145/3282894.3289739PDFCitationDetails
- Elagroudy, Passant; Abdelrahman, Yomna; Faltaous, Sarah; Schneegass, Stefan; Davis, Hilary: Workshop on Amplified and Memorable Food Interactions. In: MUM 2018: Proceedings of the 17th International Conference on Mobile and Ubiquitous Multimedia. ACM, Cairo, Egypt 2018. doi:https://doi.org/10.1145/3282894.3286059PDFCitationDetails
- Auda, Jonas; Hoppe, Matthias; Amiraslanov, Orkhan; Zhou, Bo; Knierim, Pascal; Schneegass, Stefan; Schmidt, Albrecht; Lukowicz, Paul: LYRA - smart wearable in-flight service assistant. In: ISWC '18: Proceedings of the 2018 ACM International Symposium on Wearable Computers. ACM, Singapore, Singapore 2018. doi:https://doi.org/10.1145/3267242.3267282PDFCitationDetails
- Faltaous, Sarah; Baumann, M.; Schneegass, Stefan; Chuang, Lewis: Design Guidelines for Reliability Communication in Autonomous Vehicles. In: AutomotiveUI '18: Proceedings of the 10th International Conference on Automotive User Interfaces and Interactive Vehicular Applications. ACM, Toronto, Canada 2018. doi:https://doi.org/10.1145/3239060.3239072PDFCitationDetails
- Poguntke, Romina; Tasci, Cagri; Korhonen, Olli; Alt, Florian; Schneegass, Stefan: AVotar - exploring personalized avatars for mobile interaction with public displays. In: MobileHCI '18: Proceedings of the 20th International Conference on Human-Computer Interaction with Mobile Devices and Services Adjunct. ACM, Barcelona, Spain 2018. doi:https://doi.org/10.1145/3236112.3236113PDFCitationDetails
- Weber, Dominik; Voit, Alexandra; Auda, Jonas; Schneegass, Stefan; Henze, Niels: Snooze! - investigating the user-defined deferral of mobile notifications. In: MobileHCI '18: Proceedings of the 20th International Conference on Human-Computer Interaction with Mobile Devices and Services. ACM, Barcelona, Spain 2018. doi:https://doi.org/10.1145/3229434.3229436PDFCitationDetails
- Poguntke, Romina; Kiss, Francisco; Kaplan, Ayhan; Schmidt, Albrecht; Schneegass, Stefan: RainSense - exploring the concept of a sense for weather awareness. In: MobileHCI '18: Proceedings of the 20th International Conference on Human-Computer Interaction with Mobile Devices and Services Adjunct. ACM, Barcelona, Spain 2018. doi:https://doi.org/10.1145/3236112.3236114PDFCitationDetails
- Voit, Alexandra; Salm, Marie Olivia; Beljaars, Miriam; Kohn, Stefan; Schneegass, Stefan: Demo of a smart plant system as an exemplary smart home application supporting non-urgent notifications. In: NordiCHI '18: Proceedings of the 10th Nordic Conference on Human-Computer Interaction. ACM, Oslo, Norway 2018. doi:https://doi.org/10.1145/3240167.3240231PDFCitationDetails
- Hassib, Mariam; Schneegass, Stefan; Henze, Niels; Schmidt, Albrecht; Alt, Florian: A Design Space for Audience Sensing and Feedback Systems. In: CHI EA '18: Extended Abstracts of the 2018 CHI Conference on Human Factors in Computing Systems. ACM, Montreal, Canada 2018. doi:https://doi.org/10.1145/3170427.3188569PDFCitationDetails
- Kiss, Francisco; Boldt, Robin; Pfleging, Bastian; Schneegass, Stefan: Navigation Systems for Motorcyclists: Exploring Wearable Tactile Feedback for Route Guidance in the Real World. In: CHI '18: Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems. ACM, Montreal, Canada 2018. doi:https://doi.org/10.1145/3173574.3174191PDFCitationDetails
- Voit, Alexandra; Pfähler, Ferdinand; Schneegass, Stefan: Posture Sleeve: Using Smart Textiles for Public Display Interactions. In: CHI EA '18: Extended Abstracts of the 2018 CHI Conference on Human Factors in Computing Systems. ACM, Montreal 2018. doi:https://doi.org/10.1145/3170427.3188687PDFCitationDetails
- Auda, Jonas; Weber, Dominik; Voit, Alexandra; Schneegass, Stefan: Understanding User Preferences towards Rule-based Notification Deferral. In: CHI EA '18: Extended Abstracts of the 2018 CHI Conference on Human Factors in Computing Systems. ACM, Montreal, Canada 2018. doi:https://doi.org/10.1145/3170427.3188688PDFCitationDetails
- Hassib, Mariam; Khamis, Mohamed; Friedl, Susanne; Schneegass, Stefan; Alt, Florian: Brainatwork - logging cognitive engagement and tasks in the workplace using electroencephalography. In: MUM '17: Proceedings of the 16th International Conference on Mobile and Ubiquitous Multimedia. ACM, Stuttgart, Germany 2017. doi:https://doi.org/10.1145/3152832.3152865PDFCitationDetails
- Duente, Im; Schneegass, Stefan; Pfeiffer, Max: EMS in HCI - challenges and opportunities in actuating human bodies. In: MobileHCI '17: Proceedings of the 19th International Conference on Human-Computer Interaction with Mobile Devices and Services. ACM, Vienna, Austria 2017. doi:https://doi.org/10.1145/3098279.3119920PDFCitationDetails
- Oberhuber, Sascha; Kothe, Tina; Schneegass, Stefan; Alt, Florian: Augmented Games - Exploring Design Opportunities in AR Settings With Children. In: IDC '17: Proceedings of the 2017 Conference on Interaction Design and Children. ACM, Stanford California, USA 2017. doi:https://doi.org/10.1145/3078072.3079734PDFCitationDetails
- Knierim, Pascal; Kosch, Thomas; Schwind, Valentin; Funk, Markus; Kiss, Francisco; Schneegass, Stefan; Henze, Niels: Tactile Drones - Providing Immersive Tactile Feedback in Virtual Reality through Quadcopters. In: CHI EA '17: Proceedings of the 2017 CHI Conference Extended Abstracts on Human Factors in Computing Systems. ACM, Denver Colorado, USA 2017. doi:https://doi.org/10.1145/3027063.3050426PDFCitationDetails
- Schmidt, Albrecht; Schneegass, Stefan; Kunze, Kai; Rekimoto, Jun; Woo, Woontack: Workshop on Amplification and Augmentation of Human Perception. In: CHI EA '17: Proceedings of the 2017 CHI Conference Extended Abstracts on Human Factors in Computing Systems. ACM, Denver Colorado, USA 2017. doi:https://doi.org/10.1145/3027063.3027088PDFCitationDetails
- Alt, Florian: Stay Cool! Understanding Thermal Attacks on Mobile-based User Authentication. In: Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems (CHI '17). ACM, New York, USA 2017, p. 3751-3763. doi:https://doi.org/10.1145/3025453.3025461PDFCitationAbstractDetails
PINs and patterns remain among the most widely used knowledge-based authentication schemes. As thermal cameras become ubiquitous and affordable, we foresee a new form of threat to user privacy on mobile devices. Thermal cameras allow performing thermal attacks, where heat traces, resulting from authentication, can be used to reconstruct passwords. In this work we investigate in details the viability of exploiting thermal imaging to infer PINs and patterns on mobile devices. We present a study (N=18) where we evaluated how properties of PINs and patterns influence their thermal attacks resistance. We found that thermal attacks are indeed viable on mobile devices; overlapping patterns significantly decrease successful thermal attack rate from 100% to 16.67%, while PINs remain vulnerable (>72% success rate) even with duplicate digits. We conclude by recommendations for users and designers of authentication schemes on how to resist thermal attacks.
- Mariam Hassib, Max Pfeiffer; Stefan Schneegass, Michael Rohs; Alt, Florian: Emotion Actuator - Embodied Emotional Feedback through Electroencephalography and Electrical Muscle Stimulation. In: Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems (CHI '17). ACM, New York, USA 2017, p. 6133-6146. doi:https://doi.org/10.1145/3025453.3025953PDFCitationAbstractDetails
The human body reveals emotional and bodily states through measurable signals, such as body language and electroencephalography. However, such manifestations are difficult to communicate to others remotely. We propose EmotionActuator, a proof-of-concept system to investigate the transmission of emotional states in which the recipient performs emotional gestures to understand and interpret the state of the sender.We call this kind of communication embodied emotional feedback, and present a prototype implementation. To realize our concept we chose four emotional states: amused, sad, angry, and neutral. We designed EmotionActuator through a series of studies to assess emotional classification via EEG, and create an EMS gesture set by comparing composed gestures from the literature to sign-language gestures. Through a final study with the end-to-end prototype interviews revealed that participants like implicit sharing of emotions and find the embodied output to be immersive, but want to have control over shared emotions and with whom. This work contributes a proof of concept system and set of design recommendations for designing embodied emotional feedback systems.
- Alt, Florian: EngageMeter: A System for Implicit Audience Engagement Sensing Using Electroencephalography. In: Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems (CHI '17). ACM, New York, USA 2017, p. 5114-5119. doi:https://doi.org/10.1145/3025453.3025669CitationAbstractDetails
Obtaining information about audience engagement in presentations is a valuable asset for presenters in many domains. Prior literature mostly utilized explicit methods of collecting feedback which induce distractions, add workload on audience and do not provide objective information to presenters. We present EngageMeter - a system that allows fine-grained information on audience engagement to be obtained implicitly from multiple brain-computer interfaces (BCI) and to be fed back to presenters for real time and post-hoc access. Through evaluation during an HCI conference (Naudience=11, Npresenters=3) we found that EngageMeter provides value to presenters (a) in real-time, since it allows reacting to current engagement scores by changing tone or adding pauses, and (b) in post-hoc, since presenters can adjust their slides and embed extra elements. We discuss how EngageMeter can be used in collocated and distributed audience sensing as well as how it can aid presenters in long term use.
- Michahelles, Florian; Ilic, Alexander; Kunze, Kai; Kritzler, Mareike; Schneegass, Stefan: IoT 2016. In: IEEE Pervasive Computing, Vol 16 (2017) No 2, p. 87-89. doi:10.1109/MPRV.2017.25PDFCitationAbstractDetails
The 6th International Conference on the Internet of Things (IoT 2016) showed clear departure from the research on data acquisition and sensor management presented at previous series of this conference. Learn about this year's move toward more commercially applicable implementations and cross-domain applications.
- Stefan Schneegass, Oliver Amft: Introduction to Smart Textiles. In: Stefan Schneegass, Oliver Amft (Ed.): Smart Textiles: Fundamentals, Design, and Interaction. Springer International Publishing, 2017, p. 1-15. doi:10.1007/978-3-319-50124-6_1CitationAbstractDetails
This chapter introduces fundamental concepts related to wearable computing, smart textiles, and context awareness. The history of wearable computing is summarized to illustrate the current state of smart textile and garment research. Subsequently, the process to build smart textiles from fabric production, sensor and actuator integration, contacting and integration, as well as communication, is summarized with notes and links to relevant chapters of this book. The options and specific needs for evaluating smart textiles are described. The chapter concludes by highlighting current and future research and development challenges for smart textiles.
- Cheng, Jingyuan; Zhou, Bo; Lukowicz, Paul; Seoane, Fernando; Varga, Matija; Mehmann, Andreas; Chabrecek, Peter; Gaschler, Werner; Goenner, Karl; Horter, Hansjürgen; Schneegass, Stefan; Hassib, Mariam; Schmidt, Albrecht; Freund, Martin; Zhang, Rui; Amft, Oliver: Textile Building Blocks: Toward Simple, Modularized, and Standardized Smart Textile. In: Stefan Schneegass, Oliver Amft (Ed.): Smart Textiles. Springer International Publishing, 2017, p. 303-331. doi:10.1007/978-3-319-50124-6_14CitationAbstractDetails
Textiles are pervasive in our life, covering human body and objects, as well as serving in industrial applications. In its everyday use of individuals, smart textile becomes a promising medium for monitoring, information retrieval, and interaction. While there are many applications in sport, health care, and industry, the state-of-the-art smart textile is still found only in niche markets. To gain mass-market capabilities, we see the necessity of generalizing and modularizing smart textile production and application development, which on the one end lowers the production cost and on the other end enables easy deployment. In this chapter, we demonstrate our initial effort in modularization. By devising types of universal sensing fabrics for conductive and non-conductive patches, smart textile construction from basic, reusable components can be made. Using the fabric blocks, we present four types of sensing modalities, including resistive pressure, capacitive, bioimpedance, and biopotential. In addition, we present a multi-channel textile–electronics interface and various applications built on the top of the basic building blocks by ‘cut and sew’ principle.
- Stefan Schneegass, Oliver Amft (Ed.): Smart Textiles - Fundamentals, Design, and Interaction, Springer International Publishing, 2017 (ISBN: 978-3-319-50124-6) (ISSN: 1571-5035). CitationAbstractDetails
From a holistic perspective, this handbook explores the design, development and production of smart textiles and textile electronics, breaking with the traditional silo-structure of smart textile research and development.
Leading experts from different domains including textile production, electrical engineering, interaction design and human-computer interaction (HCI) address production processes in their entirety by exploring important concepts and topics like textile manufacturing, sensor and actuator development for textiles, the integration of electronics into textiles and the interaction with textiles. In addition, different application scenarios, where smart textiles play a key role, are presented too.
Smart Textiles would be an ideal resource for researchers, designers and academics who are interested in understanding the overall process in creating viable smart textiles.
- Schneegass, Stefan; Schmidt, Albrecht; Pfeiffer, Max: Creating user interfaces with electrical muscle stimulation. In: interactions, Vol 24 (2016) No 1, p. 74-77. doi:http://doi.acm.org/10.1145/3019606PDFFull textCitationAbstractDetails
Muscle movement is central to virtually everything we do, be it walking, writing, drawing, smiling, or singing. Even while we're standing still, our muscles are active, ensuring that we keep our balance. In a recent forum  we showed how electrical signals on the skin that reflect muscle activity can be measured. Here, we look at the reverse direction. We explain how muscles can be activated and how movements can be controlled with electrical signals.
- Schneegass, Stefan; Voit, Alexandra: GestureSleeve: using touch sensitive fabrics for gestural input on the forearm for controlling smartwatches. In: Proceedings of the 2016 ACM International Symposium on Wearable Computers (ISWC '16). ACM, New York, USA 2016, p. 108-115. doi:https://doi.org/10.1145/2971763.2971797CitationAbstractDetails
Smartwatches provide quick and easy access to information. Due to their wearable nature, users can perceive the information while being stationary or on the go. The main drawback of smartwatches, however, is the limited input possibility. They use similar input methods as smartphones but thereby suffer from a smaller form factor. To extend the input space of smartwatches, we present GestureSleeve, a sleeve made out of touch enabled textile. It is capable of detecting different gestures such as stroke based gestures or taps. With these gestures, the user can control various smartwatch applications. Exploring the performance of the GestureSleeve approach, we conducted a user study with a running application as use case. In this study, we show that input using the GestureSleeve outperforms touch input on the smartwatch. In the future the GestureSleeve can be integrated into regular clothing and be used for controlling various smart devices.
- Stefan Schneegass, Youssef Oualil; Bulling, Andreas: SkullConduct: Biometric User Identification on Eyewear Computers Using Bone Conduction Through the Skull. In: Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems (CHI '16). ACM, New York, USA 2016, p. 1379-1384. doi:https://doi.org/10.1145/2858036.2858152PDFCitationAbstractDetails
Secure user identification is important for the increasing number of eyewear computers but limited input capabilities pose significant usability challenges for established knowledge-based schemes, such as passwords or PINs. We present SkullConduct, a biometric system that uses bone conduction of sound through the user's skull as well as a microphone readily integrated into many of these devices, such as Google Glass. At the core of SkullConduct is a method to analyze the characteristic frequency response created by the user's skull using a combination of Mel Frequency Cepstral Coefficient (MFCC) features as well as a computationally light-weight 1NN classifier. We report on a controlled experiment with 10 participants that shows that this frequency response is person-specific and stable -- even when taking off and putting on the device multiple times -- and thus serves as a robust biometric. We show that our method can identify users with 97.0% accuracy and authenticate them with an equal error rate of 6.9%, thereby bringing biometric user identification to eyewear computers equipped with bone conduction technology.
- Florian Alt, Stefan Schneegass; Alireza Sahami Shirazi, Mariam Hassib; Bulling, Andreas: Graphical Passwords in the Wild: Understanding How Users Choose Pictures and Passwords in Image-based Authentication Schemes. In: Proceedings of the 17th International Conference on Human-Computer Interaction with Mobile Devices and Services (MobileHCI '15). ACM, New York, USA 2015, p. 316-322. doi:10.1145/2785830.2785882CitationAbstractDetails
Common user authentication methods on smartphones, such as lock patterns, PINs, or passwords, impose a trade-off between security and password memorability. Image-based passwords were proposed as a secure and usable alternative. As of today, however, it remains unclear how such schemes are used in the wild. We present the first study to investigate how image-based passwords are used over long periods of time in the real world. Our analyses are based on data from 2318 unique devices collected over more than one year using a custom application released in the Android Play store. We present an in-depth analysis of what kind of images users select, how they define their passwords, and how secure these passwords are. Our findings provide valuable insights into real-world use of image-based passwords and inform the design of future graphical authentication schemes.
- Max Pfeiffer, Tim Dünte; Stefan Schneegass, Florian Alt; Rohs, Michael: Cruise Control for Pedestrians: Controlling Walking Direction using Electrical Muscle Stimulation. In: In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems (CHI '15). ACM, New York, USA 2015, p. 2505-2514. doi:https://doi.org/10.1145/2702123.2702190CitationAbstractDetails
Pedestrian navigation systems require users to perceive, interpret, and react to navigation information. This can tax cognition as navigation information competes with information from the real world. We propose actuated navigation, a new kind of pedestrian navigation in which the user does not need to attend to the navigation task at all. An actuation signal is directly sent to the human motor system to influence walking direction. To achieve this goal we stimulate the sartorius muscle using electrical muscle stimulation. The rotation occurs during the swing phase of the leg and can easily be counteracted. The user therefore stays in control. We discuss the properties of actuated navigation and present a lab study on identifying basic parameters of the technique as well as an outdoor study in a park. The results show that our approach changes a user's walking direction by about 16°/m on average and that the system can successfully steer users in a park with crowded areas, distractions, obstacles, and uneven ground.
- Henze, Niels: Modeling Distant Pointing for Compensating Systematic Displacements. In: Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems (CHI '15). ACM, New York, USA 2015, p. 4165-4168. doi:https://doi.org/10.1145/2702123.2702332CitationAbstractDetails
Distant pointing at objects and persons is a highly expressive gesture that is widely used in human communication. Pointing is also used to control a range of interactive systems. For determining where a user is pointing at, different ray casting methods have been proposed. In this paper we assess how accurately humans point over distance and how to improve it. Participants pointed at projected targets from 2m and 3m while standing and sitting. Testing three common ray casting methods, we found that even with the most accurate one the average error is 61.3cm. We found that all tested ray casting methods are affected by systematic displacements. Therefore, we trained a polynomial to compensate this displacement. We show that using a user-, pose-, and distant-independent quartic polynomial can reduce the average error by 37.3%.
- Bader, Patrick; Schwind, Valentin; Henze, Niels; Schneegass, Stefan; Broy, Nora; Schmidt, Albrecht: Design and evaluation of a layered handheld 3d display with touch-sensitive front and back. In: NordiCHI '14: Proceedings of the 8th Nordic Conference on Human-Computer Interaction: Fun, Fast, Foundational. ACM, Helsinki, Finland 2014. doi:https://doi.org/10.1145/2639189.2639257PDFCitationDetails
- Stefan Schneegass, Frank Steimle; Andreas Bulling, Florian Alt; Schmidt, Albrecht: SmudgeSafe: geometric image transformations for smudge-resistant user authentication. In: Proceedings of the 2014 ACM International Joint Conference on Pervasive and Ubiquitous Computing (UbiComp '14). ACM, New York, USA 2014, p. 775-786. doi:10.1145/2632048.2636090CitationAbstractDetails
Touch-enabled user interfaces have become ubiquitous, such as on ATMs or portable devices. At the same time, authentication using touch input is problematic, since finger smudge traces may allow attackers to reconstruct passwords. We present SmudgeSafe, an authentication system that uses random geometric image transformations, such as translation, rotation, scaling, shearing, and flipping, to increase the security of cued-recall graphical passwords. We describe the design space of these transformations and report on two user studies: A lab-based security study involving 20 participants in attacking user-defined passwords, using high quality pictures of real smudge traces captured on a mobile phone display; and an in-the-field usability study with 374 participants who generated more than 130,000 logins on a mobile phone implementation of SmudgeSafe. Results show that SmudgeSafe significantly increases security compared to authentication schemes based on PINs and lock patterns, and exhibits very high learnability, efficiency, and memorability.
- Schmidt, Albrecht: Exploiting thermal reflection for interactive systems. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '14). ACM, New York, USA 2014, p. 3483-3492. doi:10.1145/2556288.2557208CitationAbstractDetails
Thermal cameras have recently drawn the attention of HCI researchers as a new sensory system enabling novel interactive systems. They are robust to illumination changes and make it easy to separate human bodies from the image background. Far-infrared radiation, however, has another characteristic that distinguishes thermal cameras from their RGB or depth counterparts, namely thermal reflection. Common surfaces reflect thermal radiation differently than visual light and can be perfect thermal mirrors. In this paper, we show that through thermal reflection, thermal cameras can sense the space beyond their direct field-of-view. A thermal camera can sense areas besides and even behind its field-of-view through thermal reflection. We investigate how thermal reflection can increase the interaction space of projected surfaces using camera-projection systems. We moreover discuss the reflection characteristics of common surfaces in our vicinity in both the visual and thermal radiation bands. Using a proof-of-concept prototype, we demonstrate the increased interaction space for hand-held camera-projection system. Furthermore, we depict a number of promising application examples that can benefit from the thermal reflection characteristics of surfaces.
- Schmidt, Albrecht: Let me catch this!: experiencing interactive 3D cinema through collecting content with a mobile phone. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '14). ACM, New York, USA 2014, p. 1011-1020. doi:10.1145/2556288.2557187Full textCitationAbstractDetails
The entertainment industry is going through a transformation, and technology development is affecting how we can enjoy and interact with the entertainment media content in new ways. In our work, we explore how to enable interaction with content in the context of 3D cinemas. This allows viewers to use their mobile phone to retrieve, for example, information on the artist of the soundtrack currently playing or a discount coupon on the watch the main actor is wearing. We are particularly interested in the user experience of the interactive 3D cinema concept, and how different interactive elements and interaction techniques are perceived. We report on the development of a prototype application utilizing smart phones and on an evaluation in a cinema context with 20 participants. Results emphasize that designing for interactive cinema experiences should drive for holistic and positive user experiences. Interactive content should be tied together with the actual video content, but integrated into contexts where it does not conflict with the immersive experience with the movie.
- Schmidt, Albrecht: FrameBox and MirrorBox: tools and guidelines to support designers in prototyping interfaces for 3D displays. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '14). ACM, New York, USA 2014, p. 2037-2046. doi:https://doi.org/10.1145/2556288.2557183Full textCitationAbstractDetails
In this paper, we identify design guidelines for stereoscopic 3D (S3D) user interfaces (UIs) and present the MirrorBox and the FrameBox, two UI prototyping tools for S3D displays. As auto-stereoscopy becomes available for the mass market we believe the design of S3D UIs for devices, for example, mobile phones, public displays, or car dashboards, will rapidly gain importance. A benefit of such UIs is that they can group and structure information in a way that makes them easily perceivable for the user. For example, important information can be shown in front of less important information. This paper identifies core requirements for designing S3D UIs and derives concrete guidelines. The requirements also serve as a basis for two depth layout tools we built with the aim to overcome limitations of traditional prototyping when sketching S3D UIs. We evaluated the tools with usability experts and compared them to traditional paper prototyping.
- Alt, Florian; Schneegass, Stefan; Auda, Jonas; Rzayev, Rufat; Broy, Nora: Using eye-tracking to support interaction with layered 3D interfaces on stereoscopic displays. In: IUI '14: Proceedings of the 19th international conference on Intelligent User Interfaces. ACM, Haifa, Israel 2014. doi:https://doi.org/10.1145/2557500.2557518PDFCitationDetails
- Kubitza, Thomas; Pohl, Norman; Dingler, Tilman; Schneegass, Stefan; Weichel, Christian; Schmidt, Albrecht: Ingredients for a New Wave of Ubicomp Products. In: IEEE Pervasive Computing, Vol 12 (2013) No 3, p. 5-8. doi:10.1109/MPRV.2013.51Full textCitationAbstractDetails
The emergence of many new embedded computing platforms has lowered the hurdle for creating ubiquitous computing devices. Here, the authors highlight some of the newer platforms, communication technologies, sensors, actuators, and cloud-based development tools, which are creating new opportunities for ubiquitous computing.
- Pascher, Max; Baumeister, Annalies; Schneegass, Stefan; Klein, Barbara; Gerken, Jens: Recommendations for the Development of a Robotic Drinking and Eating Aid - An Ethnographic Study. In: 2021, Human-Computer Interaction – Interact; Springer; Cham (Ed.): INTERACT 2021. Springer, . CitationAbstractDetails
Being able to live independently and self-determined in one's own home is a crucial factor for human dignity and preservation of self-worth. For people with severe physical impairments who cannot use their limbs for every day tasks, living in their own home is only possible with assistance from others. The inability to move arms and hands makes it hard to take care of oneself, e. g. drinking and eating independently. In this paper, we investigate how 15 participants with disabilities consume food and drinks. We report on interviews, participatory observations, and analyzed the aids they currently use. Based on our findings, we derive a set of recommendations that supports researchers and practitioners in designing future robotic drinking and eating aids for people with disabilities.
- Pascher, Max; Baumeister, Annalies; Klein, Barbara; Schneegass, Stefan; Gerken, Jens: Little Helper: A Multi-Robot System in Home Health Care Environments. In: Ecole Nationale de l'Aviation Civile [ENAC] . CitationAbstractDetails
Being able to live independently and self-determined in once own home is a crucial factor for social participation. For people with severe physical impairments, such as tetraplegia, who cannot use their hands to manipulate materials or operate devices, life in their own home is only possible with assistance from others. The inability to operate buttons and other interfaces results also in not being able to utilize most assistive technologies on their own. In this paper, we present an ethnographic field study with 15 tetraplegics to better understand their living environments and needs. Results show the potential for robotic solutions but emphasize the need to support activities of daily living (ADL), such as grabbing and manipulating objects or opening doors. Based on this, we propose Little Helper, a tele-operated pack of robot drones, collaborating in a divide and conquer paradigm to fulfill several tasks using a unique interaction method. The drones can be tele-operated by a user through gaze-based selection and head motions and gestures manipulating materials and applications.