@article{schmidmaier2025using, title = {Using Nonverbal Cues in Empathic Multi-Modal LLM-Driven Chatbots for Mental Health Support}, author = {Matthias Schmidmaier and Jonathan Rupp and Cedrik Harrich and Sven Mayer}, url = {https://sven-mayer.com/wp-content/uploads/2025/07/schmidmaier2025using.pdf}, doi = {10.1145/3743719}, year = {2025}, date = {2025-09-01}, urldate = {2025-09-01}, journal = {Proc. ACM Hum.-Comput. Interact.}, volume = {9}, number = {5}, publisher = {Association for Computing Machinery}, address = {New York, NY, USA}, abstract = {Despite their popularity in providing digital mental health support, mobile conversational agents primarily rely on verbal input, which limits their ability to respond to emotional expressions. We therefore envision using the sensory equipment of today\'s devices to increase the nonverbal, empathic capabilities of chatbots. We initially validated that multi-modal LLMs (MLLM) can infer emotional expressions from facial expressions with high accuracy. In a user study (N=200), we then investigated the effects of such multi-modal input on response generation and perceived system empathy in emotional support scenarios. We found significant effects on cognitive and affective dimensions of linguistic expression in system responses, yet no significant increases in perceived empathy. Our research demonstrates the general potential of using nonverbal context to adapt LLM response behavior, providing input for future research on augmented interaction in empathic MLLM-based systems.}, keywords = {}, pubstate = {published}, tppubtype = {article} }