Abstract
This study compares the performance of recurrent neural networks (RNNs) and transformer-based models (DistilBERT) in classifying utterances as dialogue acts. The results show that transformers consistently outperform RNNs, highlighting their usefulness in coding small group interaction. Furthermore, the study explores the impact of incorporating context, in the form of preceding and following utterances. The findings reveal that adding context leads to modest improvements in model performance. Moreover, in some cases, adding context can lead to a slight decrease in performance. The study discusses the implications of these findings for small group researchers employing AI models for text classification tasks.
| Original language | English |
|---|---|
| Pages (from-to) | 864-893 |
| Number of pages | 30 |
| Journal | Small Group Research |
| Volume | 56 |
| Issue number | 5 |
| DOIs | |
| State | Published - 2025 |
Bibliographical note
Publisher Copyright:© The Author(s) 2025.
Funding
The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This research used the Delta advanced computing and data resource which is supported by the National Science Foundation (award OAC 2005572) and the State of Illinois. Delta is a joint effort of the University of Illinois Urbana-Champaign and its National Center for Supercomputing Applications.
| Funders | Funder number |
|---|---|
| University of Illinois, Urbana-Champaign | |
| National Science Foundation Arctic Social Science Program | OAC 2005572 |
Keywords
- communication
- content analysis
- interaction analysis
- meetings
ASJC Scopus subject areas
- Social Psychology
- Applied Psychology