Qualitative Coding Analysis through Open-Source Large Language Models: A User Study and Design Recommendations
Abstract
Qualitative data analysis is labor-intensive, yet the privacy risks associated with commercial Large Language Models (LLMs) often preclude their use in sensitive research. To address this, we introduce ChatQDA, an on-device framework powered by open-source LLMs designed for privacy-preserving open coding. Our mixed-methods user study reveals that while participants rated the system highly for usability and perceived efficiency, they exhibited "conditional trust", valuing the tool for surface-level extraction while questioning its interpretive nuance and consistency. Furthermore, despite the technical security of local deployment, participants reported epistemic uncertainty regarding data protection, suggesting that invisible security measures are insufficient to foster trust. We conclude with design recommendations for local-first analysis tools that prioritize verifiable privacy and methodological rigor.
Source: arXiv:2602.18352v1 - http://arxiv.org/abs/2602.18352v1 PDF: https://arxiv.org/pdf/2602.18352v1 Original Link: http://arxiv.org/abs/2602.18352v1