ORIGINAL RESEARCH article
Front. Big Data
Sec. Machine Learning and Artificial Intelligence
Depression Detection through Dual-Stream Modeling with Large Language Models: A Fusion-Based Transfer Learning Framework Integrating BERT and T5 Representations
Provisionally accepted- 1The University of Newcastle School of Information and Physical Sciences, Callaghan, Australia
- 2School of Automation, Guangdong Polytechnic Normal University, Guangzhou, China
- 3Universiti Putra Malaysia Fakulti Kejuruteraan, Serdang, Malaysia
- 4Universiti Malaya Department of Electrical Engineering, Federal Territory of Kuala Lumpur, Malaysia
Select one of your emails
You have multiple emails registered with Frontiers:
Notify me on publication
Please enter your email address:
If you already have an account, please login
You don't have a Frontiers account ? You can register here
Millions of people around the world suffer from depression. While early diagnosis is essential for timely intervention, it remains a significant challenge due to limited access to clinically diagnosed data and privacy restrictions on mental health records. These limitations hinder the training of robust AI models for depression detection. To tackle this, this article proposes a parallel transfer learning framework for depression detection that integrates BERT and T5 through a fusion mechanism, combining the complementary advantages of these two large language models. By integrating their semantic embeddings, the method captures a broader range of linguistic cues from transcribed speech. These embeddings are processed through a model with two parallel branches: a one-dimensional convolutional neural network and a dense neural network are used to construct each branch for preliminary prediction, which are then fused for final prediction. Evaluations on the E-DAIC dataset demonstrate that the proposed method outperforms baseline models, achieving a 3.0% increase in accuracy (91.3%), a 6.9% increase in precision (95.2%), and a 1.7% improvement in F1-score (90.0%). The experimental results verify the effectiveness of BERT and T5 fusion in enhancing depression detection performance and highlight the potential of transfer learning for scalable and privacy-conscious mental health applications.
Keywords: 1DCNN, BERT, Depression, E-DAIC, T5, text, Transfer Learning, transformer
Received: 25 Jun 2025; Accepted: 30 Dec 2025.
Copyright: © 2025 Wang, Zhang, Kamil, Renner, Abdul RAHMAN AL-HADDAD, Ibrahim and Zhao. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
* Correspondence:
Weijia Zhang
Raja Kamil
Zhen Zhao
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.
