论文标题

抑郁症状从社交媒体文本进行建模:半监督学习方法

Depression Symptoms Modelling from Social Media Text: A Semi-supervised Learning Approach

论文作者

Farruque, Nawshad, Goebel, Randy, Sivapalan, Sudhakar, Zaiane, Osmar

论文摘要

基于社交媒体语言的临床抑郁模型的基本组成部分是抑郁症状检测(DSD)。不幸的是,没有任何DSD数据集反映出自披露的抑郁症人群样本中的临床见解和抑郁症状的分布。 In our work, we describe a Semi-supervised Learning (SSL) framework which uses an initial supervised learning model that leverages 1) a state-of-the-art large mental health forum text pre-trained language model further fine-tuned on a clinician annotated DSD dataset, 2) a Zero-Shot learning model for DSD, and couples them together to harvest depression symptoms related samples from our large self-curated Depression Tweets Repository (DTR)。我们的临床医生注释的数据集是同类数据集中最大的数据集。此外,DTR是由自张开的抑郁用户在两个数据集中的Twitter时间轴中创建的,其中包括从Twitter中检测到用户级抑郁症的最大基准数据集之一。这进一步有助于保留自张开的Twitter用户推文的抑郁症状分布。随后,我们使用收获的数据迭代地重新训练我们的初始DSD模型。我们讨论了该SSL过程的停止标准和局限性,并详细说明了在整个SSL过程中起着至关重要的作用的基础结构。我们表明,我们可以生产最大的数据集,这是同类产品中最大的数据集。此外,受过训练的DSD和抑郁症检测后(DPD)模型的精度明显优于初始版本。

A fundamental component of user-level social media language based clinical depression modelling is depression symptoms detection (DSD). Unfortunately, there does not exist any DSD dataset that reflects both the clinical insights and the distribution of depression symptoms from the samples of self-disclosed depressed population. In our work, we describe a Semi-supervised Learning (SSL) framework which uses an initial supervised learning model that leverages 1) a state-of-the-art large mental health forum text pre-trained language model further fine-tuned on a clinician annotated DSD dataset, 2) a Zero-Shot learning model for DSD, and couples them together to harvest depression symptoms related samples from our large self-curated Depression Tweets Repository (DTR). Our clinician annotated dataset is the largest of its kind. Furthermore, DTR is created from the samples of tweets in self-disclosed depressed users Twitter timeline from two datasets, including one of the largest benchmark datasets for user-level depression detection from Twitter. This further helps preserve the depression symptoms distribution of self-disclosed Twitter users tweets. Subsequently, we iteratively retrain our initial DSD model with the harvested data. We discuss the stopping criteria and limitations of this SSL process, and elaborate the underlying constructs which play a vital role in the overall SSL process. We show that we can produce a final dataset which is the largest of its kind. Furthermore, a DSD and a Depression Post Detection (DPD) model trained on it achieves significantly better accuracy than their initial version.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源