Table 2: Evaluation Metrics of the Double-layer BiLSTM-Attention Model [Owner-draw].
Classifications Recall Precision Accuracy F1-measure
RiskοΌ0οΌ1οΌ2οΌ3οΌ
/
/
62% 0.438
Suicidal ideation PresenceοΌ0οΌ1οΌ
0.948 0.924 90% 0.936
Severe Suicidal IdeationοΌ0οΌ1οΌ
0.956 0.835 84% 0.892
The above table shows that the model can classify
the presence of suicidal ideation with high Recall and
Accuracy. In contrast, the Accuracy has declined by
nearly 10% in determining the severity of suicidal
ideation. Therefore, the model in this paper cannot
distinguish the slight differences between the low and
moderate levels. Moreover, it usually classifies users
with low suicidal ideation as moderate.
While this model performed well in suicidal idea-
tion classification, the Accuracy and F1_macro value
are relatively lower in the four-classification task due
to the model's design and dataset applied. In
CLPsych2019, fewer users without suicidal ideation
and more in the remaining three categories. Thus, it is
easier to determine suicidal ideation as the accuracy
rate is higher under this classification criterion.
4 CONCLUSION
The task of this paper is to collect posts on social me-
dia and classify them into four categories: no, low,
moderate, and severe, according to the level of sui-
cidal ideation.
To complete this task, the author set up a two-
layer BiLSTM-Attention Model. The model first
cleans the data, removing noisy and meaningless data.
Then it sorts each user's posts by timestamp and ob-
tains the word vectors through BERT, the pre-trained
language model, which provides a better representa-
tion than the word embedding model. Next, the re-
sults are fed into the model designed in this paper, uti-
lising BiLSTM to capture long-distance dependency
(LDD) and the bidirectional information as well as
the Attention mechanism, which allows the model to
focus on the core information of the post. The author
obtains the reasonable vector representations of post
through the first layer of BiLSTM, the representations
of the users in the second layer, and the assessment
results from the classifiers. The Accuracy of the
model reaches 62%, and the macro-f1 value is 0.438.
Specifically, the Attention mechanism improves the
Macro_ Fi value by 16%. Compared with other mod-
els feeding the results directly to the classifiers after
BERT, this model improves the Macro_F1 value by
29%. As a result, the model in this paper performs
better on the processing dataset CLPsych2019.
What is more, to better assess the model's perfor-
mance, the author introduces two classifications: the
presence of suicidal ideation and the severity of sui-
cidal ideation. Results show that the distinguishabil-
ity of low and moderate suicidal risk is relatively low,
which needs to be improved in the future.
REFERENCES
Maron M E. Automatic indexing: an experimental in-
quiry[J]. Journal of the ACM (JACM), 1961, 8(3): 404-
417.
Rabiner L, Juang B. An introduction to hidden Markov
models[J]. IEEE ASSP Magazine, 1986, 3(1): 4-16.
Khalil Alsmadi M, Omar K B, Noah S A, et al. Performance
comparison of multi-layer perceptron (Back Propaga-
tion, Delta Rule and Perceptron) algorithms in neural
networks[C]//2009 IEEE International Advance Com-
puting Conference. IEEE, 2009: 296-299.
Ji S, Yu C P, Fung S, et al. Supervised learning for suicidal
ideation detection in online user content[J]. Complex-
ity, 2018.
Kalchbrenner N, Grefenstette E, Blunsom P. A convolu-
tional neural network for modelling sentences[J]. arXiv
preprint arXiv:1404.2188, 2014.
Cao L, Zhang H, Feng L. Building and Using Personal
Knowledge Graph to Improve Suicidal Ideation Detec-
tion on Social Media [J]. IEEE Transactions on Multi-
media, 2020.