%Aigaion2 BibTeX export from Idiap Publications %Saturday 21 December 2024 06:03:54 PM @ARTICLE{VILLATORO-TELLO_COGNITIVECOMPUTATION_2021, author = {Villatoro-Tello, Esa{\'{u}} and Parida, Shantipriya and Kumar, Sajit and Motlicek, Petr}, keywords = {BERT, deep learning, Natural language processing, Operant Motive Test, Psycholinguistics, Supervised Autoencoders}, projects = {Idiap}, month = jul, title = {Applying Attention-Based Models for Detecting Cognitive Processes and Mental Health Conditions}, journal = {Cognitive Computation}, year = {2021}, pages = {18}, url = {https://link.springer.com/article/10.1007/s12559-021-09901-1}, doi = {10.1007/s12559-021-09901-1}, crossref = {VILLATORO-TELLO_Idiap-RR-01-2022}, abstract = {According to the psychological literature, implicit motives allow for the characterization of behavior, subsequent success, and long-term development. Contrary to personality traits, implicit motives are often deemed to be rather stable personality characteristics. Normally, implicit motives are obtained by Operant Motives, unconscious intrinsic desires measured by the Operant Motive Test (OMT). The OMT test requires participants to write freely descriptions associated with a set of provided images and questions. In this work, we explore different recent machine learning techniques and various text representation techniques for facing the problem of the OMT classification task. We focused on advanced language representations (e.g, BERT, XLM, and DistilBERT) and deep Supervised Autoencoders for solving the OMT task. We performed an exhaustive analysis and compared their performance against fully connected neural networks and traditional support vector classifiers. Our comparative study highlights the importance of BERT which outperforms the traditional machine learning techniques by a relative improvement of 7.9\%. In addition, we performed an analysis of how the BERT attention mechanism is being modified. Our findings indicate that the writing style features acquire higher importance at the moment of accurately identifying the different OMT categories. This is the first time that a study to determine the performance of different transformer-based architectures in the OMT task is performed. Similarly, our work propose, for the first time, using deep supervised autoencoders in the OMT classification task. Our experiments demonstrate that transformer-based methods exhibit the best empirical results, obtaining a relative improvement of 7.9\% over the competitive baseline suggested as part of the GermEval 2020 challenge. Additionally, we show that features associated with the writing style are more important than content-based words. Some of these findings show strong connections to previously reported behavioral research on the implicit psychometrics theory.} } crossreferenced publications: @TECHREPORT{VILLATORO-TELLO_Idiap-RR-01-2022, author = {Villatoro-Tello, Esa{\'{u}} and Parida, Shantipriya and Kumar, Sajit and Motlicek, Petr}, keywords = {BERT, deep learning, Natural language processing, Operant Motive Test, Psycholinguistics, Supervised Autoencoders}, projects = {Idiap}, month = {1}, title = {Applying Attention Based Models for Detecting Cognitive Processes and Mental Health Conditions}, type = {Idiap-RR}, number = {Idiap-RR-01-2022}, year = {2022}, institution = {Idiap}, note = {Preprint Accepted in Cognitive Computation Journal, Springer (IF 4.3)}, crossref = {VILLATORO-TELLO_COGNITIVECOMPUTATION_2021}, abstract = {Objective: According to the psychological literature, implicit motives allow for the characterization of behavior, subsequent success, and long-term development. Contrary to personality traits, implicit motives are often deemed to be rather stable personality characteristics. Normally, implicit motives are obtained by Operant Motives, unconscious intrinsic desires measured by the Operant Motive Test (OMT). The OMT test requires participants to write freely descriptions associated with a set of provided images and questions. In this work, we explore different recent machine learning techniques and various text representation techniques for facing the problem of the OMT classification task. Methods: We focused on advanced language representations (e.g, BERT, XLM, and DistilBERT) and deep Supervised Autoencoders for solving the OMT task. We performed an exhaustive analysis and compared their performance against fully connected neural networks and traditional support vector classifiers. Results: Our comparative study highlights the importance of BERT which outperforms the traditional machine learning techniques by a relative improvement of 7.9\%. In addition, we performed an analysis of how the BERT attention mechanism is being modified. Our findings indicate that the writing style features acquire higher importance at the moment of accurately identifying the different OMT categories. Conclusions: This is the first time that a study to determine the performance of different transformers-based architectures in the OMT task is performed. Similarly, our work propose, for the first time, using Deep Supervised Autoencoders in the OMT classification task. Our experiments demonstrate that transformers-based methods exhibit the best empirical results, obtaining a relative improvement of 7.9\% over the competitive baseline suggested as part of the GermEval 2020 challenge. Additionally, we show that features associated with the writing style are more important than content-based words. Some of these findings show strong connections to previously reported behavioral research on the implicit psychometrics theory.}, pdf = {https://publications.idiap.ch/attachments/reports/2021/VILLATORO-TELLO_Idiap-RR-01-2022.pdf} }