The role of using eyetracking technology in psychodiagnostics: systematic review

Cover Page

Cite item

Full Text

Open Access Open Access
Restricted Access Access granted
Restricted Access Subscription Access

Abstract

The use of the eye movement registration and gaze fixation system — eytracing (eytracing — TA technology) as a means of psychodiagnostics is a new and very promising diagnostic technique. It allows for a deeper understanding of cognitive processes, emotional reactions and behavioral characteristics, which opens up new opportunities in the diagnosis and study of mental states at an interdisciplinary level. The purpose of the study is to assess the state and possibilities of using psychodiagnostic TA within the framework of a systematic analysis of scientific data. Materials and methods. The literature was searched in electronic databases of scientific data PubMed.gov, Elsevier.com, Google Scholar.com, Elibrary.ru. The analysis of the data obtained focused on works published between 2014 and 2024. Results. After the audit procedures of 549 works for duplication and compliance criteria, 31 studies were selected. Of these, 15 (46.9%) are devoted to evaluating the use of TA in clinical psychology, 10 (31.2%) — cognitive psychology and 7 (21.9%) — social psychology. At the same time, in most cases, the work was aimed at diagnostic optimization of TA — 15 (46.9%), in 14 (43.8%) it had the character of a scientific justification for the expediency of using TA in the framework of psychodiagnostics, and in 3 (9.4%) it provided data on the validity of TA as a psychodiagnostic tool. Conclusions. The analysis of scientific data revealed that despite the relative novelty of TA as a psychodiagnostic tool, the most popular areas have been identified that focus on the assessment of autism spectrum disorders, cognitive function and neuropsychic profile. However, despite the prevailing popularity of the study of TA as a psychodiagnostic tool, to date the world scientific community has not been able to formulate a standard or create a unified protocol for the use of this technology for the assessment of psychopathological conditions.

About the authors

E. D. Shagina

N.I. Pirogov Russian National Research Medical University of the Ministry of Health of the Russian Federation

Author for correspondence.
Email: shaginenok@gmail.com

Institute of Clinical Psychology and Social Work, Head of the Laboratory of Practical Skills, Assistant of the Department of Clinical Psychology

Russian Federation, 125047, Moscow, 1st Miusskaya str., 22/24, bldn. 2

S. A. Matveicheva

N.I. Pirogov Russian National Research Medical University of the Ministry of Health of the Russian Federation

Email: markshakhov03@yandex.ru

Institute of Clinical Psychology and Social Work, Student

Russian Federation, 125047, Moscow, 1st Miusskaya str., 22/24, bldn. 2

Yu. V. Nekhaeva

N.I. Pirogov Russian National Research Medical University of the Ministry of Health of the Russian Federation

Email: markshakhov03@yandex.ru

Institute of Clinical Psychology and Social Work, Student

Russian Federation, 125047, Moscow, 1st Miusskaya str., 22/24, bldn. 2

S. E. Lukina

N.I. Pirogov Russian National Research Medical University of the Ministry of Health of the Russian Federation

Email: markshakhov03@yandex.ru

Institute of Clinical Psychology and Social Work, Student

Russian Federation, 125047, Moscow, 1st Miusskaya str., 22/24, bldn. 2

M. V. Shakhov

N.I. Pirogov Russian National Research Medical University of the Ministry of Health of the Russian Federation

Email: markshakhov03@yandex.ru

Institute of Clinical Psychology and Social Work, Student

Russian Federation, 125047, Moscow, 1st Miusskaya str., 22/24, bldn. 2

I. E. Kondratiev

N.I. Pirogov Russian National Research Medical University of the Ministry of Health of the Russian Federation

Email: markshakhov03@yandex.ru

Institute of Clinical Psychology and Social Work, Student

Russian Federation, 125047, Moscow, 1st Miusskaya str., 22/24, bldn. 2

A. A. Shishkin

N.I. Pirogov Russian National Research Medical University of the Ministry of Health of the Russian Federation

Email: doc.shishkin@mail.ru

Z.P. Solovyov Institute of Preventive Medicine, Candidate of Medical Sciences, researcher at the Laboratory of the Latest technologies of health-improving motor activity

Russian Federation, 129110, Moscow, Panteleevskaya str., 10

N. A. Demin

N.I. Pirogov Russian National Research Medical University of the Ministry of Health of the Russian Federation

Email: deminnic1@gmail.com

Z.P. Solovyov Institute of Preventive Medicine, Junior researcher at the Laboratory of the Latest technologies of recreational motor activity

Russian Federation, 129110, Moscow, Panteleevskaya str., 10

References

  1. Bessonova Ju.V., Oboznov A.A. Ajtreking v diagnostike pravdy–lzhi. Institut psihologii Rossijskoj akademii nauk. Organizacionnaja psihologija i psihologija truda. 2019. V. 4. № 1. P. 46–86. (In Russian)
  2. Bojko V.V. Pravila jemocional’nogo povedenija. Saint Petersburg. 1998. 88 p. (In Russian)
  3. Zotova N.E., Zotov M.V., Korotkova I.C. Neproizvol’naja integracija rezul’tatov perceptivnoj aktivnosti drugogo individa v sobstvennuju perceptivnuju aktivnost’. Kul’turno-istoricheskaja psihologija. 2022. V. 18. № 2. P. 73–81. https://doi.org/10.17759/chp.2022180208 (In Russian)
  4. Kajdalova D.A., Kuznecova A.A., Nikitina E.A. Osobennosti glazodvigatel’nyh reakcij pri vosprijatii statichnyh izobrazhenij u junoshej i devushek so sklonnost’ju k igrovoj komp’juternoj zavisimosti. Izvestija Jugo-Zapadnogo gosudarstvennogo universiteta. Serija: lingvistika i pedagogika. 2023. V. 13. № 4. P. 164–175. https://doi.org/10.21869/2223-151X-2023-13-4-164-175 (In Russian)
  5. Lihacheva Je.V., Nikolaeva L.P., Ognev A.S. et al. Okulograficheskie pokazateli preobladanija pozitivnyh ili negativnyh jemocional’nyh sostojanij. Chelovecheskij kapital. 2020. № 9(141). P. 188–199. https://doi.org/10.25629/HC.2020.09.17. EDN KVKPSV (In Russian)
  6. Ognev A.S., Lihacheva Je.V., Nikolaeva L.P. et al. Vozmozhnosti ispol’zovanija izobrazhenij pesochnyh kompozicij v okulometricheskoj diagnostike psihologicheskih osobennostej lichnosti. Chelovecheskij kapital. 2022. № 6(162). P. 91–95. https://doi.org/10.25629/HC.2022.06.11. EDN WMJBBM (In Russian)
  7. Pochinkova P.A., Gorbatova M.A., Narkevich A.N. et al. Obnovlennye kratkie rekomendacii po podgotovke i predstavleniju sistematicheskih obzorov: chto novogo v PRISMA-2020? Morskaja medicina. 2022. V. 8. № 2. P. 88–101. https://doi.org/10.22328/2413-5747-2022-8-2-88-101 (In Russian)
  8. Prirodova O.F., Lazarenko V.A., Nikishina V.B. et al. Ocenka ustanovochnogo komponenta kognitivnoj zadachi po strukturno-dinamicheskim harakteristikam dvizhenija glaz prepodavatelja vuza. Medicinskoe obrazovanie i professional’noe razvitie. 2019. V. 10. № 1. P. 73–83. https://doi.org/10.24411/2220-8453-2019-11008 (In Russian)
  9. Sagalakova O.A., Truevcev D.V., Zhirnova O.V. et al. Jeffekt social’noj trevogi i styda v razvitii golosov: metakognitivnaja obrabotka neblagoprijatnogo opyta otnoshenij i bditel’nost’ k social’nym ugrozam. Klinicheskaja i special’naja psihologija. 2023. V. 12. № 2. P. 25–53. https://doi.org/10.17759/cpse.2023120202 (In Russian)
  10. Sapronov F.A., Makarov I.M., Gorbunova E.S. Kategorizacija v gibridnom poiske: issledovanie s ispol’zovaniem registracii dvizhenij glaz. Jeksperimental’naja psihologija. 2023. V. 16. № 3. P. 121–138. https://doi.org/10.17759/exppsy.2023160308 (In Russian)
  11. Solov’eva V.A., Venig S.B., Belyh T.V. Analiz okulomotornoj aktivnosti, nabljudaemoj pri izuchenii obrazovatel’nogo materiala s jekrana. Integracija obrazovanija. 2021. V. 25. № 1. P. 91–109. https://doi.org/10.15507/1991-9468.102.025.202101.091-109 (In Russian)
  12. Jaroshenko E.I. Primenenie tehnologii ajtrekinga dlja vyjavlenija social’no-psihologicheskih osobennostej jemocional’nogo vygoranija lichnosti. Organizacionnaja psihologija. 2019. № 9(1). P. 96–115. (In Russian)
  13. Ahmed I.A., Senan E.M., Rassem T.H. et al. Eye tracking-based diagnosis and early detection of autism spectrum disorder using machine learning and deep learning techniques. Electronics. 2022. № 11(4). Art. 530. https://doi.org/10.3390/electronics11040530
  14. Alblas M.C., Mollen S., Fransen M.L. et al. Food at first sight: Visual attention to palatable food cues on TV and subsequent unhealthy food intake in unsuccessful restrained eaters. Appetite. 2020. № 147. Art. 104574. https://doi.org/10.1016/j.appet.2019.104574
  15. Armstrong T., Olatunji B.O. Eye tracking of attention in the affective disorders: a meta-analytic review and synthesis. Clin Psychol Rev. 2012. V. 32. № 8. P. 704–723. https://doi.org/10.1016/j.cpr.2012.09.004
  16. Baker R.S., Fields H.W Jr., Beck F.M. et al. Objective assessment of the contribution of dental esthetics and facial attractiveness in men via eye tracking. Am J Orthod Dentofacial Orthop. 2018. № 153(4). P. 523–533. https://doi.org/10.1016/j.ajodo.2017.08.016
  17. Baron M., Hunfalvay M. Eye tracking technology detects traumatic brain injury with sensitivity and specificity. Practical Neurology. 2023. https://practicalneurology.com/news/eye-tracking-technology-detects-traumatic-brain-injury-with-sensitivity-and-specificity
  18. Brunet M., Taddei A., Py J. et al. Social contact, own-group recognition bias and visual attention to faces. Br J Psychol. 2023. V. 114 № 1. P. 112–133. https://doi.org/10.1111/bjop.12603
  19. Brunyé T.T., Drew T., Weaver D.L. et al. A review of eye tracking for understanding and improving diagnostic interpretation. Cogn. Research. 2019. № 4(7). https://doi.org/10.1186/s41235-019-0159-2
  20. Cain M.S., Adamo S.H., Mitroff S.R. A taxonomy of errors in multiple-target visual search. Visual Cognition. 2013. № 21. P. 899–921. https://doi.org/10.1080/13506285.2013.843627
  21. Cipresso P., Mauri M., Semonella M. et al. Looking at one’s self through Facebook increases mental stress: a computational psychometric analysis by using Eye-Tracking and psychophysiology. Cyberpsychol Behav Soc Netw. 2019. № 22(5). P. 307–314. https://doi.org/10.1089/cyber.2018.0602
  22. Di Stasi L.L., Díaz-Piedra C., Ruiz-Rabelo J.F. et al. Quantifying the cognitive cost of laparo-endoscopic single-site surgeries: Gaze-based indices. Appl Ergon. 2017. № 65. P. 168–174. https://doi.org/10.1016/j.apergo.2017.06.008
  23. Jenner L.A., Farran E.K., Welham A. et al. The use of eye-tracking technology as a tool to evaluate social cognition in people with an intellectual disability: a systematic review and meta-analysis. J Neurodev Disord. 2023. № 15(1). Art. 42. https://doi.org/10.1186/s11689-023-09506-9
  24. Jones W., Klin A. Eye-tracking device predicts expert autism diagnoses. HCP Live. 2023. https://www.hcplive.com/view/eye-tracking-device-predicts-expert-autism-diagnoses
  25. Kerr-Gaffney J., Jones E., Mason L. et al. Social attention in anorexia nervosa and autism spectrum disorder: Role of social motivation. Autism. 2022. № 26(7). P. 1641–1655. https://doi.org/10.1177/13623613211060593
  26. Koutsogiorgi C.C., Michaelides M.P. Response tendencies due to item wording using eye-tracking methodology accounting for individual differences and item characteristics. Behav Res Methods. 2022. № 54(5). P. 2252–2270. https://doi.org/10.3758/s13428-021-01719-x
  27. Kovesdi C., Spielman Z., LeBlanc K. et al. Application of eye tracking for measurement and evaluation in human factors studies in control room modernization. Nuclear Technology. 2018. № 202(2–3). P. 220–229. https://doi.org/10.1080/00295450.2018.1455461
  28. Lee J.E., Kim S.H., Shin S.K. et al. Attentional engagement for pain-related information among individuals with chronic pain: the role of pain catastrophizing. Pain Res Manag. 2018. Art. 6038406. https://doi.org/10.1155/2018/6038406
  29. Lee M., Nayar K., Maltman N. et al. Understanding social communication differences in autism spectrum disorder and first-degree relatives: a study of looking and speaking. J Autism Dev Disord. 2020. № 50(6). P. 2128–2141. https://doi.org/10.1007/s10803-019-03969-3
  30. Lev A., Braw Y., Elbaum T. et al. Eye tracking during a continuous performance test: utility for assessing adhd patients. J Atten Disord. 2022. № 26(2). P. 245–255. https://doi.org/10.1177/1087054720972786
  31. Li N., Yang X., Du W. et al. Exploratory research on key technology of human-computer interactive 2.5-minute fast digital early warning for mild cognitive impairment. Comput Intell Neurosci. 2022. Art. 2495330. https://doi.org/10.1155/2022/2495330
  32. Lim J.Z., Mountstephens J., Teo J. Emotion recognition using eye-tracking: taxonomy, review and current challenges. Sensors (Basel). 2020. № 20(8). Art. 2384. https://doi.org/10.3390/s20082384
  33. Liu Q., Yang X., Chen Z., et al. Using synchronized eye movements to assess attentional engagement. Psychol Res. 2023. № 87(7). P. 2039–2047. https://doi.org/10.1007/s00426-023-01791-2
  34. MacNeill L.A., Fu X., Buss K.A. et al. Do you see what I mean?: Using mobile eye tracking to capture parent-child dynamics in the context of anxiety risk. Dev Psychopathol. 2022. № 34(3). P. 997–1012. https://doi.org/10.1017/S0954579420001601
  35. Malsert J., Palama A., Gentaz E. Emotional facial perception development in 7, 9 and 11 year-old children: The emergence of a silent eye-tracked emotional other-race effect. PLoS One. 2020. № 15(5). Art. e0233008. https://doi.org/10.1371/journal.pone.0233008
  36. Matsuda Y., Fedotov D., Takahashi Y. et al. EmoTour: Estimating emotion and satisfaction of users based on behavioral cues and audiovisual data. Sensors (Basel). 2018. № 18(11). Art. 3978. https://doi.org/10.3390/s18113978
  37. Metternich B., Gehrer N.A., Wagner K. et al. Eye-movement patterns during emotion recognition in focal epilepsy: An exploratory investigation. Seizure. 2022. № 100. P. 95–102. https://doi.org/10.1016/j.seizure.2022.06.018
  38. Nayar K., Gordon P.C., Martin G.E. et al. Links between looking and speaking in autism and first-degree relatives: insights into the expression of genetic liability to autism. Mol Autism. 2018. V. 9. № 51. https://doi.org/10.1186/s13229-018-0233-5
  39. Nayar K., Kang X., Xing J. et al. A cross-cultural study showing deficits in gaze-language coordination during rapid automatized naming among individuals with ASD. Sci Rep. 2021. № 11(1). Art. 13401. https://doi.org/10.1038/s41598-021-91911-y
  40. Noyes B., Biorac A., Vazquez G. et al. Eye-tracking in adult depression: protocol for a systematic review and meta-analysis. BMJ Open. 2023. № 13(6). Art. e069256. https://doi.org/10.1136/bmjopen-2022-069256
  41. Ognev A.S., Likhacheva E.V., Nikolaeva L.P. Oculographic manifestations of situative installations, cognitive schemes and life scenarios as different components of character. Human capital. 2020. V. 141. № 9. P. 200–210. doi: https://doi.org/10.25629/HC.2020.09.18
  42. Page M.J., McKenzie J.E., Bossuyt P.M. et al. The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. BMJ. 2021. V. 372. № 71. https://doi.org/10.1136/bmj.n71
  43. Penedo T., Polastri P.F., Rodrigues S.T. et al. Influence of obstacle color on locomotor and gaze behaviors during obstacle avoidance in people with Parkinson’s disease. Exp Brain Res. 2018. № 236(12). P. 3319–3325. https://doi.org/10.1007/s00221-018-5385-1
  44. Putra P.U., Shima K., Alvarez S.A. et al. Identifying autism spectrum disorder symptoms using response and gaze behavior during the Go/NoGo game CatChicken. Sci Rep. 2021. № 11(1). Art. 22012. https://doi.org/10.1038/s41598-021-01050-7
  45. Samadani U., Spinner R.J., Dynkowski G. et al. Eye tracking for classification of concussion in adults and pediatrics. Front. Neurol. 2022. V. 13. Art. 1039955. https://doi.org/10.3389/fneur.2022.1039955
  46. Tahri Sqalli M., Aslonov B., Gafurov M. et al. Eye tracking technology in medical practice: a perspective on its diverse applications. Front Med Technol. 2023 № 5. Art. 1253001. https://doi.org/10.3389/fmedt.2023.1253001
  47. Tokarskaya L.V., Kotyusov A.I. Diagnostic tools for children with severe multiple developmental disorders: eye-tracking and electroencephalogram approaches. Lurian Journal. 2021. № 2(3). P. 45–61. https://doi.org/10.15826/Lurian.2021.2.3.5
  48. Tomer E., Lupu T., Golan L. et al. Eye tracking as a mean to detect feigned cognitive impairment in the word memory test. Appl Neuropsychol Adult. 2020. № 27(1). P. 49–61. https://doi.org/10.1080/23279095.2018.1480483
  49. Wegner-Clemens K., Rennig J, Beauchamp M.S. A relationship between Autism-Spectrum Quotient and face viewing behavior in 98 participants. PLoS One. 2020. № 15(4). Art. e0230866. https://doi.org/10.1371/journal.pone.0230866
  50. Wolf A., Tripanpitak K., Umeda S. et al. Eye-tracking paradigms for the assessment of mild cognitive impairment: a systematic review. Front Psychol. 2023. № 14. Art. 1197567. https://doi.org/10.3389/fpsyg.2023.1197567
  51. Zhu L., Lv J. Review of studies on user research based on EEG and eye tracking. Applied Sciences. 2023. № 13(11). Art. 6502. https://doi.org/10.3390/app13116502

Supplementary files

Supplementary Files
Action
1. JATS XML

Copyright (c) 2025 Russian Academy of Sciences

Согласие на обработку персональных данных с помощью сервиса «Яндекс.Метрика»

1. Я (далее – «Пользователь» или «Субъект персональных данных»), осуществляя использование сайта https://journals.rcsi.science/ (далее – «Сайт»), подтверждая свою полную дееспособность даю согласие на обработку персональных данных с использованием средств автоматизации Оператору - федеральному государственному бюджетному учреждению «Российский центр научной информации» (РЦНИ), далее – «Оператор», расположенному по адресу: 119991, г. Москва, Ленинский просп., д.32А, со следующими условиями.

2. Категории обрабатываемых данных: файлы «cookies» (куки-файлы). Файлы «cookie» – это небольшой текстовый файл, который веб-сервер может хранить в браузере Пользователя. Данные файлы веб-сервер загружает на устройство Пользователя при посещении им Сайта. При каждом следующем посещении Пользователем Сайта «cookie» файлы отправляются на Сайт Оператора. Данные файлы позволяют Сайту распознавать устройство Пользователя. Содержимое такого файла может как относиться, так и не относиться к персональным данным, в зависимости от того, содержит ли такой файл персональные данные или содержит обезличенные технические данные.

3. Цель обработки персональных данных: анализ пользовательской активности с помощью сервиса «Яндекс.Метрика».

4. Категории субъектов персональных данных: все Пользователи Сайта, которые дали согласие на обработку файлов «cookie».

5. Способы обработки: сбор, запись, систематизация, накопление, хранение, уточнение (обновление, изменение), извлечение, использование, передача (доступ, предоставление), блокирование, удаление, уничтожение персональных данных.

6. Срок обработки и хранения: до получения от Субъекта персональных данных требования о прекращении обработки/отзыва согласия.

7. Способ отзыва: заявление об отзыве в письменном виде путём его направления на адрес электронной почты Оператора: info@rcsi.science или путем письменного обращения по юридическому адресу: 119991, г. Москва, Ленинский просп., д.32А

8. Субъект персональных данных вправе запретить своему оборудованию прием этих данных или ограничить прием этих данных. При отказе от получения таких данных или при ограничении приема данных некоторые функции Сайта могут работать некорректно. Субъект персональных данных обязуется сам настроить свое оборудование таким способом, чтобы оно обеспечивало адекватный его желаниям режим работы и уровень защиты данных файлов «cookie», Оператор не предоставляет технологических и правовых консультаций на темы подобного характера.

9. Порядок уничтожения персональных данных при достижении цели их обработки или при наступлении иных законных оснований определяется Оператором в соответствии с законодательством Российской Федерации.

10. Я согласен/согласна квалифицировать в качестве своей простой электронной подписи под настоящим Согласием и под Политикой обработки персональных данных выполнение мною следующего действия на сайте: https://journals.rcsi.science/ нажатие мною на интерфейсе с текстом: «Сайт использует сервис «Яндекс.Метрика» (который использует файлы «cookie») на элемент с текстом «Принять и продолжить».