Literature
Although ecological momentary assessment (EMA) allows for rich insights into dynamic psychological processes, measurement issues, including reactivity, reliability, validity, and feedback mechanisms, pose significant methodological challenges. Importantly, recent work by Vogelsmeier et al. (2024) shows that measurement practices in EMA, such as using validated items or assessing reliability, are not regularly done, often because researchers do not know how to approach such methodological questions. In the past years, members of the MITNB consortium have made substantial contributions to improve the quality of measurement in EMA. Some of this work is highlighted on this page.
Methodological Best Practices
Well-designed EMA studies depend on choosing what to measure, how often, and how to maintain psychometric rigor. These two papers are summaries of current state-of-the art practices.
Fritz et al. (2024)
Offers a practical guide of 10 essential considerations, ranging from study design to data analysis.Tuerlinckx et al. (2025)
Reviews emerging advancements in EMA methodology, including best practices around reliability, validity, and timing of assessments.
Reactivity, Invariance & Response Quality in EMA
Although repeated assessments provide rich data, they can influence how participants respond, leading to phenomena such as measurement reactivity and careless responding. These papers summarize the issue and point towards analytical solutions to deal with them.
König et al. (2022)
Systematic review and meta-analysis of reactivity effects in EMA studies on health behavior. It highlights common patterns and offers mitigation strategies.Vogelsmeier et al. (2022)
Introduces how latent Markov factor analysis can be used to assess measurement invariance in EMA data.Vogelsmeier et al. (2024)
Applies a combination of confirmatory mixture item response theory model with latent Markov factor analysis to detect careless responding in EMA data.
Validation of EMA Items
Although EMA is widely used in psychological research, rigorous validation of constructs remains underdeveloped. Many studies rely on items that have not undergone thorough psychometric evaluation. A first good start is to scan the ESM item repository. The following contributions help in assessing what a good item is and how mixed-methods approaches can help reveal problems with validity in EMA research.
Eisele et al. (2025)
Presents the ESM-Q, a consensus-based tool for assessing the quality of EMA items. It provides a practical framework for evaluating construct validity during study design.Schorrlepp et al. (accepted)
Demonstrates how qualitative methods can serve as validity checks in clinical EMA settings and provides concrete implementation examples.Olthof et al. (2024)
Uses a mixed-methods EMA approach, showing how quantitative rating scales sometimes do not match the lived experience of participants.
Reliability & Validity of EMA Data
Assessing psychometric properties in EMA studies is complicated by the dynamic nature of emotions and behaviors. These papers describe multiple ways to assess reliability and validity in EMA data and show two recent examples of EMA questionnaires that were validated using multiple methods.
Castro-Alvarez et al. (2024)
Offers an overview of statistical approaches to quantify reliability of dynamic psychological constructs.Dejonckheere et al. (2022)
Describes guidelines for evaluating reliability in single-items in EMA, which are widely used.Siepe et al. (2025)
Provides a tutorial for using descriptive statistics and data-visualization techniques to increase the understanding of the performance of EMA items.Cloos et al. (2023)
Develops and evaluates the validity of different items to measure positive and negative affect in EMA protocolsKochhar et al. (2025)
Assesses the psychometric properties of an EMA protocol for cigarette-smoking youth using a mixed-methods approach.
Design Features: Sampling Strategies & Item Design
The quality of EMA data depends not only on what is measured, but also when, how often, and how items are presented. These papers investigate how these choices can impact data quality and provide guidelines for designing EMA studies.
Eisele et al. (2021) & Eisele et al. (2023)
Explore how different sampling protocols and individual participant traits influence reactivity and reliability.Cloos et al. (2025)
Proposes using intensity profile drawings to better capture continuous emotional experiences.Dejonckheere et al. (2024)
Recommends showing participants their previous rating to improve accuracy of EMA ratings.Haslbeck et al. (2025)
Compares Likert scales vs. Visual Analogue Scales (VAS), addressing a long-standing debate over optimal scale format, providing evidence that VAS scales may be superior.Cloos, Siepe et al. (2025)
Provide evidence for the accuracy and consistency of VAS in EMA.
Note: This paper is a direct output from a collaboration formed during the MITNB annual meeting 2024.Henninger et al. (2025)
Shows that extreme response styles may distort assessments of emotion variability over time.Panayiotou, Razum et al. (preprint)
Shows that the PHQ-9, a leading screening instrument, is misinterpreted by a majority of participants, casting doubts about the validity of this measure.
Note: This paper is a direct output from a collaboration formed during the MITNB annual meeting 2024.
Beyond Rating Scales: Capturing Daily Life Complexity
EMA typically uses standard ratings scales, but such scales often miss important nuances. These tutorial papers show different methods beyond standard rating scales to better capture the complexity of daily life.
Stadel et al. (2025)
Compares categorical response scoring vs. open-ended qualitative analysis to study daily activities.Stadel et al. (2024)
Combines EMA with personal social network mapping to explore how social contexts shape psychological experience.
Feedback in EMA: Enhancing Participant Engagement
Feedback systems in EMA research not only improve participant retention, but also create therapeutic opportunities in psychological interventions.
Leertouwer et al. (2021)
Reviews explicit and implicit assumptions behind personalized feedback using self-report EMA data, encouraging critical reflection among practitioners.Bringmann et al. (2025)
Introduces a feedback system in the m-Path app, which combines qualitative insights and quantitative EMA data in cognitive behavioral interventions.Rimpler et al. (2024)
Presents the FRED software, a user-friendly tool for creating feedback reports based on EMA data.