Literature
Although Experience Sampling Methods (ESM) allows for rich insights into dynamic psychological processes, measurement issues, including reactivity, reliability, validity, and feedback mechanisms, pose significant methodological challenges. Importantly, recent work by Vogelsmeier et al. (2024) shows that measurement practices in ESM, such as using validated items or assessing reliability, are not regularly done, often because researchers do not know how to approach such methodological questions. In the past years, members of the MITNB consortium have made substantial contributions to improve the quality of measurement in ESM. Some of this work is highlighted on this page.
Methodological Best Practices
Well-designed ESM studies depend on choosing what to measure, how often, and how to maintain psychometric rigor. These two papers are summaries of current state-of-the art practices.
Fritz et al. (2024)
Offers a practical guide of 10 essential considerations, ranging from study design to data analysis.Tuerlinckx et al. (2025)
Reviews emerging advancements in ESM methodology, including best practices around reliability, validity, and timing of assessments.
Reactivity, Invariance & Response Quality in ESM
Although repeated assessments provide rich data, they can influence how participants respond, leading to phenomena such as measurement reactivity and careless responding. These papers summarize the issue and point towards analytical solutions to deal with them.
König et al. (2022)
Systematic review and meta-analysis of reactivity effects in ESM studies on health behavior. It highlights common patterns and offers mitigation strategies.Vogelsmeier et al. (2022)
Introduces how latent Markov factor analysis can be used to assess measurement invariance in ESM data.Vogelsmeier et al. (2024)
Applies a combination of confirmatory mixture item response theory model with latent Markov factor analysis to detect careless responding in ESM data.
Validation of ESM Items
Although ESM is widely used in psychological research, rigorous validation of constructs remains underdeveloped. Many studies rely on items that have not undergone thorough psychometric evaluation. A first good start is to scan the ESM item repository. The following contributions help in assessing what a good item is and how mixed-methods approaches can help reveal problems with validity in ESM research.
Eisele et al. (2025)
Presents the ESM-Q, a consensus-based tool for assessing the quality of ESM items. It provides a practical framework for evaluating construct validity during study design.Schorrlepp et al. (accepted)
Demonstrates how qualitative methods can serve as validity checks in clinical ESM settings and provides concrete implementation examples.Olthof et al. (2024)
Uses a mixed-methods ESM approach, showing how quantitative rating scales sometimes do not match the lived experience of participants.
Reliability & Validity of ESM Data
Assessing psychometric properties in ESM studies is complicated by the dynamic nature of emotions and behaviors. These papers describe multiple ways to assess reliability and validity in ESM data and show two recent examples of ESM questionnaires that were validated using multiple methods.
Castro-Alvarez et al. (2024)
Offers an overview of statistical approaches to quantify reliability of dynamic psychological constructs.Dejonckheere et al. (2022)
Describes guidelines for evaluating reliability in single-items in ESM, which are widely used.Siepe et al. (2025)
Provides a tutorial for using descriptive statistics and data-visualization techniques to increase the understanding of the performance of ESM items.Cloos et al. (2023)
Develops and evaluates the validity of different items to measure positive and negative affect in ESM protocols.Kochhar et al. (2025)
Assesses the psychometric properties of an ESM protocol for cigarette-smoking youth using a mixed-methods approach.
Design Features: Sampling Strategies & Item Design
The quality of ESM data depends not only on what is measured, but also when, how often, and how items are presented. These papers investigate how these choices can impact data quality and provide guidelines for designing ESM studies.
Eisele et al. (2021) & Eisele et al. (2023)
Explore how different sampling protocols and individual participant traits influence reactivity and reliability.Cloos et al. (2025)
Proposes using intensity profile drawings to better capture continuous emotional experiences.Dejonckheere et al. (2024)
Recommends showing participants their previous rating to improve accuracy of ESM ratings.Haslbeck et al. (2025)
Compares Likert scales vs. Visual Analogue Scales (VAS), addressing a long-standing debate over optimal scale format, providing evidence that VAS scales may be superior.Cloos, Siepe et al. (2025)
Provide evidence for the accuracy and consistency of VAS in ESM.
Note: This paper is a direct output from a collaboration formed during the MITNB annual meeting 2024.Henninger et al. (2025)
Shows that extreme response styles may distort assessments of emotion variability over time.Panayiotou, Razum et al. (preprint)
Shows that the PHQ-9, a leading screening instrument, is misinterpreted by a majority of participants, casting doubts about the validity of this measure.
Note: This paper is a direct output from a collaboration formed during the MITNB annual meeting 2024.
Beyond Rating Scales: Capturing Daily Life Complexity
ESM typically uses standard ratings scales, but such scales often miss important nuances. These tutorial papers show different methods beyond standard rating scales to better capture the complexity of daily life.
Stadel et al. (2025)
Compares categorical response scoring vs. open-ended qualitative analysis to study daily activities.Stadel et al. (2024)
Combines ESM with personal social network mapping to explore how social contexts shape psychological experience.
Feedback in ESM: Enhancing Participant Engagement
Feedback systems in ESM research not only improve participant retention, but also create therapeutic opportunities in psychological interventions.
Leertouwer et al. (2021)
Reviews explicit and implicit assumptions behind personalized feedback using self-report ESM data, encouraging critical reflection among practitioners.Bringmann et al. (2025)
Introduces a feedback system in the m-Path app, which combines qualitative insights and quantitative ESM data in cognitive behavioral interventions.Rimpler et al. (2024)
Presents the FRED software, a user-friendly tool for creating feedback reports based on ESM data.