Abstract
Objectives
The Five Facet Mindfulness Questionnaire (FFMQ) is a popular self-report instrument for mindfulness assessment. However, several studies report mixed evidence regarding its reliability and validity. While recent replication studies have shown several issues regarding its latent structure, first-order facets seemed to replicate successfully. This study proposes an exploratory approach to these facets on an item level in one sample, with cross-validation in another sample.
Methods
Using a snowball sampling, 1008 participants were recruited in the first sample. Psychometric networks were applied to explore relations between items and item clusters. We compared these exploratory latent variable proposals with previous literature. A second sample of 1210 participants was collected from an FFMQ validation study, and confirmatory factor analyses were applied to cross-validate findings on the first sample.
Results
The FFMQ showed a positively correlated network. Exploratory analyses suggested the 5-facet structure as stable with alternatives of 4-facet (merging Observe and Non-Judging) and 6-facet (splitting Acting with Awareness in two) solutions. However, the CFAs in the second sample did not provide clear support to any solution.
Conclusions
The FFMQ showed unclear evidence on its latent structure. We propose researchers and users of the FFMQ to use the most fitting solution between the 5 and 6-facet solutions in their data, since the 4-facet solution is difficult to interpret. We also propose cautionary notes and guidelines for researchers and applied users of the FFMQ and regarding this instrument. We conclude that more research is needed in mindfulness assessment to provide robust measurements.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
Mindfulness research has experimented a rapid growth in interest and production (American Mindfulness Research Association, 2021). This interest has drawn even political attention (The Mindfulness All-Party Parliamentary Group, 2015), and great importance in academia exploring its positive effects (for a review, see Creswell, 2016). Thus, reliable and valid measurement models and instruments are necessary to validate and expand this demand for evidence-based mindfulness research (for a review, see Bergomi et al., 2013b). The Five Facet Mindfulness Questionnaire (FFMQ; Baer et al., 2008) is a well-known and used self-report instrument to measure mindfulness, since a large body of literature using it.
FFMQ items measure mindfulness as an overall psychological construct with five main facets, named: “Observe,” as the tendency to observe stimuli (e.g., sensations of the wind in the body); “Describe,” as the tendency to verbally describe those stimuli; “Acting with Awareness” or “Actaware,” as the tendency to attend to one’s activities in the present moment (in contrast with behaving mechanically, or driven by “automatic pilot”); “Non-Judging of Inner Experience” or “Non-Judging,” as the tendency to not judge a particular inner experience as good or bad; “Non-Reacting to Inner Experience” or “Non-Reacting,” as the tendency to not immediately react to a particular inner experience and “take a step back” to gain perspective. The popularity of the FFMQ is somehow due to a high number of international validations of the questionnaire (Lecuona et al., 2020). Nevertheless, scientific literature has drawn some criticisms toward the FFMQ. For example, evidence suggests that mindfulness facets seem to be stable and valid while the overall mindfulness factor seems unstable. At the same time, the external validity of the instrument remains robust except for the Observe facet in the general population (i.e., non-meditators). Structuring these comments, criticisms provide challenging evidence on (1) construct, (2) external, and (3) content validity.
As regards (1) construct validity, a recent review and replication study reported several issues (Lecuona et al., 2020). First, most validation studies of the FFMQ cast concerns regarding their data analysis, which questions the apparent robustness of the FFMQ in the literature. Second, the latent structure could not be reproduced as a whole, requiring the inclusion of method factors and the removal of the Observe facet. Therefore, the replication concluded that the FFMQ seems to have an unstable overall mindfulness factor in the general population. However, mindfulness facets (first-level latent factors) seem to replicate successfully.
As regards (2) external validity, a recent meta-analysis (Carpenter et al., 2019) suggests similar conclusions as for construct validity (i.e., facets behaving psychometrically correct except for the Observe facet) and expands them. Concretely, it points out that although the FFMQ seems to be inversely related to negative affect and psychopathological symptoms, the Observe facet is not meaningfully correlated in the same way within general population (i.e., majority of non-meditators). Therefore, the Observe facet seems to not fit well to contribute to the overall mindfulness score, as well as to the external validity of the instrument.
Finally, as regards (3) content validity, a study provided evidence on how Buddhist expert meditators have little agreement on the dimensions of FFMQ as mindfulness components, and also had a consistent disagreement with the Observe facet (Belzer et al., 2013). Furthermore, another study showed that the FFMQ failed to show discriminant validity between a mindfulness-based intervention (MBI) and active controls (Goldberg et al., 2015).
Therefore, it appears that the FFMQ has mixed evidence in the scientific literature. This state-of-the-art can produce undesirable consequences for the field. Skeptics could oversimplify or exaggerate this criticism (e.g., arguing MBIs hold null effects due to the FFMQ issues). It is also predictable for a large force of applied users to ignore criticisms and use the standard (and apparently not robust) model of the FFMQ. This dissonance between technical and applied research has taken place in other fields, like Cronbach’s α misuse (Mcneish, 2017). Therefore, while research develops new scalable instruments to measure mindfulness, an in-depth exploration of its properties seems necessary to (1) acknowledge which components of the FFMQ are valid, and (2) offer an intermediate solution for both applied and research uses.
It can be argued that although the FFMQ can be used to assess mindfulness facets, the structure of said facets needs more evidence to be considered solid, especially regarding the Observe facet. The reason is that most of the FFMQ validation studies assessed the hierarchical model (assuming a second-order overall mindfulness factor). In contrast, previous studies did not prioritize the correlated 5-facet model. Due to this, inter-facet correlations, cross-loadings, or correlations between items of different facets were not explored in depth. In this regard, the field requires more exploratory techniques than CFA. Exploratory Factor Analysis (EFA) and Exploratory Structural Equation Models (ESEM) are examples of these techniques. However, a recent technique can explore more freely how items relate to each other, namely the psychometric network models (Epskamp et al., 2018). Also known as Regularized Partial Correlation Networks (RPCNs), they have been recently introduced in psychological research (e.g., transdiagnostic models in psychopathology).
RPCNs allow to freely estimate correlations between psychological variables without being caused by an unobserved latent entity. In addition, they provide more unbiased estimations with a three-step approach: First, they compute the correlation matrix (in the case for Likert variables, polychoric correlations); second, they partialize all correlations for the effect of all other correlations; and third, they regularize them (i.e., forcing all small correlations to zero), assuming them to be non-meaningful and aiming for a parsimonious solution (further reading in Epskamp et al., 2018). Thus, RPCNs stand out as a more flexible technique to explore complex interactions between observed variables. Moreover, RPCNs can be used to estimate stable groups of variables via Exploratory Graph Analysis (EGA, Golino & Epskamp, 2017). These groups can be interpreted as emergent phenomena that arise from the networks, which is compatible with latent variables. Given EGA’s performance has shown to be as precise as standard EFA techniques like parallel analysis (Golino et al., 2020), it seems suitable as an interesting option to explore the FFMQ at a first-order level. However, since RPCNs and EGA are novel techniques, a cross-validation technique should benefit the study. In this sense, confirmatory factor analysis (CFA) seems like the optimal technique since it provides testing for specific solutions, like the outputs produced by the EGA.
The FFMQ seems to provide stable facets (except for Observe in some contexts), but literature has not explored them deeply. Thus, this study explores the FFMQ at a facet level via psychometric networks. More concretely, it explores the facet solutions that may underlie the FFMQ via RPCNs and EGA. We expect the five-facet theoretical model to be reproduced in our data, and therefore replicate the facet structure of the FFMQ.
Method
Participants
We collected two samples for this study, for exploratory and confirmatory analyses, respectively. The first sample was a snowball sample of 1008 participants. We recruited participants for the first sample using a three-step approach. First, we contacted university students from the Faculties of Psychology of the Autonomous University of Madrid and the King Juan Carlos University Madrid asking for participation. Second, we contacted a mindfulness institute to recruit participants before starting mindfulness-based interventions. Third, we posted on our social media profiles (i.e., Facebook, Twitter, Whatsapp, and Telegram) a call for participants. In all steps, all participants were asked to post on their social media profiles our call for participants and contact their relatives to ask for participation. We conducted a power analysis for the estimated networks extracting three indices: True-estimated network correlation, sensitivity (true positives rate), and specificity (true negatives rate) (Epskamp & Fried, 2018). For our overall sample, we obtained a high correlation (> 0.95) and specificity values (around 88%), while sensitivity (around 68%) provided more modest values. However, figures of the power analysis (available at https://osf.io/65abr) show how sensitivity values are close to asymptotic. Since RPCNs are designed to prefer higher specificity than sensitivity, these results are expectable. Participants identified predominantly as women (75.89%), young (M = 35.84 years, SD = 13.31), with college studies or above (60.90%), and had practiced meditation on at least two occasions (72.59%).
The second sample was collected by Meda et al. (2015) as an incidental student sample of 1210 participants. Participants identified predominantly as women (67.62%), young (M = 20.10 years, SD = 5.15), with college studies or above (90.10%). Meditation experience was not collected.
Procedure
All recruited participants signed an informed consent of participation, which informed them about the privacy, anonymity, and confidentiality of the treatment of their data, following the Helsinki protocol. Then, we administered a copy of the FFMQ along with demographic items. All data were collected in Spanish. We explained the rules of responding to the questionnaire. The participants could ask any question or issue that they had to the researcher in charge of the data collection, and they could withdraw from the questionnaire at their convenience. The participants did not receive any compensation for their participation. Finally, the second sample was collected by contacting the researchers in charge of the sample and agreeing on providing the dataset (details are provided in Meda et al., 2015).
Measures
The Five Facets of Mindfulness Questionnaire
The FFMQ consists of 39 self-reported items with a Likert response format from 1 (totally disagree) to 5 (totally agree) (Baer et al., 2008; validated to Spanish population by Cebolla et al., 2012). As discussed, it has a proposed latent structure of five facets and a hierarchical or bifactor overall facet with potential method factors, with an alternative structure of six facets (Karl et al., 2020; Lecuona et al., 2020).
Data Analyses
Descriptive statistics were obtained for all variables and plots were examined. We applied RPCNs (Epskamp et al., 2018) to explore relations among items without assuming any structure or model. RPCNs import the network analysis framework to relations between psychological variables. Thus, network nodes (the basic components of the network) represent variables, and edges (links between nodes) represent partial regularized correlations. Mathematically, RPCNs take a correlation matrix (in this case, the polychoric correlation matrix since all items were ordinal) as an input to partialize all correlations for all the other present correlations in the matrix. Then, it regularizes the false positive rate of the network forcing near-zero correlations to zero, assuming them to be essentially uncorrelated. We chose the Expected Bayesian Inference Criteria (EBIC) estimation with the graphical Least Absolute Shrinkage Optimization (gLASSO) regularization as the network estimation method since it provides the most adaptable estimations for ordinal, approximately normal data. Besides, a proportion of explained variance for each node by the network can be obtained with an R2 statistic. The final product of this process is a partial regularized correlation matrix, named as “weight matrix.”
Following Mullarkey et al. (2018), we examined item standard deviations with an exclusion criterion of below 2.5 SDs of standard deviations since small-variance items can influence the final estimates of the network. No items displayed standard deviations below the criterion. Also, conceptually or empirically overlapping items can bias network estimations since they could measure latent traits without relevant, unique content. Therefore, we searched for empirically overlapped items with a three-step approach: (1) we selected pairs of items with potential conceptual overlap and estimated their correlations, (2) obtained their mean and SD to form a criterion of 1 SD as empirically overlapping items, and (3) created composite scores for each resulting pair, reducing the network. Two pair of items met our criteria. Once the final network was reduced and estimated, we applied bootstrap techniques to the network to assess replicability of the network. Among others, we implemented the correlation stability coefficient (CS) to assess stability of edges, with values above 0.25 as minimum, and values above 0.5 as ideal (Epskamp et al., 2018). Once the replicability of the network is assessed, we proceeded to represent the network graphically (see Fig. 1 for an example). We chose the Fruchterman-Reingold algorithm since it provides visually clear and intuitive displays (all data are available at the Open Science Framework-OSF, in Supplementary Materials: https://osf.io/65abr/).
Once the network is estimated, it is relevant to assess the relative importance of each node. Two analytic concepts can be provided: centrality and clustering. Centrality involves the intensity, closeness and inter-connectivity of each node. Clustering involves how much a node is closely connected to a closed group of other nodes (e.g. items clustered due to a latent variable). Centrality measures will not be considered in this study due to bias and interpretation issues when latent variables are involved (Bringmann et al., 2019).
Nevertheless, clustering was assessed with a five-step approach: First, we marked nodes with colors corresponding to theoretical facets to assess if current networks reproduced the model. Second, since Fruchterman-Reingold plots do not provide interpretable distances between nodes, we implemented multi-dimensional scaling plots (see OSF Supplementary Materials) for better assessment of clusters. Third, we performed an Exploratory Graph Analysis (EGA, Golino & Epskamp, 2017) to explore latent variables in the network. This technique has shown comparable performance to traditional latent variable extraction methods, such as Parallel Analysis (Golino et al., 2020). Fourth, we also implemented a community analysis (mathematically equivalent to EGA) comparing other algorithms than the EGA’s default (i.e., walktrap and Louvain), like spinglass and multilevel. Model fit was assessed with entropy indices (Golino et al., 2019). More concretely, the Entropy Index (EFI) with Shannon entropy, the EFI with Von Neumann entropy (EFIvn), and the Total EFIvn (TEFIvn). In all cases, lower values indicate a better fit. TEFIvn was prioritized since it shows higher accuracy in over factored solutions. Fifth, we bootstrapped the EGAs to assess its replicability. We implemented two network estimation methods (gLASSO and Triangular Maximally Filtered Graph, or TMFG) crossed with walktrap and Louvain algorithms. Dimension stability was assessed with the median of proportions of proposed solutions with a confidence interval, while item stability was assessed with the median of replicated item correspondences following a certain solution.
Finally, we implemented an Exploratory Factor Analysis for this sample as a comparison to RPCNs and EGA estimates with a more standard technique (see OSF Supplementary Materials for technical details and outputs).
To cross-validate our exploratory results, we implemented confirmatory factor analysis (CFAs) to the second sample with the structures proposed by EGA. As we mentioned before, all our items were Likert-type so we implemented polychoric correlations, except for composite scores of overlapping items. The chosen estimation method was weighted least squares mean-and-variance adjusted (WLSMV) since it provides better performance to other methods, like maximum likelihood robust (MLR) or unweighted least squares (ULS) (Li, 2016). Fit measures were the chi-square statistic (reported but not interpreted), the Comparative Fit Index (CFI) and Tucker-Lewis Index (TLI), both with values > 0.90 as a good fit, the root mean square error approximation (RMSEA), with values < 0.07 as a good fit, and the standardized root mean square residual (SRMR), with values < 0.08 as good fit (following Hair et al., 2014, p. 584). If a CFA model did not meet the fit criteria, we implemented modification indices that provided a theoretical and empirical improvement of the model. A maximum of three modification indices were allowed. We compared CFA models with formal tests as Δχ2 (optimized for robust measures), and direct differences between CFI, TLI, RMSEA, and SRMR. As a general assessment, we interpret increases in CFI and TLI and decreases of RMSEA and SRMR as indicative of better fit.
All analyses were computed using the R environment (R Development Core Team, 2020). Descriptive statistics were computed using psych (Rewelle, 2014), while the network analyses were computed using bootnet (Epskamp et al., 2018), mgm (Haslbeck & Waldorp, 2015), and networktools (Jones, 2020). The EGAs were computed using EGAnet (Golino & Christensen, 2019), and the CFAs were computed using lavaan (Rosseel, 2012). Data is available at https://osf.io/65abr/ along with scripts and supplementary materials.
Results
For both samples, item descriptive statistics did not show relevant skewness except for item 23 in sample 2, although they were prone to platykurtic distributions (Table 1). Therefore, we selected the mean and standard deviation (SD) as their central tendency and dispersion estimators. All items had means around 3 (the central category) and SDs around 1. No items showed informativeness issues (none had SDs relevantly below the rest).
Overlapping item analysis proposed two pairs of items as empirically overlapping: Items 5 and 13 (“When I do something, my mind wanders and I get easily distracted,” and “I get easily distracted,” respectively), and items 25 and 30 (“I tell myself I shouldn’t be thinking the way I’m thinking” and “I think some of my emotions are bad or inappropriate and I shouldn’t feel them,” respectively). Therefore, we merged both pairs into two composite scores. Regarding network analyses, the FFMQ items showed a positively correlated network (Fig. 1) with moderate-to-high predictability (mean of R2 = 0.444, SD = 0.126), with only one item showing an R2 lower than 0.25 (item 11, R2 = 0.142). Bootstrap analysis revealed a generally stable network (see OSF Supplementary Materials), with a correlation stability coefficient of 0.75, which is interpreted as stable. Based on visual inspection, theoretical facets show a good preliminary correspondence within the network, with all items belonging to a certain facet being close to each other and little or no edges connecting between groups.
Multi-dimensional scaling plots (Fig. 2) show a similar behavior, with items belonging to each theoretical facet being close and connected with each other. However, two cautionary patterns arise: First, the Observe and Non-Reacting facets seem more disperse than the rest of the facets, with some interconnections between them. This could be due to these facets being merged in this sample, although they struggle to find a theoretical meaning following their content. Second, the Acting with Awareness items appears to split into two subgroups, with low correlations between each other. While proximate between them, the first half of items of the Actaware are highly connected between them, while only one weak connection for each item with other Actaware items. The items of each subgroup are the first half of the Actaware facet (5 and 13 composite, 8, and 18) and the second half (23, 28, 34, and 38). Attending to the content of the items, a plausible interpretation can be made. While both measure components of mindlessness or “automatic pilot,” the first subgroup seem to be more related to distractibility (e.g., item 13: “I am easily distracted”), and the second subgroup to mindless actions (e.g., item 38: “I find myself doing things without paying attention”). Item distribution with this proposed split is displayed in Table 2. Finally, connections between theoretical facets are few and low, except for item 2.
Attending the rest of the network, the Describe and Non-Judging items could display a behavior like Actaware but were not regarded in this way due to more connections between items. Therefore, we conclude that the theoretical five facets seem to hold, but two alternative models arise as potentially meaningful: A 4-facet model with Observe and Non-Reacting merged into a single facet (Fig. 3), and a 6-facet model with Actaware split in two (Fig. 4).
Exploratory Graph Analysis
The default EGA supported the 5-facet model for the FFMQ. Different number of steps in the walktrap algorithm threw the same solution except for 8 and 9 steps, proposing the 4-facet solution. Regarding the spinglass algorithm, a different outlook was displayed. While the default algorithm proposed the 5-facet model, most variations of spins (from 25 to 500 spins by steps of 50 stems) proposed the 6-facet model. Finally, the multilevel algorithm proposed the 5-facet model without variations.
Entropy indices were convergent with the walktrap proposal, getting the best fit for the 5-facet model (Table 3). The bootstrap analysis proposed in all cases the 5-facet solution with confidence intervals not reaching values below 4 or above 6 (see OSF Supplementary Materials). The 4 and 6-facet models received very little support. Dimension and item stability analysis reported high stability for almost all facets in all cases. The only exception was Non-Reacting, due to low or medium recovery rates for item 4, with the rest of the items displaying acceptable or good recovery rates. Therefore, EGA proposes the 5-facet structure as the most fitting and stable solution for the FFMQ.
Finally, regarding the Exploratory Factor Analysis, no clear support was provided for the 4, 5 or 6-facet model (see OSF Supplementary Materials). Nevertheless, the 5-facet model is perhaps the most fitting model due to partial support of Parallel Analysis and factor loadings following a simple structure with theoretical meaning.
Confirmatory Factor Analysis
As mentioned before, we imported the three resulting solutions from EGA and applied CFAs to cross-validate its results in the second sample. Fit measures are displayed in Table 4. None of the raw models displayed a good fit. Thus, we examined modification indices. We found three cross-loadings that improved fit: (1) Item 17 measuring Observe, which can be explained as that being attentive would benefit directly the tendency to label thoughts as good or bad since it measures internal monitoring; (2) Item 22 measuring Non-Judging, which can be explained as that a non-judging attitude would benefit directly the descriptions of sensations since it’s a more neutral approach to experience; and (3) Item 16 measuring Non-Judging, which can be explained similarly than item 22, but with emotions instead of sensations. By adding modification indices, model fit was appropriate for CFI, but not for TLI, RMSEA, and SRMR in all models. However, fit comparisons offered a significant improvement for all models (all Δχ2 with p < 0.001), and with improvements in all fit indices (ΔCFI, ΔTLI, ΔRMSEA, and ΔSRMR). Therefore, both the 5-facet and 6-facet structures provided a questionable fit. Factor loadings were positive, significant, and above 0.3 in all models, and all covariances between factors were positive and significant. This suggests a potential second-order latent solution, thus we estimated a hierarchical model. However, fit measures were not satisfactory (χ2(624) = 13,028.06, p < 0.001, CFI = 0.687, TLI = 0.666, RMSEA = 0.118, SRMR = 0.126), even when modification indices were considered. Therefore, we discard a second-order stable structure.
To conclude this section, cross-validation results indicated a questionable fit of FFMQ in all models, although support partially the 5 and 6-facet models with minor modifications.
Discussion
This study aimed to explore the FFMQ facets to generate an interpretable solution for existing and future research. To do so, we implemented RPCNs and EGAs to explore and contrast optimal solutions for a large sample. Results show that the 5-facet solution produces a suitable network supported by EGA, although two alternative structures are also proposed: A 4-facet structure, merging Observe and Non-Reacting, and a 6-facet structure, splitting Acting with Awareness in two, named Distractibility and Mindless Actions.
As regards the 4-facet solution, the Observe and Non-Reacting facets were merged. Technically, results suggest items of both facets could be considered measuring the same phenomenon. One possible interpretation is a theoretically authentic latent variable, although it is difficult to label and interpret. We could think that these items assess a contemplative mindset characterized by an equanimity attitude to the phenomena that constitute the experience (i.e., sensations, thoughts, emotions, sounds). As a whole, these items seem to evaluate the tendency to notice and pay attention to internal and external phenomena, establishing a relationship with them characterized by decentering (i.e., the meta-cognitive observation of the contents of consciousness; Bernstein et al., 2015), and detachment (i.e., not-identification with the contents of the consciousness). Both processes would reduce the tendency to react to thoughts, feelings, and emotions and promotes the tendency to open to them, and allow and let them be without a reactive attitude. In other words, these items of the FFMQ seem to evaluate the ability of the individual to remain present in an attitude of equanimity. We understand equanimity as “an even-minded mental state or dispositional tendency toward all experiences or objects, regardless of their affective valence (pleasant, unpleasant or neutral) or source … (that) allows awareness to be even and unbiased by facilitating an attitude of non-attachment and nonresistance” (Desbordes et al., 2015, p. 357–358).
Another possibility is that those items represent a method factor. The Observe facet is known for its mixed functioning. Also, the Non-Reacting facet could present a mixed behavior due to a lack of comprehension of items in a general population sample, as suggested in previous studies (Lecuona et al., 2020) and stability analysis of EGA. It could be the most distant facet to the Western culture to understand, therefore prone to be influenced by confusion or response biases. Generically, the western population experiences that thoughts and emotions constitute the self, and they do not experience the feelings, emotions, and thoughts as phenomena that flow through the field of consciousness that are possible to observe (Roeser & Peck, 2009). The fact of not experiencing the “self-as-an-observant” of internal phenomena can make it challenging to understand the items of the Non-reacting facet since they refer to relate with thoughts and emotions from a “self-as-an-observant” (e.g., item 9, “I watch my feelings without getting lost in them” or item 19, “when I have distressing thoughts or images, I step back and am aware of the thought or image without getting taken over by it”). Thus, this merged facet of Observe and Non-React could represent a general bias latent factor of mixed or confused item responses. Nevertheless, both hypotheses fail to provide substantive arguments to illustrate the common grounds between these two phenomena. In addition, more research is needed on the cross-cultural potential differences on the FFMQ and mindfulness phenomena (Karl et al., 2020).
Regarding the 6-facet solution, the Actaware facet split into two sub-facets. A content analysis revealed the best labels were Distractibility (i.e., tendency or proneness to distraction) and Mindless actions (this is, tendency or proneness to perform actions without conscience of doing them). One similar phenomenon to mindless actions is the “automatic pilot” concept (Kabat-Zinn, 1990), but this characteristic appeared to be in both sub-facets, as revealed by the content analysis. From a theoretical point of view, we can distinguish distractibility and mindless actions, although it might not be essentially different phenomena, or at least, they would be overlapped. However, this solution has been replicated in a recent cross-cultural comparison (Karl et al., 2020) with a large sample size (N > 8000). Thus, more research is needed to ensure the likelihood of this structure.
The EGA displays the 4 and 6-facet solutions as unstable and more disorganized in this sample, prioritizing the 5-facet structure. Nevertheless, the CFA prioritized almost equally the 5-facet and 6-facet solutions, while content analysis converged with CFA. Therefore, we conclude that while the 5-facet solution is the standard of mindfulness literature, both solutions count with strong evidence supporting them. Thus, the 6-facet solution provides a competitive alternative to the 5-facet solution. We conclude that the data does not fully support neither structure, thus we call for caution and encourage future studies to explore the replicability of our findings. In addition, this study points out that the internal validity of the FFMQ is still unclear and more developments (evidence and improved instruments) are required in future studies. In this sense, a promising area of development is behavioral instruments since they are less biased by cultural or verbal representations. Examples of these instruments are the breath-counting task (Isbel et al., 2020; Levinson et al., 2014) or the simulated thoughts paradigm (Amir et al., 2021; for a review, see Hadash & Bernstein, 2019). However, this research field is in early development. Future studies should expand it and aim to develop theoretically and empirically robust behavioral instruments for mindfulness.
We propose some guidelines to make optimal use of the FFMQ. First, we advise considering complementary mindfulness instruments, like the Freiburg Mindfulness Questionnaire (FMI, Sauer et al., 2013) or the Comprehensive Inventory of Mindfulness Experiences (CHIME-β, Bergomi et al., 2013a). In this sense, we prioritize psychometric stability over losing theoretical richness, but we leave that to the discretion of the user. Second, we recommend scoring facets of the FFMQ, but not the overall mindfulness facet. The reason to advice against the overall mindfulness factor is the lack of fit when including it that is supported by other studies as well (e.g. Karl et al., 2020; Lecuona et al., 2020). Finally, we advise researchers to test the 5 and 6-facet models when analyzing samples with the FFMQ and choosing the model with best fit, since literature seems to support both structures.
To conclude, this study shows partial support for the 5-facet model of the FFMQ. However, an alternative model of 6 facets (with the Acting with Awareness split in two) shows similar support. Mindfulness researchers and applied users should take caution in using the FFMQ and consider using other self-reports (e.g., FMI or CHIME-β), and if using it, assess the most fitting option between scoring the five or six facets, instead of the overall mindfulness factor.
Limitations and Future Research
This study has some limitations. First, the design was cross-sectional, which obscures possible longitudinal features of the FFMQ, i.e., invariance between pre and post participation in a mindfulness-based intervention. Also, all our measures were self-reports, which impedes control for single method bias. Another limitation is that the FFMQ was the only psychological instrument included in the study, which hinders the possibility of exploring relations of the FFMQ with other relevant constructs. Finally, although both sample sizes were large, future studies can assess meditation practice as a possible moderator of the latent facets and their interactions. Future research should also focus on finding more robust ways of assessing mindfulness, and if the FFMQ is used, explore the most suitable model for that specific data.
Change history
31 October 2021
The article is now online but Funding Note is missing in the published version.
References
American Mindfulness Research Association (AMRA) (2021). Mindfulness publications by year, 1980–2020. Retrieved March 10, 2019, from https://goamra.org/Library.
Amir, I., Ruimi, L., & Bernstein, A. (2021). Simulating thoughts to measure and study internal attention in mental health. Scientific Reports, 11(1), 1–17. https://doi.org/10.1038/s41598-021-81756-w
Baer, R. A., Smith, G. T., Lykins, E., Button, D., Krietemeyer, J., Sauer, S., Walsh, E., Duggan, D., & Williams, J. M. G. (2008). Construct validity of the five facet mindfulness questionnaire in meditating and nonmeditating samples. Assessment, 15(3), 329–342. https://doi.org/10.1177/1073191107313003
Belzer, F., Schmidt, S., Lucius-Hoene, G., Schneider, J. F., Orellana-Rios, C. L., & Sauer, S. (2013). Challenging the construct validity of mindfulness assessment—A cognitive interview study of the freiburg mindfulness inventory. Mindfulness, 4(1), 33–44. https://doi.org/10.1007/s12671-012-0165-7
Bergomi, C., Tschacher, W., & Kupper, Z. (2013a). Measuring mindfulness: First steps towards the development of a comprehensive mindfulness scale. Mindfulness, 4(1), 18–32. https://doi.org/10.1007/s12671-012-0102-9
Bergomi, C., Tschacher, W., & Kupper, Z. (2013b). The assessment of mindfulness with self-report measures: Existing scales and open issues. Mindfulness, 4(3), 191–202. https://doi.org/10.1007/s12671-012-0110-9
Bernstein, A., Hadash, Y., Lichtash, Y., Tanay, G., Shepherd, K., & Fresco, D. M. (2015). Decentering and related constructs. Perspectives on Psychological Science, 10(5), 599–617. https://doi.org/10.1177/1745691615594577
Bringmann, L., Elmer, T., Epskamp, S., Krause, R., Schoch, D., Wichers, M., Wingman, J. T. W., & Snippe, E. (2019). What do centrality measures measure in psychological networks? Journal of Abnormal Psychology, 128(8), 892–903. https://doi.org/10.1037/abn0000446
Carpenter, J. K., Conroy, K., Gomez, A. F., Curren, L. C., & Hofmann, S. G. (2019). The relationship between trait mindfulness and affective symptoms: A meta-analysis of the five facet mindfulness questionnaire (FFMQ). Clinical Psychology Review, 74, 101785. https://doi.org/10.1016/j.cpr.2019.101785
Cebolla, A., García-Palacios, A., Soler, J., Guillen, V., Baños, R., & Botella, C. (2012). Psychometric properties of the spanish validation of the five facets mindfulness questionnaire (FFMQ). The European Journal of Psychiatry, 26(2), 118–126. https://doi.org/10.4321/S0213-61632012000200005
Creswell, J. D. (2016). Mindfulness interventions. Annual Review of Psychology, 68, 491–516. https://doi.org/10.1146/annurev-psych-042716-051139
Desbordes, G., Gard, T., Hoge, E. A., Hölzel, B. K., Kerr, C., Lazar, S. W., Olendzki, A., & Vago, D. R. (2015). Moving beyond mindfulness: Defining equanimity as an outcome measure in meditation and contemplative research. Mindfulness, 6(2), 356–372. https://doi.org/10.1007/s12671-013-0269-8
Epskamp, S., & Fried, E. I. (2018). A tutorial on regularized partial correlation networks. Psychological Methods, 23(4), 617–634. https://doi.org/10.1037/met0000167
Epskamp, S., Borsboom, D., & Fried, E. I. (2018). Estimating psychological networks and their accuracy: A tutorial paper. Behavior Research Methods, 50(1), 195–212. https://doi.org/10.3758/s13428-017-0862-1
Goldberg, S. B., Wielgosz, J., Dahl, C., Schuyler, B., MacCoon, D. S., Rosenkranz, M., Lutz, A., Sebranek, C. A., & Davidson, R. J. (2015). Does the five facet mindfulness questionnaire measure what we think it does? Construct validity evidence from an active controlled randomized clinical trial. Psychological Assessment, 28(8), 1009–1014. https://doi.org/10.1037/pas0000233
Golino, H., Shi, D., Garrido, L., Christensen, A., Nieto, M., Sadana, R., Thiyagarajan, J. A., & Pérez-Molina, A. (2020). Investigating the performance of exploratory graph analysis and traditional techniques to identify the number of latent factors: A simulation and tutorial. Psychological Methods, 25(3), 292–320. https://doi.org/10.1037/met0000255
Golino, H., & Christensen, A. (2019). EGAnet: Exploratory graph analysis-a framework for estimating the number of dimensions in multivariate data using network psychometrics. R package version 0.8.0.
Golino, H. F., & Epskamp, S. (2017). Exploratory graph analysis: A new approach for estimating the number of dimensions in psychological research. PLoS ONE, 12(6), e0174035. https://doi.org/10.1371/journal.pone.0174035
Golino, H., Moulder, R., Shi, D., Christensen, A. P., Nieto, M. D., Nesselroade, J. R., Sadana, R., Thiyagarajan, J. A., & Boker, S. M. (2019). Entropy fit index: New fit measures for assessing the structure and dimensionality of multiple latent variables. Multivariate Behavioral Research. https://doi.org/10.1080/00273171.2020.1779642
Hadash, Y., & Bernstein, A. (2019). Behavioral assessment of mindfulness: Defining features, organizing framework, and review of emerging methods. Current Opinion in Psychology, 28, 229–237. https://doi.org/10.1016/j.copsyc.2019.01.008
Hair, J. F., Black, W. C., Babin, B. J., & Anderson, R. E. (2014). Multivariate data analysis. Pearson Education Limited.
Haslbeck, J. M. B., & Waldorp, L. J. (2015). mgm: Estimating time-varying mixed graphical models in high-dimensional data. Journal of Statistical Software. http://arxiv.org/abs/1510.06871.
Isbel, B., Stefanidis, K., & Summers, M. J. (2020). Assessing mindfulness: Experimental support for the discriminant validity of breath counting as a measure of mindfulness but not self-report questionnaires. Psychological Assessment, 32(12), 1184. https://doi.org/10.1037/pas0000957
Jones, P. (2020). networktools: Tools for identifying important nodes in networks. R package, version 1.2.3. https://cran-r-project.org/web/packages/networktools/.
Kabat-Zinn, J. (1990). Full catastrophe living: Using the wisdom of your body and mind to face stress, pain, and illness. Bantam Dell.
Karl, J. A., Méndez Prado, S. M., Gračanin, A., Verhaegen, P., Ramos, A., Prasun Mandal, S., Michalak, J., Zhang, C.-Q., Schmidt, C., Tran, U. S., Druica, E., Solem, S., Astani, A., Xinghua, L., Luciano, J. V., Tkalčić, M., Lilja, J. L., Dundas, I., Wong, S. Y. S., & Fischer, R. (2020). The cross-cultural validity of the five-facet mindfulness questionnaire across 16 countries. Mindfulness, 11(5), 1226–1237. https://doi.org/10.1007/s12671-020-01333-6
Lecuona, O., García-Garzón, E., García-Rubio, C., & Rodríguez-Carvajal, R. (2020). A psychometric review and conceptual replication study of the five facets mindfulness questionnaire latent structure. Assessment, 27(5), 859–872. https://doi.org/10.1177/1073191119873718
Levinson, D. B., Stoll, E. L., Kindy, S. D., Merry, H. L., & Davidson, R. J. (2014). A mind you can count on: Validating breath counting as a behavioral measure of mindfulness. Frontiers in Psychology, 5, 1202. https://doi.org/10.3389/fpsyg.2014.01202
Li, C.-H. (2016). Confirmatory factor analysis with ordinal data: Comparing robust maximum likelihood and diagonally weighted least squares. Behavior Research Methods, 48(3), 936–949. https://doi.org/10.3758/s13428-015-0619-7
Mcneish, D. (2017). Thanks coefficient alpha, we’ll take it from here. Psychological Methods., 23(3), 412–433. https://doi.org/10.1037/met0000144
Meda, R. M., Herrero, M., Blanco-Donoso, L. M., Moreno-Jiménez, B., & Palomera, A. (2015). Propiedades psicométricas del “cuestionario de cinco facetas de la conciencia plena” (five facets mindfulness questionnaire, FFMQ-M) en México. Behavioral Psychology, 23(3), 467–487.
Mullarkey, M. C., Stewart, R. A., Wells, T. T., Shumake, J., & Beevers, C. G. (2018) Self-dislike and sadness are central symptoms of depression in college students: A network analysis. PsyArXiv, osf.io/fujmb.
R Development Core Team (2020). R: A language and environment for statistical computing. Vienna, Austria: R Foundation for Statistical Computing. Retrieved from http://www.r-project.org/.
Rewelle, W. (2014). psych. In Procedures for personality and psychological research. Northwestern University, Evanston, Illinois, USA: Version 1.4.8. Retrieved from http://cran.r-project.org/package=psych.
Roeser, R. W., & Peck, S. C. (2009). An education in awareness: Self, motivation, and self-regulated learning in contemplative perspective. Educational Psychologist, 44(2), 119–136. https://doi.org/10.1080/00461520902832376
Rosseel, Y. (2012). lavaan: An R package for structural equation modeling. Journal of Statistical Software, 48(2), 1–36. Retrieved from http://www.jstatsoft.org/v48/i02/.
Sauer, S., Ziegler, M., Danay, E., Ives, J., & Kohls, N. (2013). Specific objectivity of mindfulness: A Rasch analysis of the freiburg mindfulness inventory. Mindfulness, 4, 45–54. https://doi.org/10.1007/s12671-012-0145-y
The Mindfulness All-Party Parliamentary Group. (2015). Mindful nation UK. London (UK). Retrieved from http://themindfulnessinitiative.org.uk/images/reports/Mindfulness-APPG-Report_Mindful-Nation-UK_Oct2015.pdf.
Funding
Open Access funding provided thanks to the CRUE-CSIC agreement with Springer Nature.
Author information
Authors and Affiliations
Contributions
OL designed and executed the study, analyzed the data, wrote the paper and responses to peer-review, and supervised the peer-review contributions. CGR assisted the interpretation and discussion of results, provided critical feedback and helped shape the manuscript, and assisted the responses to peer-review. SdR provided critical feedback and helped shape the manuscript, and assisted the responses to peer-review. JEMJ provided critical feedback and helped shape the manuscript, and assisted the responses to peer-review. RMML provided the data of the second sample. RRC assisted the design and execution of the study provided critical feedback and helped shape the manuscript, and assisted the responses to peer-review.
Corresponding author
Ethics declarations
Ethics Approval
All procedures performed in this study were following the ethical standards of the two institutional research committees (King Juan Carlos University; Autonomous University of Madrid) and with the Helsinki declaration and its later amendments or comparable ethical standards.
Informed Consent
Informed consent was obtained from all individual participants included in the studies.
Conflict of Interest
The authors declare no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Lecuona, O., García-Rubio, C., de Rivas, S. et al. A Network Analysis of the Five Facets Mindfulness Questionnaire (FFMQ). Mindfulness 12, 2281–2294 (2021). https://doi.org/10.1007/s12671-021-01704-7
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s12671-021-01704-7