- 1Oral and Maxillofacial Radiology, Applied Oral Sciences and Community Dental Care, Faculty of Dentistry, The University of Hong Kong, Hong Kong, China
- 2Oral and Maxillofacial Surgery, Faculty of Dentistry, The University of Hong Kong, Hong Kong, China
Introduction: There are various methods available for conducting meta-analyses of fMRI data, with coordinate-based meta-analysis (CBMA) being a frequently used approach due to the limited availability of effect size and statistical maps. Since the literature has accumulated many fMRI meta-analyses, several reports have been published to reveal the prevalence of numerous meta-analytic software packages without investigating into the recency of their versions used. To address this gap, a literature survey was conducted to identify the software packages and version numbers used for fMRI meta-analyses published between 2019 and 2024.
Methods: The online databases of Web of Science Core Collection (WOSCC) and Scopus were queried to identify relevant papers. After screening, the analysis consisted of data manually extracted from 820 papers.
Results: The most frequently used software was GingerALE (407 out of 820 papers, 49.6%), followed by SDM-PSI (27.4%) and Neurosynth (11.0%). Overall, 540 papers (65.9%) fully disclosed the names and version numbers of the software packages used in their analyses, whereas 19 papers (2.3%) reported neither the names nor the version numbers. For GingerALE, the most frequently used versions were 2.3.6 and 3.0.2, but it should be noted that versions prior to 2.3.6 have an issue of inflated false positive rates. For SDM-PSI, the most frequently used versions were 5.141, 5.15, 6.21, and 6.22, but the meta-analytic method adopted for version 6 differs from those used in prior versions.
Discussion: To ensure transparency and reproducibility in research, researchers should clearly report the name and version number of software package used.
Introduction
The fMRI literature has accumulated a number of meta-analytic studies, which used a huge variety of software packages to conduct meta-analyses. FMRI meta-analysis is necessary because individual fMRI studies often suffer from small sample sizes due to the high cost of scanning, the risk of false positives arising from multiple testing and inadequate correction methods, and variability in analysis techniques that can lead to inconsistent results (Button et al., 2013; Yeung et al., 2023).
FMRI meta-analysis can be broadly divided into two approaches: the estimation of effect size versus effect location (Fox et al., 1998). For the former, researchers may extract the effect size reported from individual studies (e.g., percent signal change from a region-of-interest), or other effect measures and process with general meta-analytic software such as Review Manager (RevMan). Since many fMRI studies reported the coordinates of the significantly activated clusters of voxels but not the effect size (Chen et al., 2017), the majority of fMRI meta-analyses focus on the latter, that is effect location, with the mainstream approach called coordinate-based meta-analysis (CBMA), aimed at identifying robust convergence of brain activation across studies. CBMA requires data of the activation foci in the format of brain coordinates, or peak coordinates, with reference to a standard space (Salimi-Khorshidi et al., 2009). Notable CBMA methods included activation likelihood estimation (ALE), commonly performed with the software package called GingerALE (Eickhoff et al., 2009; Eickhoff et al., 2012; Turkeltaub et al., 2012); multilevel kernel density analysis (MKDA), commonly performed with the MKDA toolbox (Wager et al., 2007); and Analysis of Brain Coordinates (ABC), commonly performed with the NeuRoi toolbox (Tench et al., 2017; Tench et al., 2022). It is worth noting that Neurosynth is a website/database that hosts large-scale automated CBMAs resulted from its own methodology, and allows users to type words to generate corresponding CBMAs (Yarkoni et al., 2011). CBMAs have the advantages of good data accessibility (requiring brain coordinates from published studies rather than brain maps), broad literature coverage, data standardization (brain coordinates in Talairach or MNI space), and ease of use; at the same time, they have the disadvantages of data reduction (loss of information), limited spatial precision, and publication bias (Jennings and Van Horn, 2012; Ioannidis et al., 2014; Samartsidis et al., 2017; Chen et al., 2022). On the other hand, if the collected dataset contains information at the voxel-level (e.g., t-maps from statistical parametric mapping, or SPM t-maps), then image-based meta-analysis (IBMA) can be performed. For instance, Seed-based d Mapping with Permutation of Subject Images (SDM-PSI) is a hybrid method that can pool data from studies with only peak coordinates with studies with SPM t-maps, commonly performed with the SDM-PSI toolbox (Albajes-Eizagirre et al., 2019). Please refer to Table 1 for a brief introduction of some representative fMRI meta-analytic software packages.
Some neuroimaging meta-analytic overviews have been published in the past. For example, (Acar et al., 2024) queried the PubMed database in September 2023 to identify papers that mentioned various meta-analysis methods with the term “meta-analysis” in their title or abstract. Their investigation concluded that ALE was the most frequently used method, with a peak of approximately 90 papers in 2022, to be followed by SDM-PSI and Neurosynth, with their respective peaks at approximately 40 and 30 papers in 2022, whereas few papers mentioned MKDA. There were almost no IBMA papers. Meanwhile, a study published in 2019 analyzed neuroimaging meta-analyses on gustation and found that 20 out of 23 meta-analyses used GingerALE, whereas SDM-PSI and MKDA were each used by one study only (Yeung et al., 2019). Besides, an earlier paper reported that 77% of published neuroimaging meta-analyses indexed in PubMed until 2018 used ALE method, followed by SDM (17%) based on an undisclosed search method (Tahmasian et al., 2019). The dominance of ALE and SDM methods have been reaffirmed by a recent preprint, which reported numbers of 72.7 and 21.2%, respectively, based on an analysis of 899 papers (Oudyk et al., 2025). On a separate note (Tench et al., 2022) mentioned that the most popular CBMA method is probably the ALE algorithm, pointed out that the ALE algorithm has changed over time, and proposed a new method of their own. A research gap exists because Acar et al. (2024), Tahmasian et al. (2019), and Oudyk et al. (2025) reported on the prevalence of methods but not software packages, whereas Yeung et al. (2019) reported the use of software packages for gustatory neuroimaging research only.
This work revisits the topic and provides an update that focuses on the current situation of the fMRI literature. It is important to know not only which software package was used to conduct the fMRI meta-analysis but also its version number to improve transparency and ensure reproducible research (Niso et al., 2022). This is particularly crucial with older versions of GingerALE, which contain bugs that could inflate the false positive rate before version 2.3.6 (Eickhoff et al., 2017). A software package may have a long history with multiple versions, where some updates provide minor enhancements or additional functionality, while others fix major errors or bugs, making it important for users to be aware of these changes especially the latter ones. For example, readers may refer to (Yeung, 2024) for a timeline that illustrates the various versions of GingerALE that introduced and fixed major errors. Field experts have made efforts to consolidate CBMA and IBMA methods with a simple and shared interface to reduce brand loyalty to any particular algorithm and to encourage between-method comparisons, one of which is called Neuroimaging Meta-Analysis Research Environment (NiMARE) with its initial release in November 2019 (Salo et al., 2022). Hence, it would be interesting to know whether NiMARE has had an increasing publication share over the past few years.
Given this context, the purpose of this literature survey is to identify what software packages and their version numbers that have been used for fMRI meta-analyses published from 2019 to 2024, inclusive.
Methods
The online databases of Web of Science Core Collection (WOSCC) and Scopus were queried on 9 May 2025. The title, abstract and keyword fields were searched with the following search string: (fMRI OR “functional magnetic resonance imaging” OR “functional MRI”) AND (meta-analy*). Papers were only included if they were labeled as article or review, written in English, and published during 2019–2024. The search yielded 1,065 papers from WOSCC, and 1,512 papers from Scopus. The records of the papers were compiled into a list and passed into Microsoft Excel for de-duplication with reference to the Digital Object Identifiers (DOIs) and paper titles. After de-duplication, 1,618 papers remained. Among these 1,618 papers, 792 were excluded due to no meta-analysis on fMRI data after accessing their full text, 5 were excluded due to no access to the full text, and 1 was excluded due to article retraction. For the remaining 820 papers that entered the analysis, the name and version number of the software packages used to perform fMRI meta-analysis were manually extracted, if present (Figure 1). Both authors independently conducted the paper screening and data extraction. Any disagreements were resolved through discussion to reach a mutual consensus.
Results
The coded data sheet was provided as the Supplementary File 1. Among the 820 papers that conducted fMRI meta-analyses, some used multiple software programs. As a result, the total count of software usage reached 853. The most frequently used software was GingerALE (407 out of 820 papers, 49.6%), followed by SDM-PSI (27.4%) and Neurosynth (11.0%) (Figure 2).

Figure 2. Frequency count of different software packages used to perform fMRI meta-analysis. Out of the 820 papers that performed fMRI meta-analysis, some used more than one software. Therefore, the total frequency count amounted to 853.
Meanwhile, Table 2 shows the frequency counts per year. The temporal trends of the 3 software packages with the highest publication share are as follows. The annual publication share of GingerALE was always slightly over one-half until 2023, when it dropped to 47.4% and further to 42.6% in 2024. The publication share of SDM-PSI increased every year from 20.8% in 2019 to 32.3% in 2024. Neurosynth had approximately 10% publication share per year during 2019–2024.
Overall, 540 papers (65.9%) fully disclosed the names and version numbers of the software packages used in their analyses (Table 3). In contrast, 19 papers (2.3%) reported neither the names nor the version numbers of the software packages used. The most frequently used versions for GingerALE were 2.3.6 and 3.0.2, whereas the most frequently used versions for SDM-PSI were 5.141, 5.15, 6.21, and 6.22 (Tables 4, 5).

Table 3. Breakdown of the disclosure of the names and version numbers of software packages used in the analyzed papers.
Among GingerALE papers, the publication share of version 2.3.6 (released on 26 April 2016) decreased from 76.4% in 2019 to 7.6% in 2024, as the share of version 3.0.2 (released on 16 May 2019) increased from 3.6% in 2019 to 81.8% in 2024 (Figure 3). Their changes were most prominent from 2019 to 2021, and became more mild since then. It is worth noting that two versions of GingerALE, namely versions 2.3.7 and 3.0.3, were each reported in one paper, but they are not officially documented. Furthermore, 15 papers (3.7% of 407) used older versions of GingerALE (prior to 2.3.6), which suggests a higher risk of false positive results.

Figure 3. Change in the annual publication share of commonly used versions of GingerALE. “? Version” refers to papers that did not report the exact version number. The publication share is computed from GingerALE papers, not all meta-analysis papers.
Among SDM-PSI papers, version 5.141 (released in December 2016) was the dominant version in 2019 and 2020 (at least 20%), but the publication shares of versions 5.15 (released in April 2018) and 6.21 (released in November 2019) gradually increased and peaked at 38.8 and 40.8% in 2023, respectively (Figure 4). The publication share of versions 5.15 and 6.21 showed clear decline from 2023 to 2024, during which the share of version 6.22 showed a sharp increase from 4.1 to 14.0%. The ratio of SDM-PSI papers that did not report the version number dropped from 47.6% in 2019 to 36.0% in 2024.

Figure 4. Change in the annual publication share of commonly used versions of SDM-PSI. “? Version” refers to papers that did not report the exact version number. The publication share is computed from SDM-PSI papers, not all meta-analysis papers.
Discussion
This literature survey has confirmed that GingerALE was the most frequently used software package for fMRI meta-analysis whereas the publication share of SDM-PSI has continued to increase from 2019 to 2024. At the same time, more and more papers have used other software packages such as Neurosynth and NiMARE. The fMRI meta-analytic literature has been dominated by these 2 CBMA meta-analytic software packages. Foundational work by Turkeltaub et al. (2002) and Chein et al. (2002), summarized in Laird et al. (2005), established the methodological framework for ALE, emphasizing their ability to integrate coordinate-based data for cross-study comparisons. After testing convergence of activation across studies, methodological advancements have enabled meta-analytic connectivity modeling (MACM) (Robinson et al., 2010) and co-activation-based parcellation (CBP) (Eickhoff et al., 2011). It is unclear why the prevalence of SDM-PSI was on the rise over the last few years, but it does have advantages over ALE in terms of effect size integration and handling of both positive and negative findings (Albajes-Eizagirre and Radua, 2018; Albajes-Eizagirre et al., 2019). However, no method is perfect. The dominance of one or two methods may reduce the incentive for researchers to develop alternative methods, and novice users may be more hesitant to try these alternatives.
A minority of papers that used GingerALE have either reported the use of older versions (prior to 2.3.6) or not reported the version number. This is not a good phenomenon, as the developer team of GingerALE has already released a debugged version in 2016 that fixed some major issues with inflated false positive rate, with a clear documentation published in 2017 (Eickhoff et al., 2017). As the surveyed meta-analytic papers were all published since 2019, ideally all of them should be using newer versions of GingerALE that have been debugged. The effects on the results could be potentially huge. For instance, an fMRI meta-analysis on perceptual decision making that used GingerALE version 2.3 originally reported 10 clusters with significant activation for a contrast of task > control (Keuken et al., 2014). After re-analyzing data with a debugged version of 2.3.6, all 10 clusters became non-significant (Keuken et al., 2017). Other re-analyses showed less dramatic changes to the initial conclusions, such as a reduction from 36 significant clusters (Belyk et al., 2015) to 4 (Belyk et al., 2017), from 12 significant clusters (Garrigan et al., 2016) to 8 (Garrigan et al., 2016), and an exclusion of three previously significant brain regions (unclear number of significant clusters) (Smith et al., 2016; Smith and Delgado, 2017). Readers should be aware that the examples listed here may be affected by publication bias. It could be possible that many re-analyses showing no significant changes from the original findings ended up as unpublished findings.
Though the issue of inflated false positive rate has not been reported for SDM-PSI, it is still advisable to report version number and use most updated version of the software package whenever possible due to a different issue. A major change of SDM-PSI from version 5 to version 6 was the adoption of permutation so that it tested whether effects were non-null in a given voxel, but no longer tested whether findings across studies tended to converge around it (Albajes-Eizagirre and Radua, 2018; Albajes-Eizagirre et al., 2019). According to the Readme document of SDM-PSI version 6.23, the first officially documented version 6 of SDM-PSI was version 6.11 released in April 2019. Ideally, most of the surveyed meta-analysis papers should have used version 6.11 or newer versions, except those papers published in early 2019. It was largely unclear why the publication share of version 5.15 (released in April 2018) continued to increase year by year and nearly matched the publication share of version 6.21 (released in November 2019). In 2024, though, both versions declined whereas versions 6.22 and 6.23 began to get some publication share. Comparing meta-analytic results between studies using the old (version 5 or older) and new versions (version 6) of SDM-PSI may be conceptually inaccurate due to the fundamentally different methodology. Though recent guidelines for conducting and reporting neuroimaging meta-analysis did not explicitly recommend researchers to list the exact software version number during reporting (Müller et al., 2018; Tahmasian et al., 2019; Manuello et al., 2022), this general practice is highly recommended in reporting an fMRI study (Poldrack et al., 2008).
On a separate note, not many studies in the dataset used NiMARE to conduct fMRI meta-analysis. The concept behind NiMARE is very good: An environment/ecosystem that allows users to conduct multiple types of meta-analyses based on a number of algorithms available from the literature and interact with online databases of brain coordinates and fMRI images. However, since NiMARE runs on Python and is operated through either command line interface (CLI) or application programing interface (API), some researchers may prefer to use existing toolboxes or software packages with a guided user interface (GUI) that allow for operation by pressing buttons, rather than typing computer code. Specifically, those who possess extensive expertise in the domains being meta-analyzed, such as physiology, psychology, or pathology, may not have the programing skills necessary to efficiently utilize CLI or API. This is particularly relevant given the interdisciplinary nature of neuroimaging research, where specialists from diverse backgrounds collaborate. The preference for GUIs is evident in the continued use of software packages such as GingerALE and SDM-PSI. However, analyses with CLI may have better provenance and reproducibility than with GUI, as users may press buttons wrongly and there is usually no action log. Some surveyed studies in the dataset opted to use both GingerALE and SDM-PSI to conduct CBMAs and compare results instead of performing them “centrally” through NiMARE. Hence, NiMARE and similar solutions could consider to build a simple GUI besides command line interface, to attract users from a more diverse background.
This study has some limitations. First, the analysis was restricted to papers published in a short period. The current findings may not be readily applied to the past literature. Second, some papers not indexed by WOSCC or Scopus might be omitted by this study, though they might only account for a small share of the literature. Third, manual data extraction is labor demanding and may be prone to human mistakes. The future development of text scraper with artificial intelligence and availability of literature full texts with open access should enable a replication of this study in a much larger-scale. Future studies can also consider more advanced analyses, such as to reveal any relationship between the impact factors of the journals and software selection.
Conclusion
Based on a literature survey of papers published during 2019–2024, GingerALE was the most frequently used software package for conducting fMRI meta-analysis, followed by SDM-PSI and Neurosynth. The issue of not reporting the version number of the software package was more serious for SDM-PSI than GingerALE, but the situation has been improving quickly for the former. Another issue for using SDM-PSI was the continued use of its old version together with its latest version. Unlike the old versions, its latest version (version 6) uses permutation to test whether effects were non-null in a given voxel, instead of whether findings across studies tended to converge around it. The methodologies to meta-analyze fMRI data are diverse. Researchers should clearly report the details including the name and version number of software packages used, so that readers can better understand what has been done and how results could be compared across studies.
Data availability statement
The original contributions presented in the study are included in the article/Supplementary material, further inquiries can be directed to the corresponding authors.
Author contributions
AY: Writing – original draft, Writing – review and editing. NW: Writing – review and editing.
Funding
The author(s) declare that no financial support was received for the research and/or publication of this article.
Conflict of interest
AY is an editorial board member of Frontiers at the time of submission.
The remaining author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The authors declare that no Generative AI was used in the creation of this manuscript.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fnhum.2025.1580808/full#supplementary-material
References
Acar, F., Maumet, C., Heuten, T., Vervoort, M., Bossier, H., Seurinck, R., et al. (2024). Improving the eligibility of task-based fMRI studies for meta-analysis: A review and reporting recommendations. Neuroinformatics 22, 5–22. doi: 10.1007/s12021-023-09643-5
Albajes-Eizagirre, A., and Radua, J. (2018). What do results from coordinate-based meta-analyses tell us? Neuroimage 176, 550–553. doi: 10.1016/j.neuroimage.2018.04.065
Albajes-Eizagirre, A., Solanes, A., Vieta, E., and Radua, J. (2019). Voxel-based meta-analysis via permutation of subject images (PSI): Theory and implementation for SDM. Neuroimage 186, 174–184. doi: 10.1016/j.neuroimage.2018.10.077
Belyk, M., Kraft, S., and Brown, S. (2015). Stuttering as a trait or state – an ALE meta-analysis of neuroimaging studies. Eur. J. Neurosci. 41, 275–284. doi: 10.1111/ejn.12765
Belyk, M., Kraft, S., and Brown, S. (2017). Stuttering as a trait or a state revisited: Motor system involvement in persistent developmental stuttering. Eur. J. Neurosci. 45, 622–624. doi: 10.1111/ejn.13512
Button, K., Ioannidis, J., Mokrysz, C., Nosek, B., Flint, J., Robinson, E., et al. (2013). Power failure: Why small sample size undermines the reliability of neuroscience. Nat. Rev. Neurosci. 14, 365–376. doi: 10.1038/nrn3475
Chein, J., Fissell, K., Jacobs, S., and Fiez, J. (2002). Functional heterogeneity within Broca’s area during verbal working memory. Physiol. Behav. 77, 635–639. doi: 10.1016/s0031-9384(02)00899-5
Chen, G., Taylor, P., and Cox, R. (2017). Is the statistic value all we should care about in neuroimaging? Neuroimage 147, 952–959. doi: 10.1016/j.neuroimage.2016.09.066
Chen, G., Taylor, P., Stoddard, J., Cox, R., Bandettini, P., and Pessoa, L. (2022). Sources of information waste in neuroimaging: Mishandling structures, thinking dichotomously, and over-reducing data. Apert Neuro 2:2. doi: 10.52294/apertureneuro.2022.2.zrji8542
Eickhoff, S., Bzdok, D., Laird, A., Kurth, F., and Fox, P. (2012). Activation likelihood estimation meta-analysis revisited. Neuroimage 59, 2349–2361. doi: 10.1016/j.neuroimage.2011.09.017
Eickhoff, S., Bzdok, D., Laird, A., Roski, C., Caspers, S., Zilles, K., et al. (2011). Co-activation patterns distinguish cortical modules, their connectivity and functional differentiation. Neuroimage 57, 938–949. doi: 10.1016/j.neuroimage.2011.05.021
Eickhoff, S., Laird, A., Fox, P., Lancaster, J., and Fox, P. (2017). Implementation errors in the GingerALE Software: Description and recommendations. Hum. Brain Mapp. 38, 7–11. doi: 10.1002/hbm.23342
Eickhoff, S., Laird, A., Grefkes, C., Wang, L., Zilles, K., and Fox, P. (2009). Coordinate-based activation likelihood estimation meta-analysis of neuroimaging data: A random-effects approach based on empirical estimates of spatial uncertainty. Hum. Brain Mapp. 30, 2907–2926. doi: 10.1002/hbm.20718
Fox, P., Parsons, L., and Lancaster, J. (1998). Beyond the single study: Function/location metanalysis in cognitive neuroimaging. Curr. Opin. Neurobiol. 8, 178–187. doi: 10.1016/s0959-4388(98)80138-4
Garrigan, B., Adlam, A., and Langdon, P. (2016). The neural correlates of moral decision-making: A systematic review and meta-analysis of moral evaluations and response decision judgements. Brain Cogn. 108, 88–97. doi: 10.1016/j.bandc.2016.07.007
Ioannidis, J., Munafò, M., Fusar-Poli, P., Nosek, B., and David, S. (2014). Publication and other reporting biases in cognitive sciences: Detection, prevalence, and prevention. Trends Cogn. Sci. 18, 235–241. doi: 10.1016/j.tics.2014.02.010
Jennings, R., and Van Horn, J. (2012). Publication bias in neuroimaging research: Implications for meta-analyses. Neuroinformatics 10, 67–80. doi: 10.1007/s12021-011-9125-y
Keuken, M., Müller-Axt, C., Langner, R., Eickhoff, S., Forstmann, B., and Neumann, J. (2014). Brain networks of perceptual decision-making: An fMRI ALE meta-analysis. Front. Hum. Neurosci. 8:445. doi: 10.3389/fnhum.2014.00445
Keuken, M., Müller-Axt, C., Langner, R., Eickhoff, S., Forstmann, B., and Neumann, J. (2017). Corrigendum: Brain networks of perceptual decision-making: An fMRI ALE meta-analysis. Front. Hum. Neurosci. 11:139. doi: 10.3389/fnhum.2017.00139
Laird, A., Fox, P., Price, C., Glahn, D., Uecker, A., Lancaster, J., et al. (2005). ALE meta-analysis: Controlling the false discovery rate and performing statistical contrasts. Hum. Brain Mapp. 25, 155–164. doi: 10.1002/hbm.20136
Manuello, J., Costa, T., Cauda, F., and Liloia, D. (2022). Six actions to improve detection of critical features for neuroimaging coordinate-based meta-analysis preparation. Neurosci. Biobehav. Rev. 137:104659. doi: 10.1016/j.neubiorev.2022.104659
Müller, V., Cieslik, E., Laird, A., Fox, P., Radua, J., Mataix-Cols, D., et al. (2018). Ten simple rules for neuroimaging meta-analysis. Neurosci. Biobehav. Rev. 84, 151–161. doi: 10.1016/j.neubiorev.2017.11.012
Niso, G., Botvinik-Nezer, R., Appelhoff, S., De La Vega, A., Esteban, O., Etzel, J., et al. (2022). Open and reproducible neuroimaging: From study inception to publication. Neuroimage 263:119623. doi: 10.1016/j.neuroimage.2022.119623
Oudyk, K., Dockès, J., Peraza, J., Kent, J., Torabi, M., Wang, M., et al. (2025). Meta all the way down: An overview of neuroimaging meta-analyses. Neuroimage 263:119623. doi: 10.1101/2025.1101.1121.634094
Poldrack, R., Fletcher, P., Henson, R., Worsley, K., Brett, M., and Nichols, T. (2008). Guidelines for reporting an fMRI study. Neuroimage 40, 409–414. doi: 10.1016/j.neuroimage.2007.11.048
Radua, J., and Mataix-Cols, D. (2009). Voxel-wise meta-analysis of grey matter changes in obsessive-compulsive disorder. Br. J. Psychiatry 195, 393–402. doi: 10.1192/bjp.bp.108.055046
Robinson, J., Laird, A., Glahn, D., Lovallo, W., and Fox, P. (2010). Metaanalytic connectivity modeling: Delineating the functional connectivity of the human amygdala. Hum. Brain Mapp. 31, 173–184. doi: 10.1002/hbm.20854
Salimi-Khorshidi, G., Smith, S., Keltner, J., Wager, T., and Nichols, T. (2009). Meta-analysis of neuroimaging data: A comparison of image-based and coordinate-based pooling of studies. Neuroimage 45, 810–823. doi: 10.1016/j.neuroimage.2008.12.039
Salo, T., Yarkoni, T., Nichols, T. E., Poline, J.-B., Bilgel, M., Bottenhorn, K. L., et al. (2022). NiMARE: Neuroimaging meta-analysis research environment. NeuroLibre 1:7. doi: 10.52294/001c.87681
Samartsidis, P., Montagna, S., Nichols, T., and Johnson, T. (2017). The coordinate-based meta-analysis of neuroimaging data. Stat. Sci. 32, 580–599. doi: 10.1214/17-STS624
Smith, D., and Delgado, M. (2017). Meta-analysis of psychophysiological interactions: Revisiting cluster-level thresholding and sample sizes. Hum. Brain Mapp. 38, 588–591. doi: 10.1002/hbm.23354
Smith, D., Gseir, M., Speer, M., and Delgado, M. (2016). Toward a cumulative science of functional integration: A meta-analysis of psychophysiological interactions. Hum. Brain Mapp. 37, 2904–2917. doi: 10.1002/hbm.23216
Tahmasian, M., Sepehry, A., Samea, F., Khodadadifar, T., Soltaninejad, Z., Javaheripour, N., et al. (2019). Practical recommendations to conduct a neuroimaging meta-analysis for neuropsychiatric disorders. Hum. Brain Mapp. 40, 5142–5154. doi: 10.1002/hbm.24746
Tench, C., Tanasescu, R., Constantinescu, C., Auer, D., and Cottam, W. (2022). Easy to interpret coordinate based meta-analysis of neuroimaging studies: Analysis of brain coordinates (ABC). J. Neurosci. Methods 372, 109556. doi: 10.1016/j.jneumeth.2022.109556
Tench, C., Tanasescu, R., Constantinescu, C., Auer, D., and Cottam, W. (2017). Coordinate based random effect size meta-analysis of neuroimaging studies. Neuroimage 153, 293–306. doi: 10.1016/j.neuroimage.2017.04.002
Turkeltaub, P., Eden, G., Jones, K., and Zeffiro, T. (2002). Meta-analysis of the functional neuroanatomy of single-word reading: Method and validation. Neuroimage 16, 765–780. doi: 10.1006/nimg.2002.1131
Turkeltaub, P., Eickhoff, S., Laird, A., Fox, M., Wiener, M., and Fox, P. (2012). Minimizing within-experiment and within-group effects in activation likelihood estimation meta-analyses. Hum. Brain Mapp. 33, 1–13. doi: 10.1002/hbm.21186
Wager, T., Lindquist, M., and Kaplan, L. (2007). Meta-analysis of functional neuroimaging data: Current and future directions. Soc. Cogn. Affect. Neurosci. 2, 150–158. doi: 10.1093/scan/nsm015
Yarkoni, T., Poldrack, R., Nichols, T., Van Essen, D., and Wager, T. (2011). Large-scale automated synthesis of human functional neuroimaging data. Nat. Methods 8, 665–670. doi: 10.1038/nmeth.1635
Yeung, A. (2024). The reverberation of implementation errors in a neuroimaging meta-analytic software package: A citation analysis to a technical report on GingerALE. Heliyon 10:e38084. doi: 10.1016/j.heliyon.2024.e38084
Yeung, A., Robertson, M., Uecker, A., Fox, P., and Eickhoff, S. (2023). Trends in the sample size, statistics, and contributions to the BrainMap database of activation likelihood estimation meta-analyses: An empirical study of 10-year data. Hum. Brain Mapp. 44, 1876–1887. doi: 10.1002/hbm.26177
Keywords: coordinate based meta-analysis, GingerALE, SDM, neurosynth, NiMARE
Citation: Yeung AWK and Wong NSM (2025) Which software packages did researchers use to meta-analyze fMRI data? A literature survey from 2019 to 2024. Front. Hum. Neurosci. 19:1580808. doi: 10.3389/fnhum.2025.1580808
Received: 21 February 2025; Accepted: 26 June 2025;
Published: 10 July 2025.
Edited by:
Changming Wang, Capital Medical University, ChinaCopyright © 2025 Yeung and Wong. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Andy Wai Kan Yeung, bmR5ZXVuZ0Boa3UuaGs=