Your new experience awaits. Try the new design now and help us make it even better

SYSTEMATIC REVIEW article

Front. Artif. Intell., 01 October 2025

Sec. AI in Finance

Volume 8 - 2025 | https://doi.org/10.3389/frai.2025.1658375

This article is part of the Research TopicImplementing Anti-Financial Crime Risk Control Measures Using Artificial Intelligence: Challenges for Advanced Economies and Emerging MarketsView all 3 articles

Big data in financial risk management: evidence, advances, and open questions: a systematic review

  • Department of Management Science and Technology, Panepistemio Patron, Patras, Greece

Introduction: The intersection of big data analytics and financial risk management has spurred significant methodological innovation and organizational change. Despite growing research activity, the literature remains fragmented, with notable gaps in comparative effectiveness, cross-sectoral applicability, and the use of non-traditional data sources.

Methods: Following the PRISMA 2020 protocol, a systematic review was conducted on 21 peer-reviewed studies published between 2016 and June 2025. The review evaluated the methodological diversity and effectiveness of machine learning and hybrid approaches in financial risk management.

Results: The analysis mapped the relative strengths and limitations of neural networks, ensemble learning, fuzzy logic, and hybrid optimization across credit, fraud, systemic, and operational risk. Advanced machine learning techniques consistently demonstrated strong predictive accuracy, yet real-world deployment remained geographically concentrated, primarily in Chinese and European banking and fintech sectors. Applications involving alternative and unstructured data, such as IoT signals and behavioral analytics, were largely experimental and faced both technical and governance challenges.

Discussion/conclusion: The findings underscore the scarcity of systematic benchmarking across risk types and organizational contexts, as well as the limited attention to explainability in current implementations. This review identifies an urgent need for comparative, cross-jurisdictional studies, stronger field validation, and open science practices to bridge the gap between technical advances and their operational impact in big data–enabled financial risk management.

1 Introduction

Financial markets and institutions today operate amid unprecedented volatility and complexity, driven by the proliferation of digital technologies and the relentless expansion of data (Theodorakopoulos et al., 2022). The inadequacy of legacy risk management frameworks has been laid bare by a string of high-profile disruptions, from cyberattacks and algorithmic trading shocks to the systemic reverberations of global crises. In this turbulent environment, the intersection of big data analytics and financial risk management has emerged as both a crucible of innovation and a source of unresolved tension, offering transformative potential but also surfacing profound methodological and practical challenges (Osman and El-Gendy, 2025).

1.1 Research gaps

Current research on big data applications in financial risk management exhibits persistent fragmentation, with most studies constrained by methodological insularity and narrow empirical scope. The literature overwhelmingly prioritizes technical innovation; benchmarks of novel machine learning models, ensemble methods, or optimization techniques, while seldom evaluating their comparative effectiveness across diverse financial risk categories or organizational contexts (Addy et al., 2024). Real-world validation is frequently sidelined, as empirical analyses rely on proprietary, simulated, or narrowly scoped datasets. The result is a conspicuous absence of robust evidence on scalability, deployment challenges, or cross-sectoral generalizability (Nguyen et al., 2023). Moreover, although the promise of integrating non-traditional data sources such as IoT (Internet of Things) streams, social media, or behavioral signals is often acknowledged, few studies operationalize this fusion in a manner that meaningfully advances predictive accuracy or risk governance. Equally striking is the literature’s neglect of explainability, interpretability, and practical adoption. Technical performance metrics are foregrounded, yet questions of model transparency, managerial usability, and regulatory compliance are largely unaddressed (Hilbert and Darmon, 2020). There is a near-total lack of cross-jurisdictional or comparative regulatory analysis, leaving the field ill-equipped to inform global best practices or anticipate sector-specific policy impacts.

This review directly addresses the most pressing and empirically tractable gaps by systematically mapping and synthesizing recent advances in: (i) the comparative strengths and limitations of major big data techniques across financial risk categories; (ii) the extent and nature of real-world deployment and scalability; (iii) the operationalization and impact of non-traditional data integration; and (iv) the influence of regulatory, geographical, and sectoral contexts on adoption and effectiveness. However, while the review highlights the critical importance of explainability, interpretability, and practical adoption, it recognizes that the current evidence base remains too limited to support a comprehensive synthesis or actionable guidance on these dimensions. By clarifying both the boundaries and the core contributions of this systematic review, the work aims to provide a transparent, critical foundation for both scholarly advancement and future research priorities in big data-enabled financial risk management.

1.2 Aim of the review and research questions

This review aims to systematically map, compare, and critically synthesize the recent empirical and conceptual advances in the application of big data analytics to financial risk management. The overarching goal is to clarify which big data-driven techniques are most effective across different risk types and sectors, to assess the extent of real-world deployment and scalability, to evaluate how the integration of non-traditional and unstructured data can enhance risk prediction, and to illuminate the influence of regulatory, geographical, and sectoral contexts. In pursuing these objectives, this review is guided by the following four research questions:

RQ1: What are the comparative strengths, limitations, and practical trade-offs of different big data-driven analytical techniques (such as neural networks, ensemble machine learning, fuzzy logic, and information fusion) in managing various categories of financial risk (credit, fraud, systemic, and operational) across sectors?

RQ2: How do real-world applications of big data and AI models in financial risk management perform when deployed at scale, and what challenges or gaps remain regarding generalizability, data diversity, and integration with organizational processes?

RQ3: To what extent does the integration of non-traditional and unstructured data sources—such as IoT signals, social media, and behavioral analytics—enhance the predictive accuracy and early warning capabilities of financial risk models, and what barriers persist in achieving widespread adoption?

RQ4: What regulatory, geographical, or sectoral differences shape the adoption, effectiveness, and governance of big data techniques for financial risk management, and where do current studies fail to provide comparative or global perspectives?

By addressing these questions through a rigorous synthesis of 21 recent studies, this review seeks to advance both scholarly understanding and practical innovation in big data-enabled financial risk management. Unlike prior reviews, which have tended to focus narrowly on either technical innovation or sector-specific case studies, this work offers a comprehensive, comparative mapping across methods, risk types, data modalities, and organizational contexts. By foregrounding not only algorithmic advances but also barriers to real-world adoption and cross-jurisdictional applicability, this review establishes a broader, more integrated agenda for both research and practice.

1.3 Structure of the paper

Charting a clear path through this multidisciplinary and rapidly evolving terrain requires both conceptual clarity and methodological precision. To that end, the paper is structured to guide the reader from foundational concepts to analytical synthesis and actionable insight. Section 2 establishes the conceptual framework, tracing the evolution of big data and its integration into financial risk management. Section 3 sets out the materials and methods, detailing the systematic review protocol, search strategy, and data extraction process. Section 4 presents the results, mapped to the four guiding research questions and reinforced by visual and tabular evidence. Section 5 engages in critical discussion, weaving together methodological, sectoral, and practical perspectives. Section 6 delineates the study’s limitations, while Section 7 articulates key priorities for future research. Finally, Section 8 concludes by distilling the review’s broader implications for scholarship, industry practice, and policy in the data-driven financial landscape.

2 Literature review

2.1 Big data: from 3 Vs to 8 Vs and beyond

Big data emerged not from theoretical abstraction, but from organizations’ desperate attempts to navigate a deluge of information that rendered traditional analytics and management tools obsolete. Early scholars and practitioners distilled this phenomenon into three foundational dimensions: Volume, capturing the unprecedented scale of digital records; Velocity, reflecting the relentless inflow of data that often arrives in real time or near real time; and Variety, signaling the proliferation of data types, from tidy relational tables to sprawling unstructured formats like text, images, and streaming sensor feeds. These original “3 Vs” offered a new vocabulary for a technological transformation, but practice soon revealed that the challenge was far deeper (Theodorakopoulos et al., 2024).

As digital ecosystems expanded and organizational ambition grew, so did the need to capture the full spectrum of big data’s complexity. The paradigm evolved, moving well beyond the initial triad to encompass five additional attributes: Veracity, acknowledging the constant struggle with data uncertainty, errors, and bias; Value, focusing analytical attention on extracting actionable insights and meaningful outcomes rather than collecting data for its own sake; Variability, recognizing the fluctuating meaning, structure, and context of data as it moves across systems and time; Visualization, underlining the importance of rendering complex analytical outputs comprehensible and usable for decision-makers; and Validity, the critical assurance that data and derived results are trustworthy, accurate, and truly reflect the phenomena under analysis. Together, these eight Vs encapsulate not only the technical hurdles of handling massive, messy, and dynamic datasets, but also the strategic, organizational, and epistemological imperatives now entwined with data-driven innovation (Ezzahra et al., 2019).

Today, the big data landscape is characterized by this multidimensionality. Organizations face not just a flood of data, but a tangled web of challenges around integrating unstructured sources, safeguarding accuracy and reliability, and bridging the gap between algorithmic power and managerial judgment. The rise of distributed computing frameworks, the proliferation of sophisticated machine learning architectures, and the automation of data preparation pipelines have redefined the art of the possible; yet the real advantage lies not in technical wizardry alone, but in the persistent work of transforming raw, multidimensional data into decisions that are both credible and consequential. In this evolving landscape, the 8 Vs serve less as a checklist than as a reminder that every new technical advance must still reckon with questions of meaning, context, and trust (Mumuni and Mumuni, 2024).

Finally, we make explicit that data vulnerability is intrinsic to big data in finance. Data are not only voluminous and heterogeneous, they are also targets and moving parts. They can be leaked, tampered with, or subtly poisoned at any stage of the pipeline, from collection to labeling to deployment. In this review we treat vulnerability as part of Veracity and Validity, and we use that lens when discussing provenance, access control, and continuous monitoring across later sections (Figure 1).

Figure 1
Diagram of a gear labeled

Figure 1. 8 Vs of big data.

2.2 Financial risk management: core concepts and mitigation strategies

Financial risk management has long stood as a pillar of prudent organizational governance, a discipline defined by its relentless preoccupation with uncertainty and its commitment to safeguarding value in the face of volatility. At its core, the field is anchored in a cycle of anticipation, quantification, surveillance, and intervention; a perpetual process designed to identify risks before they crystallize into losses. Traditionally, risk managers have focused on a taxonomy that includes credit risk, the perennial challenge of counterparties defaulting on their obligations; market risk, the threat posed by unpredictable fluctuations in prices, rates, or currencies; operational risk, a sprawling category that encompasses breakdowns in processes, systems, and even human error; liquidity risk, which can leave otherwise healthy organizations paralyzed in times of stress; and systemic risk, the lurking specter of contagion that can transform local shocks into global crises (Huynh et al., 2025).

Mitigation of these diverse risks has given rise to a formidable toolkit. Statistical models, from classical regression to more intricate time series and copula-based approaches, have served as the analytical backbone for decades. Scenario analysis and stress testing have evolved into sophisticated exercises, allowing managers to peer into the fog of potential futures and assess resilience under extreme, if improbable, conditions. Value-at-risk (VaR), though not without controversy, remains a lingua franca for risk quantification, providing a common metric for both regulators and practitioners. Regulatory compliance frameworks, shaped by successive waves of Basel Accords and the increasingly intricate directives of the European Union and other bodies, now exert immense influence, not merely as constraints but as drivers of innovation in risk measurement and reporting. Insurance, derivatives, and risk transfer mechanisms form the second line of defense, while robust internal controls, governance structures, and audit practices round out the institutional response (Lentini and Thayasivam, 2025).

Yet, this traditional machinery is under growing strain. The past decade has forced risk managers to grapple with threats that defy familiar categorization: cyber risk, where attacks can materialize from distant actors and propagate invisibly across networks; reputational risk, magnified by the velocity and reach of digital media; and the emergent, tangled webs of interdependence that characterize modern financial ecosystems. The collapse of a single node, whether due to fraud, error, or malfeasance, can ripple outward, challenging the very assumptions upon which risk models were built. At the same time, the regulatory environment has become both more demanding and more dynamic. Basel III, for example, has not only tightened capital requirements and stress test protocols but also heightened expectations for data quality, model governance, and risk transparency. Parallel initiatives in data privacy and digital finance are raising the bar for how risk data is managed, reported, and secured (González García and Álvarez-Fernández, 2022).

Against this backdrop, the role of risk analytics is being fundamentally reimagined. Static, backward-looking models are being displaced by agile, real-time analytics that can ingest vast, heterogeneous datasets and generate early warning signals before conventional indicators even twitch. The competitive imperative is no longer just to comply, but to anticipate, adapt, and innovate, transforming risk management from a reactive function into a strategic engine for resilience and value creation. This transition, however, is far from straightforward; it brings its own risks, from model overfitting and opacity to data governance and ethical dilemmas. The next era of financial risk management will be shaped by how effectively organizations can navigate these new uncertainties, marrying technological capability with judgment, foresight, and a relentless commitment to trustworthiness (Rozony et al., 2024).

2.3 The intersection: big data in financial risk management

The meeting point of big data analytics and financial risk management has become one of the defining frontiers in both domains—a collision of scale, speed, and sophistication that is fundamentally reshaping how organizations anticipate, interpret, and respond to risk. No longer bound by the limits of periodic reporting or narrow, siloed data streams, risk managers today can at least in theory draw from oceans of information: high-frequency market data, transactional histories, IoT device telemetry, social sentiment, and real-time macroeconomic signals. This vast, heterogeneous influx is neither orderly nor straightforward, yet it holds the raw potential to illuminate risks that once lurked in statistical shadows (Kalashnikov and Kartbayev, 2024). At the same time, this capability comes with a wider attack and failure surface. When models ingest streaming or unstructured sources, vulnerability becomes a design constraint rather than a peripheral concern, affecting feature stability, provenance, and the trust placed in any downstream decision.

Machine learning, deep learning, and other AI-powered methods have dramatically altered the risk analytics landscape. Rather than relying solely on regression or scenario-based stress tests, contemporary approaches can identify subtle patterns, anomalies, or outlier behaviors that might signal emerging fraud, credit deterioration, or liquidity shortfalls often surfacing weak signals invisible to classic models. The embrace of neural networks and ensemble techniques has made it possible to process and learn from both structured records and unstructured data, unlocking predictive and diagnostic capabilities at a scale previously unthinkable. Network analytics, meanwhile, now enable the mapping of interdependencies and contagion pathways within and across institutions, revealing vulnerabilities that are systemic rather than merely idiosyncratic. Information fusion bringing together disparate data sources further amplifies analytic depth, supporting early-warning systems and dynamic risk scoring in environments of profound uncertainty (Udeh et al., 2024).

At the same time, big data analytics have injected a distinct behavioral and sentiment-driven layer into the architecture of risk management, gradually challenging the long-standing dominance of hard financial ratios and macroeconomic indicators. For the first time, market sentiment, the rhythm of public discourse, and subtle inflections in consumer behavior can be systematically quantified, tracked, and mapped to evolving risk profiles. This infusion of “soft data” promises explanatory insights that classic numerical models could not approach, allowing for earlier detection of market shifts or emerging crises fueled by rumor, panic, or collective exuberance. Yet this expansion is not without cost; questions of causality become knottier, as signals drawn from news feeds or social platforms can be clouded by noise, bias, and fleeting trends. The interpretability of such hybrid models, blending qualitative nuance with quantitative rigor, remains fiercely debated, as organizations confront the possibility that algorithmic complexity might obscure, rather than clarify, critical risk signals (Tiwari et al., 2025).

The pursuit of more holistic intelligence, blending structured fundamentals with a mosaic of alternative data sources, is now a focal point of both academic inquiry and industry innovation. It is precisely in this context that the need for systematic synthesis becomes urgent. The literature is fragmented; some studies chase technical novelty, others focus on deployment, but few map the full landscape of methodologies, application domains, and practical challenges. As the boundaries of financial risk management blur and the appetite for predictive, multidimensional analytics grows, a rigorous, evidence-driven assessment of the field is essential. This imperative shaped the present review’s methodological design, motivating the adoption of a transparent, protocol-driven approach that could cut through the complexity, evaluate the comparative strengths of emerging techniques, and expose persistent gaps. The following section details the Materials and Methods underpinning this synthesis, setting out the systematic process by which relevant literature was identified, screened, and mapped in service of a clearer, more integrated understanding of big data’s role in the evolving risk management landscape.

3 Materials and methods

This review adopts a systematic and transparent methodological framework, rooted in the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) 2020 protocol (Page et al., 2021), to critically map the intersection of big data analytics and financial risk management. Recognizing the rapid methodological evolution and sectoral expansion of this research domain, our approach emphasizes both breadth and analytical depth. PRISMA 2020 is used to transparently document the review process and decisions as well as support the construction of a coherent evidence base capable of addressing both technical and contextual research questions. Methodological rigor was further reinforced by the use of a five-step process, visually summarized in Figure 2, that guided each phase from topic formulation to evidence extraction and synthesis. The review’s methodological design was explicitly tailored to surface not only dominant trends and high-performing algorithms, but also latent gaps, sectoral blind spots, and barriers to real-world impact.

Figure 2
Flowchart illustrating research steps: 1. Define Topic, 2. Develop Research Questions, 3. Identify Keywords, 4. Identify & Search Databases, 5. Access & Read Articles. A person holds a magnifying glass near a chart.

Figure 2. 5-step PRISMA process.

All stages of the review process were executed sequentially and iteratively to maximize both coverage and relevance. The process began with the precise definition of the review’s thematic focus: the empirical and conceptual landscape of big data techniques in financial risk management across organizational and regulatory boundaries. Research questions were then developed, rooted in an initial scoping of the literature and aligned with recognized gaps in comparative methodology, deployment, data integration, and contextual adaptation. Comprehensive search strings were constructed, anchored by keywords such as “Big Data,” “Big Data Analysis,” “Machine Learning,” “Financial Risk Management” and “Systemic Risk,” and systematically applied across major scholarly databases. Anchoring on “Big Data” was deliberate: the review’s aim is to map work that self-identifies with this term in financial risk management, a corpus we found to be thin and fragmented relative to the broader Machine Learning (ML) literature. The Scopus database served as the primary repository, employing the following search syntax to ensure retrieval of peer-reviewed, English-language journal articles published between 2016 and June 2025:

TITLE-ABS-KEY ((“Big Data” OR “Big Data Analysis”) AND (“Machine Learning”) AND (“Financial Risk Management” OR “Systemic Risk”)) AND PUBYEAR > 2015 AND PUBYEAR < 2026 AND (LIMIT-TO (LANGUAGE, “English”)) AND (LIMIT-TO (DOCTYPE, “ar”)).

To further broaden coverage and capture relevant computer science research, equivalent queries were subsequently executed in the DBLP database. All identified articles underwent a multi-stage screening process, beginning with title and abstract review for initial relevance, followed by full-text reading and methodological mapping. This allowed for the rigorous identification of empirical, theoretical, and review studies that addressed at least one dimension of the guiding research questions. Only articles meeting strict quality and topicality criteria were retained for in-depth analysis and evidence synthesis.

3.1 Inclusion criteria

The following Inclusion Criteria were established to ensure the methodological rigor and thematic relevance of the studies selected for review (Table 1).

Table 1
www.frontiersin.org

Table 1. Inclusion criteria of this SLR study.

3.2 Exclusion criteria

The exclusion criteria were applied to remove studies that did not align with the review’s temporal, linguistic, or topical scope (Table 2).

Table 2
www.frontiersin.org

Table 2. Exclusion criteria of this SLR study.

3.3 Study selection process

After removal of duplicate records, all retrieved articles were screened sequentially according to the inclusion and exclusion criteria. Initial screening was conducted on titles and abstracts to eliminate irrelevant or off-topic works. Remaining papers underwent full-text review to confirm eligibility and topical relevance. The overall flow of study selection, including the number of records identified, screened, included, and excluded at each stage, is detailed in Figure 2, which presents the PRISMA 2020 flow diagram for this review.

3.4 Data extraction and mapping

For each article retained after full-text screening, a structured data extraction protocol was implemented. Key variables extracted included authorship, publication year, sector and geographic context, type of data and big data technique (s) used, risk category addressed, methodological approach, key results, and reported limitations. This standardized mapping enabled both narrative and comparative synthesis across all studies, facilitating alignment with the review’s research questions and supporting evidence-based tabulation (see Table 3). Data extraction and coding were managed using Microsoft Excel. Where bibliometric or keyword co-occurrence analysis was relevant, VOSviewer was considered as the primary tool for network visualization and mapping. Visuals such as the PRISMA flow diagram were designed using Canva.

Table 3
www.frontiersin.org

Table 3. Studies identified through the PRISMA 2020 protocol.

3.5 Quality assessment

We applied the Mixed Methods Appraisal Tool (MMAT; 2018) as a structured checklist adapted to algorithmic, non-interventional studies. Items not applicable were marked N/A; insufficient reporting was recorded as “Cannot tell.” We report item-level patterns and did not compute overall scores.

3.6 Synthesis approach

Data from eligible studies were synthesized using a hybrid narrative-comparative approach. Each paper was mapped to one or more of the guiding research questions, enabling cross-study comparison by technique, risk category, data type, and context. Major patterns, methodological trade-offs, and sectoral or geographical distinctions were identified both narratively and in summary tables. This multi-dimensional synthesis supported an integrated analysis of dominant trends, persistent gaps, and priorities for future research.

Before synthesis, we evaluated whether a quantitative meta-analysis was feasible. Given substantial heterogeneity in tasks, metrics, datasets, and incomplete variance reporting across studies, a formal meta-analysis was not appropriate. We therefore apply a structured narrative synthesis that contrasts techniques, risk domains, data regimes, and evaluation choices. Where three or more non-overlapping studies report the same task and metric, we summarize ranges descriptively without pooling. We also flag when included studies supply governance artefacts (e.g., documentation, oversight design, monitoring plans) aligned with emerging regulatory guidance, as these shape real-world deployability.

4 Results

The systematic review process resulted in a curated evidence base of 21 primary research articles, each critically examined to reveal methodological advances, sectoral trends, and enduring limitations in the deployment of big data analytics for financial risk management. The results are presented as an integrated narrative, closely tied to the PRISMA 2020 framework, and are substantiated throughout by a series of targeted visualizations designed to clarify both the process and the evolving landscape of this multidisciplinary domain.

The foundation of this synthesis is established by the PRISMA 2020 flow diagram (Figure 3), which documents the multi-stage selection process applied to the initial pool of retrieved records. From a broad sweep of database search results, systematic application of inclusion and exclusion criteria, coupled with rigorous screening of titles, abstracts, and full texts, led to the exclusion of duplicates, non-English articles, and off-topic studies. Ultimately, this process distilled a large and heterogeneous initial corpus to a highly relevant set of empirical and conceptual studies, providing transparency and reproducibility in the construction of the final evidence base.

Figure 3
Flowchart depicting article identification via databases. From Scopus (461 records) and DBLP (11 records), 468 records remained after removing duplicates. After screening titles, 432 were excluded, leaving 36 eligible. Finally, 21 studies were included in the systematic review.

Figure 3. PRISMA flow diagram.

The temporal evolution of research activity within the field is depicted in Figure 4, which presents the distribution of included studies by year of publication. This visual underscores a dramatic escalation in scholarly output from 2021 onward, coinciding with the widespread adoption of machine learning and AI techniques in finance, the proliferation of digitized transaction data, and mounting global concern over systemic and cyber risks. The spike in recent years also reflects intensified academic and industry interest following high-profile incidents of financial disruption and regulatory transformation.

Figure 4
Line graph titled

Figure 4. Number of studies published per year.

Figure 5 maps the distribution of studies by publisher, highlighting both the diffusion of research across major international publishing houses and the growing role of open-access platforms in disseminating methodological innovation. Notably, a significant proportion of the included literature appears in interdisciplinary journals and venues at the intersection of computer science, engineering, and finance, signaling the cross-domain migration of big data methods and the field’s accelerating convergence.

Figure 5
Horizontal bar chart depicting various publishers with color-coded bars representing their respective counts. Categories include

Figure 5. Number of studies published per publisher.

A deeper layer of thematic structure emerges in Figure 6, which visualizes the keyword co-occurrence network derived from the included studies. This mapping reveals a dense and interconnected constellation of concepts, with prominent clusters centered on “machine learning,” “deep learning,” “fraud detection,” “credit risk,” “IoT,” and “blockchain.” The architecture of the network confirms that while a few analytic paradigms, particularly neural networks and ensemble learning, anchor the majority of empirical work, there is an increasing proliferation of hybrid methods and cross-cutting applications spanning domains from e-commerce to supply chain management. The visibility of terms such as “regulatory compliance,” “explainability,” and “unstructured data” at the network periphery also points to a latent research agenda around gaps and challenges yet to be systematically addressed.

Figure 6
Network graph illustrating connections between concepts related to big data, risk management, financial risk, and fintech. Nodes include terms like artificial intelligence, internet of things, neural networks, Twitter data analysis, and machine learning. Lines indicate relationships, with varying colors representing different thematic clusters.

Figure 6. Co-occurrence keyword network.

These visuals provide a panoramic snapshot of the current evidence base, documenting not only the volume and distribution of research activity, but also the evolving methodological priorities and sectoral reach of the field. With this context established, the results are presented thematically according to the four guiding research questions, each supported by integrated narrative, summary tables, and direct cross-referencing to the mapped literature. This approach facilitates a transparent and comprehensive synthesis, aligning empirical patterns and research frontiers with the review’s overarching objectives.

Because outcome definitions and metrics are not commensurate across the corpus, we do not pool effects; instead we report structured comparisons and, when directly comparable, descriptive ranges.

To ground the subsequent thematic synthesis in methodological transparency and allow for precise cross-referencing, Table 3 presents a detailed mapping of the 21 studies included in this review. Each entry documents essential elements, such as authorship, publication year, analytical methodology, risk domain, sectoral or organizational context, and principal findings, enabling both a high-level overview and granular comparison across the evidence base. This tabular synthesis serves as both an analytical scaffold for the narrative discussion and a navigational aid for readers seeking to trace the evolution of big data approaches within specific financial risk categories, institutional settings, or methodological families. By foregrounding this structured evidence map, the review underscores its commitment to systematic rigor and supports transparent engagement with the breadth and diversity of recent scholarship at the intersection of big data and financial risk management.

To situate the evidence map in terms of study quality and reporting, we summarize an item-level appraisal using the Mixed Methods Appraisal Tool (MMAT; Version 2018 user guide) (Hong et al., 2018), adapted to algorithmic, non-interventional studies. Of the 21 items, 18 are empirical (quantitative non-randomized n = 15; quantitative descriptive n = 3) and 3 are conceptual/system pieces recorded as S2 = No and not domain-scored. Patterns are straightforward: measurement definitions and metrics are generally adequate (3.2 = Yes), while sampling frames, outcome completeness, and confounder handling are often under-reported (3.1/3.3/3.4 = mostly CT). Intervention adherence is inapplicable in this corpus (3.5 = N/A). We report distributions rather than any composite score, and use these ticks only to weight interpretation and claims of generalizability. Table 4 presents the MMAT summary (N/A = not applicable; CT = “cannot tell”).

Table 4
www.frontiersin.org

Table 4. MMAT appraisal summary for the included studies.

Taken together, this appraisal gives us guardrails for interpretation rather than a scoreboard. Where studies report transparent data pipelines, suitable metrics, and some form of external validation, we treat their findings as more persuasive; where sampling frames, outcome completeness, or confounder handling are unclear, we read results as promising but provisional. These signals shape how we compare techniques and deployment contexts in the next section and help us flag what may or may not travel across jurisdictions, data regimes, and risk types. The Discussion uses these cues to weigh evidence, draw practical takeaways for financial risk management, and point to reporting practices that would materially strengthen the field (Table 4).

5 Discussion

The growing body of research on big data analytics in financial risk management is marked by methodological variety, contextual specificity, and uneven practical maturity. Given the heterogeneity and limited variance reporting in the primary studies, treating results as a quantitative meta-estimate would be misleading; our comparisons are therefore narrative and, where possible, anchored by descriptive ranges rather than pooled effects. While significant advances have been made in algorithm development, data integration, and early empirical applications, the field remains defined by persistent fragmentation and unresolved questions regarding comparative effectiveness, real-world impact, and transferability across domains. To provide a rigorous, evidence-based response to these challenges, the following sections synthesize findings from twenty-one recent studies, mapping both the state-of-the-art and the enduring limitations in the literature. The analysis is structured around four guiding research questions, each targeting a critical dimension of the current landscape: the comparative performance of big data techniques, their real-world deployment and scalability, the integration of non-traditional data sources, and the influence of regulatory and sectoral context. By addressing these questions systematically, the review aims to distill both actionable insights and priority areas for future scholarly and practical innovation.

Several use cases in our corpus (credit scoring/underwriting, fraud controls, some insurance risk models) fall within the EU AI Act’s high-risk scope, which triggers specific obligations on risk management, data governance, technical documentation/logging, transparency and human oversight, accuracy/robustness, conformity assessment and post-market monitoring. The Act entered into force on 1 Aug 2024 with phased application (selected prohibitions from Feb 2025; GPAI/governance rules from Aug 2025; most high-risk obligations fully applicable by Aug 2026–2027). For financial actors, this means treating model governance as a compliance-critical capability, not a research afterthought. In parallel, the OECD (Organization for Economic Co-operation and Development) AI Principles (human-centred values, transparency/explainability, robustness/safety, accountability) provide a pragmatic lens we can operationalize in FRM (Financial Risk Management) by (i) documenting provenance and feature lineage, (ii) reporting model cards and decision-support affordances for reviewers, and (iii) tracking fairness/impact metrics alongside utility. We reflect these requirements by reading empirical results as promising where governance artefacts exist, and as provisional where sampling, completeness or oversight are under-reported.

As big data pipelines increasingly ingest unstructured, behavioral, and real-time signals, vulnerability operates as a first-class property of the data itself, not just an external threat to be patched later. Data lineage breaks, adversarial contamination, and silent drift can alter distributions and semantics long before models see the inputs, which means that veracity, validity, and value are all conditional on how exposure to failure and attack is managed across the lifecycle. Treating vulnerability as intrinsic aligns the 8 Vs with contemporary practice, where provenance, resilience, and recoverability must be designed in from collection to consumption.

RQ1: What are the comparative strengths, limitations, and practical trade-offs of different big data-driven analytical techniques (such as neural networks, ensemble machine learning, fuzzy logic, and information fusion) in managing various categories of financial risk (credit, fraud, systemic, and operational) across sectors?

The landscape of big data-driven analytical techniques in financial risk management has evolved into a complex mosaic, reflecting the confluence of rapid technological progress and growing risk complexity across sectors. Recent years have witnessed an escalating shift from traditional statistical and rule-based models toward increasingly sophisticated paradigms, including advanced neural networks, ensemble learning, hybrid optimization, fuzzy multi-criteria systems, and network-based information fusion. Each of these paradigms responds to distinct practical pressures: the need for heightened predictive power in credit scoring and fraud detection, the imperative of robustness under uncertainty in market and investment risk, and the growing demand for systemic risk monitoring amid networked financial environments and contagion threats. Despite notable methodological progress, this literature remains shaped by sectoral silos, data accessibility, and the persistent tension between algorithmic accuracy, interpretability, and operational scalability. Benchmarking studies rarely cross boundaries between families of techniques or rigorously test their performance in diverse, real-world deployment scenarios. Instead, evidence is often context-bound; banking, fintech, or supply chain, relying on proprietary, simulated, or highly localized datasets. These conditions have produced a field where technical sophistication outpaces its integration into decision processes, and where the practical strengths and limitations of competing approaches remain underexplored at scale.

Neural network-based approaches, particularly those leveraging hybrid or optimized architectures, consistently demonstrate high predictive accuracy for credit risk, bankruptcy forecasting, and enterprise financial evaluation (Zhou et al., 2019; Liu, 2022; Tayfor et al., 2024; Cui and Yang, 2024; Bi and Liang, 2022; Yue et al., 2021). Particle Swarm-optimized Back Propagation (PSO-BP) neural networks, for instance, outperform conventional ML in processing IoT-driven loan data, with notable gains in both convergence and screening efficiency (Zhou et al., 2019). Deep fuzzy neural networks, particularly when enhanced by advanced optimization techniques, routinely achieve classification accuracies exceeding 95% on benchmark bankruptcy datasets, but such results are largely confined to simulation environments or proprietary data (Tayfor et al., 2024; Cui and Yang, 2024).

Ensemble machine learning models, notably Random Forest, XGBoost, and selective ensemble frameworks, dominate fraud detection, credit scoring, and operational risk applications, consistently outperforming classical single-model approaches (Xia et al., 2024; Nivetha et al., 2024; Murugan, 2023; Bi and Liang, 2022). Selective ensemble models (Xia et al., 2024) and XGBoost (Murugan, 2023) are shown to deliver superior robustness in highly imbalanced digital financial datasets, offering both high recall and computational scalability. However, these approaches often demand significant data preprocessing, feature engineering, and substantial computational resources, limiting ease of adoption in resource-constrained environments (Xia et al., 2024; Murugan, 2023).

Fuzzy logic and multi-criteria decision-making (MCDM) techniques find their greatest traction in contexts of investment, market, and liquidity risk, where uncertainty and conflicting stakeholder priorities are paramount (Shang et al., 2025; Zhou, 2023; Srinivasan and Kamalakannan, 2017). Interval-valued picture fuzzy sets and advanced aggregation operators, when combined with large-scale expert input, enhance robustness in volatile environments (Shang et al., 2025), but their scalability to fully automated, data-driven regimes remains limited.

Network-based and information fusion models, such as categorical graphical models and Bayesian data fusion, offer unique capabilities for systemic risk identification, especially by integrating non-traditional data sources like social media and interbank exposures (Cerchiello and Giudici, 2016a, 2016b). These approaches unlock new dimensions of market sentiment analysis and early warning, but their performance is sensitive to the quality and reliability of unstructured input, and practical deployment is often restricted to specific regulatory or market settings.

Lastly, Hybrid and optimization-driven frameworks, including genetic tabu algorithms, swarm optimization, and multi-objective evolutionary models, excel at early warning and multi-criteria decision tasks, adapting flexibly to complex organizational contexts (Liu, 2022; Deka et al., 2024; Srinivasan and Kamalakannan, 2017). Yet, the challenge of model transparency and the interpretability of results remains a significant barrier for managerial adoption (Table 5).

Table 5
www.frontiersin.org

Table 5. Summary of techniques used in financial risk management.

Taken together, these studies illustrate that no single big data technique universally outperforms others across all risk categories or sectors. Instead, the comparative strengths and limitations of each approach are shaped by domain requirements, data structures, and operational constraints. Neural networks and ensemble ML models dominate in predictive accuracy for well-structured risk tasks, while fuzzy MCDM and hybrid models are more robust under conditions of uncertainty and multi-stakeholder decision-making. Network and information fusion approaches are indispensable for systemic and contagion risk, particularly when alternative data streams, such as social media or network exposures, must be integrated. Hybrid and optimization-driven models provide flexibility and adaptability, excelling in complex, dynamic environments where early warning and multi-criteria evaluation are essential. Across all families of methods, however, the persistent trade-off between performance, interpretability, and practical scalability remains unresolved, underscoring the need for continued benchmarking, greater emphasis on transparent decision support, and more robust evidence from real-world deployments (Zhou et al., 2019; Xia et al., 2024; Shang et al., 2025; Cerchiello and Giudici, 2016a, 2016b; Liu, 2022; Tayfor et al., 2024; Nivetha et al., 2024; Murugan, 2023).

RQ2: How do real-world applications of big data and AI models in financial risk management perform when deployed at scale, and what challenges or gaps remain regarding generalizability, data diversity, and integration with organizational processes?

Despite the surge of algorithmic development in financial risk management, the real-world deployment and organizational integration of big data and AI models remain uneven and highly context-dependent. The literature reveals a sharp divergence between technical promise and practical impact: while a subset of studies demonstrate the feasibility of large-scale implementation and impressive predictive performance in operational settings, many contributions remain tethered to simulation environments, single-organization case studies, or limited-scope datasets. This landscape is characterized by a persistent tension between the scalability of advanced models, the diversity and heterogeneity of financial data streams, and the realities of integrating such systems into existing business processes and institutional routines.

Empirical evidence of successful, scalable deployment is strongest in studies leveraging substantial real-world datasets or operating within digitally mature financial environments. For instance, Zhou et al. (2019) showcase a Spark-based PSO-BP neural network deployed for IoT-enabled loan risk screening in a Chinese commercial bank, demonstrating both computational efficiency and domain-specific predictive value. Similarly, Murugan (2023) report on the application of ensemble ML algorithms, including KNN, logistic regression, and XGBoost, within large-scale banking and IoT contexts, highlighting the capacity of these systems to outperform conventional models when given access to extensive, heterogeneous transaction records. Nivetha et al. (2024) extend this evidence to the domain of fraud detection, deploying Random Forest and XGBoost on an open-access Kaggle dataset comprising nearly 300,000 credit card transactions and achieving high predictive accuracy in a high-velocity, real-world data environment.

Yet, the leap from technical validation to routine field deployment remains substantial. Multiple studies underscore significant challenges related to data quality, model transferability, and the practical integration of big data analytics with legacy IT infrastructure and risk governance processes (Li, 2018; Bi and Liang, 2022; Yue et al., 2021; Xia et al., 2024). For example, while Li (2018) develops a real-time risk control platform for multi-institutional Chinese finance, the system’s effectiveness is deeply contingent on regulatory support and organizational buy-in, variables that are rarely controlled or measured in empirical work. The majority of AI/ML deployment case studies, such as Bi and Liang’s (2022) risk assessment in IoT-driven e-commerce, remain bounded by single-firm, proof-of-concept contexts, limiting the generalizability of their results. Meanwhile, Liu (2022) and Deka et al. (2024) describe hybrid or optimization-driven models for early warning and supply chain risk, yet their field-testing is restricted to municipal enterprises or select industrial partners.

A common thread in these studies is the necessity of robust IT and data infrastructure for successful scale-up. Where banking and fintech institutions possess established data pipelines, model deployment and iterative learning are feasible (Murugan, 2023; Zhou et al., 2019). However, in less digitally mature sectors or geographies, integration with organizational routines, regulatory compliance, and workforce upskilling remain formidable barriers. Wang and Wang (2022) and Yue et al. (2021) highlight the importance of managerial engagement and regulatory alignment, yet both ultimately acknowledge that current evidence is patchy and context-specific, with successful implementation stories often failing to translate beyond their original institutional setting (Table 6).

Table 6
www.frontiersin.org

Table 6. Real-world application and deployment of big data and AI models in financial risk management.

These findings, put together, reveal a field where large-scale, real-world application of big data and AI for financial risk management is not only possible, but demonstrably valuable, when the right technical and organizational preconditions are met. However, substantial challenges persist in achieving broad generalizability, particularly regarding the transfer of models across heterogeneous organizations, geographies, and risk types. The persistent reliance on isolated case studies and benchmark datasets underscores the need for more systematic research on integration, scalability, and the organizational dynamics that enable (or inhibit) the practical impact of big data analytics in finance (Zhou et al., 2019; Li, 2018; Murugan, 2023; Bi and Liang, 2022; Xia et al., 2024; Yue et al., 2021; Nivetha et al., 2024; Deka et al., 2024; Liu, 2022; Wang and Wang, 2022).

RQ3: To what extent does the integration of non-traditional and unstructured data sources—such as IoT signals, social media, and behavioral analytics—enhance the predictive accuracy and early warning capabilities of financial risk models, and what barriers persist in achieving widespread adoption?

The integration of non-traditional and unstructured data sources, such as IoT device signals, social media streams, and e-commerce behavioral data, has been widely heralded as the next frontier for financial risk modeling. Yet, a systematic examination of recent studies reveals that while enthusiasm is high, substantive operationalization remains the exception rather than the rule. Across the reviewed literature, only a handful of empirical works move beyond theoretical endorsement to actually implement or rigorously test the predictive value of heterogeneous data integration within financial risk management systems. Where realized, these approaches demonstrate clear, quantifiable improvements in both the timeliness and accuracy of risk prediction, particularly in domains where traditional financial indicators alone have proven insufficient.

Notably, Zhou et al. (2019) provide a compelling illustration of IoT data fusion in credit risk screening, leveraging a PSO-BP neural network architecture to analyze chattel mortgage loans in a Chinese commercial bank. Their model integrates sensor-derived asset and transaction data, resulting in significant gains in default prediction efficiency relative to conventional machine learning pipelines. A similar practical step forward is found in the work of Cerchiello and Giudici (2016a, 2016b), whose network-based models combine market financial information with semantic analysis of Twitter data to model systemic risk and contagion effects among Italian banks. By merging structured market metrics with real-time sentiment streams, their approach not only enhances early warning capabilities but also opens new avenues for capturing emerging, exogenous threats often overlooked by balance sheet analysis alone.

IoT-driven advances are also evident in Murugan (2023), where ensemble machine learning techniques, incorporating KNN, logistic regression, and XGBoost, are deployed on large, sensor-rich datasets in banking and enterprise contexts. Here, the incorporation of IoT telemetry is shown to boost real-time monitoring and model responsiveness, especially in environments characterized by rapid data velocity and volume. Bi and Liang (2022) reinforce these findings at the firm level, demonstrating that machine learning models which incorporate IoT operator data and e-commerce behavioral signals outperform those trained exclusively on standard financial attributes, particularly in predicting credit and operational risks (Table 7).

Table 7
www.frontiersin.org

Table 7. Integration of non-traditional and unstructured data in financial risk management.

While these studies highlight the transformative potential of unstructured and non-traditional data, the broader literature still largely defaults to classic financial variables or treats new data types as peripheral supplements. Bibliometric and conceptual syntheses, such as those by Singh et al. (2024) and Cong (2021), trace the emergence of big data, blockchain, IoT, and behavioral analytics as rising research trends but simultaneously note the field’s lag in operationalizing truly multimodal risk frameworks. In most cases, technical, organizational, and data governance barriers, including the lack of standardized data integration pipelines, data privacy concerns, the complexity of real-time unstructured data processing, and the absence of regulatory clarity, are cited as major obstacles to widespread adoption and scalable impact (Singh et al., 2024; Cong, 2021).

Thus, while select pioneering studies provide tangible evidence that the fusion of non-traditional and unstructured data sources can materially improve risk detection and early warning performance, the financial risk management literature as a whole remains in the early stages of this transition. Systematic, sector-wide adoption, and robust, cross-organizational validation, remains a frontier rather than a settled reality. The next wave of research and practice must focus on overcoming integration challenges, developing interpretable and regulatory-compliant architectures, and providing empirical benchmarks that move beyond isolated pilots (Zhou et al., 2019; Cerchiello and Giudici, 2016a, 2016b; Murugan, 2023; Bi and Liang, 2022; Singh et al., 2024; Cong, 2021).

RQ4: What regulatory, geographical, or sectoral differences shape the adoption, effectiveness, and governance of big data techniques for financial risk management, and where do current studies fail to provide comparative or global perspectives?

The diffusion and impact of big data techniques in financial risk management are inextricably linked to regulatory environments, national contexts, and sector-specific demands. Across the reviewed literature, a pronounced geographical skew emerges: the majority of empirical and applied research is anchored in the Chinese financial system or in highly localized case studies, with only sporadic forays into European (notably Italian) banking or global bibliometric mapping. As a result, the capacity to draw robust, cross-jurisdictional lessons or articulate sector-agnostic best practices remains fundamentally constrained.

Studies such as Zhou et al. (2019), Li (2018), and Wang and Wang (2022) ground their technical advances in the regulatory and institutional landscapes of China’s banking and internet finance sectors. Here, strong state-led digital infrastructure and proactive regulatory frameworks enable the rapid adoption of IoT-enabled risk models, real-time monitoring systems, and AI-driven fraud detection at scale. These deployments often benefit from centralized oversight, unified data standards, and comparatively streamlined approval processes, all of which are less prevalent in more fragmented or heterogeneous regulatory regimes.

By contrast, European contributions, most notably Cerchiello and Giudici (2016a, 2016b), focus on the integration of alternative data sources and network-based systemic risk models in the Italian banking context. While these studies are technically innovative, their practical deployment is shaped by the realities of EU data privacy regulations (such as GDPR), the complexity of cross-border banking supervision, and a generally more cautious, compliance-driven approach to big data adoption. This regulatory backdrop both constrains and refines the nature of data fusion and analytics used, foregrounding concerns around explainability, auditability, and consumer protection.

Sectoral divergence is also pronounced. In addition to the banking and fintech sectors, a smaller but growing body of work addresses e-commerce (Cong, 2021; Bi and Liang, 2022), supply chain finance (Deka et al., 2024), and state-owned enterprises (Zhou, 2023). These domains face unique data challenges and regulatory considerations; from the handling of transactional and behavioral data in digital commerce to the public accountability and audit requirements of government-linked entities. Notably, the adoption trajectory in these sectors is heavily influenced by sector-specific compliance standards, legacy IT integration hurdles, and differing levels of risk appetite and innovation readiness.

Bibliometric and conceptual analyses (Singh et al., 2024; Chakraborty, 2019) further illuminate the global landscape, mapping regional hotspots of big data innovation and regulatory engagement. These studies reveal that while the rhetoric of digital transformation and AI-driven risk governance is nearly universal, substantive differences persist in legal environments, data access, and institutional support; often reinforcing rather than bridging regional and sectoral divides (Table 8).

Table 8
www.frontiersin.org

Table 8. Geographical, sectoral, and regulatory contexts in big data financial risk management studies.

This current evidence base reveals a field dominated by context-specific solutions and regulatory path dependency. The scarcity of direct comparative studies, either across countries or sectors, underscores a critical blind spot. Few studies attempt to systematically analyze how variations in regulatory frameworks, market structures, or organizational cultures mediate the adoption and impact of big data analytics in financial risk management. This gap not only limits the transferability of current findings but also highlights an urgent research agenda for future cross-jurisdictional and cross-sectoral inquiry (Zhou et al., 2019; Cerchiello and Giudici, 2016a, 2016b; Li, 2018; Wang and Wang, 2022; Zhou, 2023; Cong, 2021; Bi and Liang, 2022; Deka et al., 2024; Singh et al., 2024; Chakraborty, 2019).

Accordingly, our synthesis should be read as context-bound to Chinese and selected European banking/fintech regimes; transfer to under-represented jurisdictions and sectors may be limited unless models are externally validated and re-tuned to different legal, market, and organizational conditions.

6 Limitations of the study

Despite its systematic rigor and adherence to PRISMA 2020 standards, this review is subject to several inherent limitations; some reflective of broader structural gaps in the field, others a function of methodological choices and practical constraints.

First, the review’s evidence base, while diverse in methodology and sectoral reach, remains constrained by the biases and silos present in the existing literature. The overwhelming concentration of empirical studies in Chinese and select European contexts, particularly banking and fintech, limits the transferability of findings to underrepresented regions, industries, and regulatory regimes. Cross-country, cross-sector, and comparative regulatory analyses are virtually absent from the retrieved studies, precluding robust conclusions about the global generalizability or context-dependence of big data approaches for financial risk management. To mitigate this in future updates and replications, we will (i) extend retrieval to regional/non-English sources where feasible, (ii) prioritize studies that report cross-jurisdiction external validation and clearly tag jurisdictional metadata, and (iii) encourage collaborative, multi-site designs that harmonize labels and report ΔAUC/ΔF1 across countries and sectors.

Second, the rapid evolution of both big data technologies and financial risk environments means that some included studies may already be technologically outdated, especially those employing earlier machine learning architectures or proprietary datasets no longer representative of industry best practices. While the review is temporally bounded (2016–2025) to capture the field’s recent acceleration, it remains vulnerable to the inherent lag between research, peer review, and publication.

Third, the reliance on English-language, peer-reviewed journal articles indexed in Scopus and DBLP introduces inevitable publication and language bias. High-quality research published in other languages or disseminated through alternative scholarly channels (e.g., conference proceedings, industry reports, white papers) may have been excluded, potentially narrowing the analytical lens. In recognition of possible terminology bias, we tested a minimal refresh (retaining the Big Data anchor and adding “Machine Learning” and “Systemic Risk”); it produced no additional eligible papers, though some big-data-scale studies that do not self-identify with the term may remain outside our net.

Fourth, we applied the Mixed Methods Appraisal Tool (MMAT; 2018) as a structured checklist adapted to algorithmic, non-interventional studies, reporting item-level patterns and not computing composite scores. Given the heterogeneity of tasks, outcomes, and reporting, domain-specific risk-of-bias tools (e.g., PROBAST for prediction model studies, ROBINS-I for causal non-randomized designs, AMSTAR 2 for reviews) were not uniformly applicable across our corpus. Future updates may layer in PROBAST where studies explicitly develop/validate prediction models with sufficient reporting while retaining MMAT for comparability.

Finally, and perhaps most fundamentally, this review is limited by the persistent blind spots in the literature itself. Despite repeated calls for attention to explainability, interpretability, managerial usability, and practical adoption, few empirical studies operationalize these dimensions in a manner that supports systematic synthesis or actionable guidance. The absence of robust, real-world deployment studies, especially outside major banking and fintech hubs, constrains the review’s ability to draw definitive conclusions about organizational integration and impact at scale. For similar reasons, emergent topics such as the integration of unstructured data, regulatory harmonization, and the interface of AI ethics and financial governance remain identified as critical gaps rather than domains of resolved knowledge (Figure 7).

Figure 7
Infographic outlining research gaps with icons: Geographical and Sectoral Bias, showing limited generalizability from China/Europe in banking/fintech. Technological & Temporal Gaps note outdated studies due to rapid field changes. Publication & Language Bias highlight reliance on English peer-reviewed articles. Quality Appraisal mentions using MMAT (2018) with suggestions for additional tools like PROBAST. Enduring Research Gaps emphasize lack of focus on explainability, real-world adoption, and regulatory or ethical issues.

Figure 7. Limitations of this study.

7 Future research

The synthesis of current literature on big data and financial risk management reveals a rapidly advancing but unevenly distributed field, characterized by both methodological innovation and persistent blind spots. The limitations of the present review, and more importantly, the structural gaps in the evidence base, point to a series of urgent and actionable priorities for future research.

First, there is a pressing need for comparative, cross-jurisdictional, and cross-sectoral studies. Within this agenda, public finance, insurance, and non-bank financial institutions in emerging economies should be treated as priority contexts rather than peripheral cases, so that transferability is tested where it matters most. The current landscape is dominated by case studies in Chinese and select European banking and fintech contexts, with scant attention to emerging markets, public sector finance, or non-bank financial institutions. Future research should prioritize multi-country collaborations and comparative designs that explicitly examine how regulatory regimes, market structures, and organizational cultures mediate the adoption, effectiveness, and governance of big data analytics in risk management. Such studies will be crucial in moving the field beyond isolated exemplars toward transferable, context-sensitive best practices. Practically, we recommend pre-registered, multi-country protocols with shared open splits and jurisdiction tags, and an out-of-domain evaluation plan that reports ΔAUC, ΔF1, and Precision@k per country and sector. Subsequent updates to this review will broaden retrieval with sector-specific terms (e.g., tax compliance, procurement fraud, benefits-payment integrity; claims triage, underwriting leakage; microfinance, digital lending, Non-Bank Financial Institutions – NBFIs) and include policy/administration and insurance venues, as feasible, to reduce coverage gaps in emerging-market and public-sector settings.

Second, advancing the practical deployment and real-world validation of big data models remains a critical research frontier. While technical performance has improved, few studies address the full life cycle of model integration, from data infrastructure and workforce capability to end-user adoption and impact on decision quality. Experimental and longitudinal field studies, action research, and mixed-methods evaluations are needed to surface the organizational, managerial, and sociotechnical challenges that shape scalable and sustainable implementation. Special emphasis should be placed on underexplored domains such as insurance, public finance, and digital lending in developing economies.

Third, the field must move beyond technical metrics to systematically address explainability, interpretability, and ethical governance. The overwhelming focus on accuracy and predictive power risks sidelining issues of transparency, accountability, and compliance, especially as financial systems become more algorithmically mediated. Future research should develop and empirically test models and frameworks for explainable AI (XAI), risk communication, and human-in-the-loop decision support, while also probing the interplay between regulatory innovation (e.g., AI audits, algorithmic accountability) and technological change (Kumar et al., 2024). Concretely, we recommend regulatory-grade evaluations that pair performance with governance outcomes: time-to-conformity (documentation & logging readiness), human-oversight effectiveness (override calibration/Brier score), and post-market monitoring signals (incident rate, detection latency, drift alarms) under the EU AI Act’s high-risk regime; use standardized profiles (e.g., NIST AI RMF GOVERN/MAP/MEASURE/MANAGE) to make results comparable across deployments.

Fourth, operationalize cyber and data vulnerability as a design variable. Future work should specify, measure, and stress-test data exposure across the pipeline, including lineage integrity, contamination pathways, and drift under adversarial or failure conditions. This calls for benchmarks and reporting standards that couple predictive performance with resilience metrics, for example recovery time, detection latency, and robustness to distributional shift, so that vulnerability is managed alongside accuracy and explainability in high-stakes financial settings.

Fifth, the integration and operationalization of unstructured and alternative data sources demand deeper methodological and practical attention. Studies incorporating IoT, social media, and behavioral analytics remain rare and often proof-of-concept in nature. There is a need for scalable architectures and standardized protocols for multi-modal data fusion, as well as critical assessment of the value, reliability, and ethical risks posed by alternative data streams in high-stakes financial contexts (Adnan and Akbar, 2019).

Finally, future work should invest in open science practices and transparent reporting standards. The creation and dissemination of reproducible code, open-access datasets, and standardized evaluation benchmarks will accelerate field-wide learning and comparative analysis, helping to close current gaps in coverage, quality, and global relevance (Nurunnabi, 2021) (Table 9).

Table 9
www.frontiersin.org

Table 9. Future research directions in big data and financial risk management.

Building on these gaps, the following focused questions arise directly from the 21-paper corpus.

a. External validity across jurisdictions. When a credit-default or bankruptcy model is trained in one jurisdiction and evaluated in another with similar features, what is the out-of-domain performance drop (e.g., ΔAUC, ΔF1), and which feature families travel? Design: pre-register splits by country or regulatory bloc; report both relative and absolute deltas. This follows the geographic concentration we observed and the lack of cross-site testing.

b. Integrity and resilience of fraud models. Under controlled label noise or data poisoning at the ingestion stage, how quickly do monitoring systems detect degradation and how fast do models recover after rollback or retrain? Outcomes: detection latency, recovery time, and ΔAUC at fixed false-positive cost. This operationalizes “data vulnerability” we flagged.

c. Multimodal early-warning for systemic risk. Do graph + text features (exposures, news, filings) improve early-warning precision at top-k events versus tabular baselines? Outcomes: Precision@k, lead-time in days. Many papers use single-modality tabular data; few fuse unstructured signals.

d. Explainability to decision quality. Do SHAP or counterfactual explanations improve human override quality and calibration in credit or fraud review teams compared with score-only dashboards? Outcomes: Brier score, net benefit, time-to-decision in a randomized user study. This moves explainability from principle to measured effect.

e. Cost-aware thresholds under class imbalance. Given realistic base rates and asymmetric loss, which thresholding schemes (cost curves, expected utility, conformal risk control) maximize net benefit across datasets commonly used in the literature? Report decision curves rather than accuracy alone; many included studies optimize symmetric metrics.

f. Benchmarking fuzzy/NN/ensemble methods head-to-head. On the same open splits for the same task, which family wins where and why (e.g., IVPF-MCDM vs. deep ensembles vs. gradient boosting) when judged on accuracy and stability across time? Outcome: average rank and variance across rolling windows. Our synthesis found few true, shared-data comparisons.

g. Provenance and drift. Does adding immutable data lineage and feature versioning reduce drift incidents and false positives in production risk systems over a six-month window? Outcomes: incident rate ratio and mean time between incidents; ties back to pipeline vulnerability noted in multiple studies.

h. Reporting completeness and reproducibility. Is MMAT item coverage (sampling frame, outcome completeness, confounders) associated with out-of-sample performance stability across replicates? Design: correlate item-level reporting with ΔAUC across re-runs on shared splits. This leverages the appraisal patterns we documented.

Each question is tied to at least one gap we observed in the 21-paper set: limited external validation and comparability, sparse multimodal use, minimal robustness testing, and under-reporting of sampling/completeness.

8 Conclusion

The accelerating convergence of big data analytics and financial risk management is not merely a technological evolution; it represents a fundamental reordering of how uncertainty is apprehended and managed across financial sectors. This review, by systematically mapping and critically synthesizing 21 recent studies, exposes both the formidable advances achieved and the persistent, often underestimated limitations that continue to shape the field. It is clear that neural networks, ensemble learning, fuzzy logic, and hybrid optimization have each claimed territory in the analytics arsenal, driving measurable gains in prediction, early warning, and adaptive decision-making. Yet, these gains remain unevenly distributed, hemmed in by sectoral silos, data fragmentation, and organizational inertia.

Perhaps the most consequential insight is the field’s methodological pluralism, coupled with its persistent fragmentation. The practical trade-offs between predictive power, transparency, and deployment scalability are neither trivial nor resolved. While machine learning and AI have demonstrably raised the ceiling for what is technically possible, the persistent gap between technical validation and real-world adoption cannot be bridged by algorithms alone. The integration of non-traditional data, meaning IoT signals, behavioral analytics, sentiment streams, remains at an early, often experimental stage. Most empirical work is still bound to a narrow set of geographies, institutional logics, and regulatory frameworks, reinforcing rather than resolving questions about generalizability and impact.

This review also signals a pivotal moment for the discipline: methodological innovation alone is no longer enough. The appetite for explainability, managerial usability, and regulatory accountability is mounting, but the evidence base to support these imperatives is only just beginning to take shape. The path forward is clear; future research must confront head-on the challenges of cross-jurisdictional transferability, integration of heterogeneous data, and the operationalization of ethical, explainable analytics. Open science practices, multi-country studies, and stronger field validation are not optional luxuries, but prerequisites for the next era of financial risk management. Our closing questions in Section 7 translate the field’s generic calls into specific evaluations that can be run and compared, so progress is measurable rather than rhetorical.

In the end, big data analytics hold the promise of transforming not just the technical infrastructure of risk analysis, but the very culture of financial decision-making. To realize this promise, the field must move decisively beyond insular benchmarks and isolated pilots toward a more transparent, comparative, and practice-attuned science. Only then will big data fulfill its role as both a catalyst and a safeguard in the evolving landscape of global finance.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

LT: Writing – original draft, Investigation, Resources, Validation, Conceptualization, Project administration, Supervision, Methodology, Writing – review & editing. AT: Methodology, Writing – original draft, Visualization, Formal analysis, Writing – review & editing, Conceptualization. AB: Conceptualization, Writing – review & editing, Writing – original draft, Project administration, Methodology, Visualization.

Funding

The author(s) declare that no financial support was received for the research and/or publication of this article.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The authors declare that no Gen AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Addy, W. A., Ajayi-Nifise, A. O., Bello, B. G., Tula, S. T., Odeyemi, O., and Falaiye, T. (2024). Machine learning in financial markets: a critical review of algorithmic trading and risk management. Int. J. Sci. Res. Arch. 11, 1853–1862. doi: 10.30574/ijsra.2024.11.1.0292

Crossref Full Text | Google Scholar

Adnan, K., and Akbar, R. (2019). Limitations of information extraction methods and techniques for heterogeneous unstructured big data. Int. J. Eng. Bus. Manage. 11:1847979019890771. doi: 10.1177/1847979019890771

Crossref Full Text | Google Scholar

Bi, W., and Liang, Y. (2022). Risk assessment of operator’s big data Internet of Things credit financial management based on machine learning. Mob. Inf. Syst. 2022, 1–11. doi: 10.1155/2022/5346995

PubMed Abstract | Crossref Full Text | Google Scholar

Cerchiello, P., and Giudici, P. (2016a). Big data analysis for financial risk management. J. Big Data 3, 1–12. doi: 10.1186/s40537-016-0053-4

PubMed Abstract | Crossref Full Text | Google Scholar

Cerchiello, P., and Giudici, P. (2016b). Categorical network models for systemic risk measurement. Qual. Quant. 51, 1593–1609. doi: 10.1007/s11135-016-0354-x

Crossref Full Text | Google Scholar

Chakraborty, G. (2019). Evolving profiles of financial risk management in the era of digitization: the tomorrow that began in the past. J. Public Aff. 20:e2034. doi: 10.1002/pa.2034

Crossref Full Text | Google Scholar

Cong, X. (2021). “Research on financial risk management of E-commerce enterprises in the era of big data,” in Proceedings of the 7th International Conference on Frontiers of Educational Technologies. Bangkok, Thailand. (195–199). doi: 10.14733/cadaps.2024.S21.208-225

Crossref Full Text | Google Scholar

Cui, Z., and Yang, N. (2024). Optimization algorithm for enterprise decision making based on big data fusion.

Google Scholar

Deka, A., Das, P. J., and Saikia, M. J. (2024). Advanced supply chain management using adaptive serial cascaded autoencoder with LSTM and multi-layered perceptron framework. Logistics 8:102. doi: 10.3390/logistics8040102

Crossref Full Text | Google Scholar

Ezzahra, M. F., Nadia, A. F. I. F. I., and Imane, H. I. L. A. L. (2019). Big data dependability opportunities & challenges. In 2019 1st International Conference on Smart Systems and Data Science (ICSSD). San Jose, CA: IEEE, 1–4.

Google Scholar

González García, C., and Álvarez-Fernández, E. (2022). What is (not) Big Data based on its 7Vs challenges: a survey. Big Data Cogn. Comput. 6:158. doi: 10.3390/bdcc6040158

Crossref Full Text | Google Scholar

Hilbert, M., and Darmon, D. (2020). How complexity and uncertainty grew with algorithmic trading. Entropy 22:499. doi: 10.3390/e22050499

PubMed Abstract | Crossref Full Text | Google Scholar

Hong, Q. N., Fàbregues, S., Bartlett, G., Boardman, F., Cargo, M., Dagenais, P., et al. (2018). The mixed methods appraisal tool (MMAT) version 2018 for information professionals and researchers. Educ. Inf. 34, 285–291. doi: 10.3233/EFI-180221

Crossref Full Text | Google Scholar

Huynh, P. L. T., Le, P. L., and Nguyen, T. T. (2025). Sustainability of FDI and domestic firms through big data analytics expertise: the moderating role of supply chain risk management. Benchmarking.. doi: 10.1108/BIJ-09-2024-0735

Crossref Full Text | Google Scholar

Kalashnikov, R., and Kartbayev, A. (2024). Assessment of the impact of big data analysis on decision-making in stock trading processes. Procedia Comput. Sci. 231, 786–791. doi: 10.1016/j.procs.2023.12.137

Crossref Full Text | Google Scholar

Kumar, S., Datta, S., Singh, V., Datta, D., Singh, S. K., and Sharma, R. (2024). Applications, challenges, and future directions of human-in-the-loop learning. IEEE Access 12, 75735–75760. doi: 10.1109/ACCESS.2024.3401547

Crossref Full Text | Google Scholar

Lentini, M. R., and Thayasivam, U. (2025). Value-at-risk student prescription trees for price personalization. J. Big Data 12, 1–12. doi: 10.1186/s40537-024-01036-y

PubMed Abstract | Crossref Full Text | Google Scholar

Li, Y. (2018). Design a management information system for financial risk control. Clust. Comput. 22, 8783–8791. doi: 10.1007/s10586-018-1969-6

PubMed Abstract | Crossref Full Text | Google Scholar

Liu, H. (2022). Financial risk intelligent early warning system of a municipal company based on genetic tabu algorithm and big data analysis. Int. J. Inf. Technol. Syst. Approach 15, 1–14. doi: 10.4018/IJITSA.307027

Crossref Full Text | Google Scholar

Mani, M. (2024). An Exploration of Contemporary Trends in Finance Research. J. Knowl. Econ. 22, 1–26. doi: 10.1007/s13132-024-02373-2

PubMed Abstract | Crossref Full Text | Google Scholar

Mumuni, A., and Mumuni, F. (2024). Automated data processing and feature engineering for deep learning and big data applications: a survey. J. Inf. Intell. 3, 113–153. doi: 10.1016/j.jiixd.2024.01.002

Crossref Full Text | Google Scholar

Murugan, M. S. (2023). Large-scale data-driven financial risk management & analysis using machine learning strategies. Meas. Sens. 27:100756. doi: 10.1016/j.measen.2023.100756

Crossref Full Text | Google Scholar

Nguyen, D. K., Sermpinis, G., and Stasinakis, C. (2023). Big data, artificial intelligence and machine learning: a transformative symbiosis in favour of financial technology. Eur. Finan. Manag. 29, 517–548. doi: 10.1111/eufm.12365

Crossref Full Text | Google Scholar

Nivetha, V., Shameem, A., Suhashini, J., Venugopal, B., Kulshrestha, N., and Bhanushali, M. M. (2024). “Exploring the Use of Big Data in Financial Risk Management and Fraud Detection,” in 2024 7th International Conference on Contemporary Computing and Informatics (IC3I), 1466–1471. IEEE.

Google Scholar

Nurunnabi, M. (2021). Disclosure, transparency, and international financial reporting standards. New York, NY: Emerald Publishing Limited, 199–311.

Google Scholar

Osman, R., and El-Gendy, S. (2025). Interconnected and resilient: A CGE analysis of AI-driven cyberattacks in global trade. Risk Anal. 45, 846–862. doi: 10.1111/risa.14321

PubMed Abstract | Crossref Full Text | Google Scholar

Page, M. J., McKenzie, J. E., Bossuyt, P. M., Boutron, I., Hoffmann, T. C., Mulrow, C. D., et al. (2021). The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. BMJ 372, 1–24. doi: 10.26633/rpsp.2022.112

Crossref Full Text | Google Scholar

Rozony, F. Z., Aktar, M. N. A., Ashrafuzzaman, M., and Islam, A. (2024). A systematic review of big data integration challenges and solutions for heterogeneous data sources. Academic J. Bus. Admin. Innova. Sust. 4, 1–18. doi: 10.69593/ajbais.v4i04.111

Crossref Full Text | Google Scholar

Shang, N., Wang, H., and Fan, J. (2025). A robust large-scale multi-criteria decision algorithm for financial risk management with interval-valued picture fuzzy information. Symmetry 17:144. doi: 10.3390/sym17010144

Crossref Full Text | Google Scholar

Singh, S., Jaiswal, R., Gupta, S., and Kumar, C. (2024). What we know and what should we know about the future of blockchain in finance. F1000Res 13:1051. doi: 10.12688/f1000research.153215.1

PubMed Abstract | Crossref Full Text | Google Scholar

Srinivasan, S., and Kamalakannan, T. J. C. E. (2017). Multi criteria decision making in financial risk management with a multi-objective genetic algorithm. Comput. Econ. 52, 443–457. doi: 10.1007/s10614-017-9683-7

Crossref Full Text | Google Scholar

Tayfor, A. E., Ali, N. B. M., and Elfaki, I. O. (2024). Optimized deep fuzzy neural network for financial risk evaluation in Fintech model. Appl. Math. 18, 1507–1517. doi: 10.18576/amis/180626

Crossref Full Text | Google Scholar

Theodorakopoulos, L., Antonopoulou, H., Mamalougou, V., and Giotopoulos, K. C. (2022). The drivers of volume volatility: a big data analysis based on economic uncertainty measures for the Greek banking system. Banks Bank Syst. 17, 49–57. doi: 10.21511/bbs.17(3).2022.05

Crossref Full Text | Google Scholar

Theodorakopoulos, L., Theodoropoulou, A., and Stamatiou, Y. (2024). A state-of-the-art review in big data management engineering: real-life case studies, challenges, and future research directions. Eng 5, 1266–1297. doi: 10.3390/eng5030068

Crossref Full Text | Google Scholar

Tiwari, A., Masilamani, M., Rao, T. S., Zope, S., Deepak, S. A., and Karthick, L. (2025). Shaping the future: emerging trends and strategic predictions in big data and AI. AI Revi.g Data, 125–154.

Google Scholar

Udeh, E. O., Amajuoyi, P., Adeusi, K. B., and Scott, A. O. (2024). The role of big data in detecting and preventing financial fraud in digital transactions. World J. Adv. Res. Rev. 22, 1746–1760. doi: 10.30574/wjarr.2024.22.2.1575

Crossref Full Text | Google Scholar

Wang, N., and Wang, K. (2022). Internet financial risk management in the context of big data and artificial intelligence. Math. Probl. Eng. 2022, 1–6. doi: 10.1155/2022/6219489

PubMed Abstract | Crossref Full Text | Google Scholar

Xia, P., Zhu, X., Charles, V., Zhao, X., and Peng, M. (2024). A novel heuristic-based selective ensemble prediction method for digital financial fraud risk. IEEE Trans. Eng. Manag. 71, 8002–8018. doi: 10.1109/TEM.2024.3385298

Crossref Full Text | Google Scholar

Yue, H., Liao, H., Li, D., and Chen, L. (2021). Enterprise financial risk management using information fusion technology and big data mining. Wirel. Commun. Mob. Comput. 2021:3835652. doi: 10.1155/2021/3835652

Crossref Full Text | Google Scholar

Zhou, X. (2023). Financial risk assessment management of state-owned enterprises based on cloud accounting in the era of big data. Appl. Math. Nonlinear Sci. 9:4. doi: 10.2478/amns.2023.2.00043

Crossref Full Text | Google Scholar

Zhou, H., Sun, G., Fu, S., Liu, J., Zhou, X., and Zhou, J. (2019). A big data mining approach of PSO-based BP neural network for financial risk management with IoT. IEEE Access 7, 154035–154043. doi: 10.1109/ACCESS.2019.2948949

Crossref Full Text | Google Scholar

Keywords: systemic risk, financial decision-making, data governance, digital transformation, Fintech

Citation: Theodorakopoulos L, Theodoropoulou A and Bakalis A (2025) Big data in financial risk management: evidence, advances, and open questions: a systematic review. Front. Artif. Intell. 8:1658375. doi: 10.3389/frai.2025.1658375

Received: 08 July 2025; Accepted: 29 August 2025;
Published: 01 October 2025.

Edited by:

Ehi Eric Esoimeme, James Hope University, Nigeria

Reviewed by:

Abiola Babajide, James Hope University, Nigeria
Amin Karami, University of East London, United Kingdom
Wence Nwoga, James Hope University, Nigeria

Copyright © 2025 Theodorakopoulos, Theodoropoulou and Bakalis. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Leonidas Theodorakopoulos, dGhlb2RsZW9AdXBhdHJhcy5ncg==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.