Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Public Health, 10 July 2025

Sec. Health Economics

Volume 13 - 2025 | https://doi.org/10.3389/fpubh.2025.1597381

This article is part of the Research TopicPublic Health Outcomes: The Role of Social Security Systems in Improving Residents' Health WelfareView all 74 articles

The impact of social security systems on public health outcomes: an economic perspective on machine translation applications


Shuhua NiuShuhua Niu1Weihua Li
Weihua Li1*Li-LiLi-Li2
  • 1Nantong University, Nantong, China
  • 2Taizhou Vocation College of Science Technology, School of Accounting Finance, Taizhou, Zhejiang, China

Introduction: The relationship between social security systems and public health outcomes has garnered significant attention due to its impact on improving health welfare and promoting economic stability. Social security systems, including pension schemes, healthcare benefits, and unemployment support, are essential for shaping societal wellbeing by influencing healthcare access, labor market participation, and overall economic resilience. However, traditional methods for evaluating these systems often fail to capture the complex dynamics of policy interventions over time.

Methods: To address this, we propose an advanced economic policy modeling framework that integrates dynamic optimization techniques with machine translation applications. Machine translation applications refer to the use of automated translation tools to facilitate communication in multilingual contexts, ensuring equal access to healthcare and social services.

Results: These applications contribute to the evaluation of social security systems by improving the accessibility and efficiency of service delivery, particularly in linguistically diverse populations.

Discussion: By incorporating both economic policy modeling and machine translation technology, our framework offers a comprehensive analysis of social security interventions, demonstrating how well-optimized policies can enhance public health outcomes while ensuring fiscal sustainability.

1 Introduction

The study of social security systems and their impact on public health outcomes has gained increasing attention due to the growing need for sustainable economic policies that enhance population wellbeing (1). Social security systems not only provide financial stability to vulnerable groups but also contribute to broader public health improvements by ensuring access to healthcare, reducing poverty-related diseases, and promoting overall social welfare (2). Moreover, the efficiency of these systems is closely tied to their ability to distribute resources effectively, particularly in multilingual societies where communication barriers can hinder access to essential services (3). Machine translation (MT) has emerged as a crucial tool in breaking down language barriers, facilitating access to healthcare services, and improving the administration of social security benefits (4). By integrating economic perspectives with technological advancements, this research aims to explore how machine translation can optimize the functioning of social security systems, ultimately leading to better public health outcomes (5). For example, several healthcare organizations, including the World Health Organization (WHO), have used machine translation tools to provide health-related information to non-native speakers, especially during global health crises like the COVID-19 pandemic, where timely and accurate communication was essential. Similarly, in countries with multilingual populations, the United States Social Security Administration (SSA) uses machine translation to provide translated documents and online services, ensuring equitable access to pension schemes, unemployment benefits, and healthcare for non-English speakers. By integrating economic perspectives with technological advancements, this research aims to explore how machine translation can optimize the functioning of social security systems, ultimately leading to better public health outcomes.

Social security systems play a critical role not only in stabilizing economic conditions but also in shaping population health. As institutional mechanisms, they encompass programs such as healthcare coverage, unemployment benefits, disability support, and pension schemes, all of which help buffer individuals from economic shocks that often lead to negative health outcomes. By reducing financial stress and improving access to medical services, these systems contribute to better health behaviors, increased utilization of preventive care, and improved management of chronic conditions. Empirical studies across OECD and developing countries have shown that more inclusive and well-funded social protection programs correlate with longer life expectancy, lower infant mortality, and reduced incidence of poverty-related diseases. Importantly, the impact of social security on health is both direct and indirect. Direct effects include subsidized healthcare access or insurance coverage, while indirect effects arise from the income security these programs provide, which influences determinants such as nutrition, housing stability, and mental wellbeing. In multilingual and socioeconomically diverse societies, the effectiveness of these systems further depends on how well they address barriers like language access and digital inclusion–factors that are increasingly recognized as determinants of health equity. Against this backdrop, it becomes crucial to explore not only the economic design of social security systems, but also the technologies, such as machine translation, that can enhance their accessibility and operational reach.

To address the limitations of early approaches, traditional economic models of social security systems initially relied on symbolic AI and knowledge representation to analyze policy impacts (6). These models were primarily rule-based and leveraged expert systems to assess economic indicators such as unemployment rates, healthcare expenditures, and pension distributions (7). By encoding economic rules and welfare policies, these systems sought to provide analytical insights into public health outcomes (8). However, these methods faced significant challenges in dealing with the complexity and variability of real-world economic and health data (9). The rigidity of symbolic AI made it difficult to adapt to dynamic policy changes and diverse linguistic contexts, particularly when analyzing the effects of social security in multilingual environments (10). Moreover, the reliance on handcrafted rules limited the scalability and generalizability of these models, reducing their effectiveness in cross-border policy evaluations. To address these issues, we sought more flexible and data-driven approaches.

With the advent of machine learning, data-driven methods began to revolutionize economic analysis and policy evaluation within social security systems (11). Statistical and machine learning models enabled we to analyze large-scale data from various sources, including healthcare records, social security databases, and economic indicators (12). These models could identify patterns in the relationship between social security expenditures and public health outcomes, providing a more empirical basis for policy recommendations (13). Machine translation also benefited from machine learning approaches, with statistical models such as phrase-based translation systems improving linguistic accessibility in social security administration (14). However, despite these advancements, traditional machine learning methods still faced challenges in handling unstructured data, capturing complex policy interactions, and ensuring real-time translation accuracy for diverse dialects and regional variations (15). Furthermore, economic evaluations still required extensive feature engineering, making it difficult to adapt machine learning models to evolving policy frameworks.

To overcome the limitations of traditional AI and machine learning models, deep learning and pre-trained language models have transformed the analysis of social security systems and their impact on public health (16). Neural networks, particularly transformer-based architectures such as BERT and GPT, have significantly improved the capabilities of machine translation by providing more context-aware and accurate translations (17). These models facilitate real-time multilingual communication in social security services, ensuring that non-native speakers can access crucial health and financial information without language barriers (18). deep learning has enhanced economic modeling by enabling automated text analysis of policy documents, social media sentiment analysis on welfare programs, and predictive analytics for public health interventions (19). Despite these advancements, challenges such as model bias, computational costs, and interpretability remain significant barriers to large-scale adoption (20). Addressing these issues requires a combination of domain-specific fine-tuning, regulatory oversight, and ethical considerations in deploying AI-driven social security systems.

Given the limitations of previous methods in addressing the economic and public health implications of social security systems, we propose a novel approach that integrates deep learning-based machine translation with real-time economic analysis. This method leverages state-of-the-art multilingual models to enhance accessibility while utilizing dynamic economic forecasting to optimize policy interventions. Unlike previous models, our approach focuses on adaptability, allowing for continuous updates based on policy changes and economic trends. By combining advanced NLP techniques with economic modeling, we aim to bridge the gap between social security administration and public health outcomes, ensuring equitable access to benefits across linguistic and demographic groups. Moreover, this framework enhances policy evaluation by incorporating real-time data streams, enabling governments to make informed decisions that balance fiscal sustainability with public wellbeing.

• Our method introduces a novel framework that combines deep learning-based machine translation with economic forecasting, providing a comprehensive approach to optimizing social security systems (Table 1).

• Unlike previous models, our system is designed to accommodate diverse policy environments and linguistic contexts, ensuring broader applicability across different countries and economic conditions.

• Experimental results demonstrate significant improvements in translation accuracy, economic forecasting precision, and public health outcome predictions, highlighting the practical benefits of our approach.

Table 1
www.frontiersin.org

Table 1. Glossary of key terms.

2 Related work

2.1 Economic impact of social security systems on public health

Social security systems play a pivotal role in shaping public health outcomes by providing financial protection and access to healthcare services (21). These systems encompass various programs, including pensions, unemployment benefits, and health insurance, aiming to mitigate economic risks associated with illness, disability, and aging. The economic impact of social security on public health is multifaceted, influencing healthcare accessibility, quality, and overall population health (22). social security systems contribute to the reduction of economic inequalities, which are closely linked to health disparities. By offering financial support to vulnerable populations, these systems help alleviate poverty-related health issues. For instance, income support programs can enable individuals to afford nutritious food, safe housing, and necessary medical care, thereby improving health outcomes. A study highlighted that comprehensive social protection is essential for reducing economic inequality and, consequently, health inequalities (23). the provision of universal health coverage through social security ensures that individuals have access to necessary medical services without financial hardship (24). This accessibility leads to early detection and treatment of diseases, reducing morbidity and mortality rates. The World Health Organization emphasizes that health systems are vital to economic performance and stability, and are key to achieving sustainable development Investing in health systems not only improves public health but also yields economic benefits by enhancing workforce productivity and reducing healthcare costs associated with advanced diseases (25). social security systems that include preventive healthcare services can lead to long-term cost savings. Preventive measures, such as vaccinations and regular health screenings, can prevent the onset of diseases, thereby reducing the need for expensive treatments. The economic case for health equity suggests that investing in preventive care is cost-effective and can lead to substantial savings in healthcare expenditure (26).

2.2 Role of preventive healthcare in economic efficiency

Preventive healthcare is a cornerstone of public health that focuses on disease prevention and health promotion. From an economic perspective, investing in preventive healthcare can lead to significant cost savings and improved economic efficiency. This approach reduces the burden of chronic diseases, decreases healthcare expenditures, and enhances workforce productivity (27). Chronic diseases, such as heart disease, diabetes, and cancer, are leading causes of morbidity and mortality worldwide. These conditions often require long-term, expensive treatments, imposing substantial economic burdens on individuals and healthcare systems. Preventive healthcare aims to reduce the incidence of these diseases through interventions like lifestyle counseling, vaccinations, and regular screenings. By preventing disease onset, healthcare systems can avoid the high costs associated with treatment and management of chronic illnesses (28). Economic evaluations have demonstrated the cost-effectiveness of preventive measures. For example, childhood immunizations have been shown to yield high returns on investment by preventing diseases that would require costly treatments. According to Healthy People 2020, for every birth cohort that receives the routine childhood vaccination schedule, direct healthcare costs are reduced by 9.9 billion, and society saves 33.4 billion in indirect costs (29). Furthermore, preventive healthcare contributes to economic efficiency by enhancing workforce productivity. Healthier individuals are more capable of maintaining employment, have fewer sick days, and are more productive at work. This increased productivity translates into economic gains for employers and the economy at large. The health capital theory posits that investments in health, such as preventive care, improve an individual's stock of health, leading to greater economic output and growth (30). However, the implementation of preventive healthcare measures requires initial investments, and the benefits may not be immediately apparent. Policymakers must consider the long-term economic advantages of such investments, including reduced healthcare costs and a more robust economy due to a healthier workforce (31). The Organisation for Economic Co-operation and Development (OECD) highlights that while preventive measures may not always lead to immediate cost savings, they are essential for improving population health and economic outcomes in the long run.

2.3 Economic benefits of machine translation in healthcare communication

Machine translation (MT) technologies have introduced several measurable economic benefits in healthcare communication, particularly in multilingual and linguistically diverse regions. MT systems now offer real-time translation capabilities that effectively bridge language barriers (32), reducing administrative burdens and operational costs by streamlining the intake process, minimizing the need for human interpreters, and improving documentation efficiency (33). These savings are especially significant in public health settings where interpreter resources are limited and communication delays can have direct health and financial consequences. Language barriers in healthcare often result in misunderstandings, misdiagnoses, and inadequate treatment plans, which can escalate medical costs and worsen health outcomes (34). MT helps mitigate these risks by facilitating clear communication between healthcare providers and patients who speak different languages, thereby reducing repeat visits and shortening hospital stays (33). Moreover, by enabling providers to serve a broader patient base without full reliance on human interpreters, MT allows for more flexible and cost-effective allocation of interpreter services to cases that require deeper cultural or contextual nuance (35). our economic framework extends the analysis by modeling indirect benefits. For example, by reducing miscommunication and delays in service delivery, MT contributes to better health outcomes among non-native speakers, which in turn leads to lower public healthcare expenditures over time. Furthermore, equitable access to healthcare via MT can support improved labor productivity and labor market participation among immigrant populations, generating broader macroeconomic gains (36). While the benefits discussed above are drawn from both literature and our modeling assumptions, we have clarified the source of each category in the revised text. We also acknowledge that this list is not exhaustive, and that ongoing innovations in natural language processing may continue to reveal new dimensions of MT's economic impact in the healthcare and social policy domains.

3 Method

3.1 Overview

Economic policy plays a fundamental role in shaping the financial and social structures of nations, influencing economic growth, stability, and distribution of wealth (Figure 1). The study of economic policy involves analyzing various mechanisms that governments employ to regulate economic activities, including monetary policy, fiscal policy, trade policy, and industrial policy. In this work, we develop a novel approach to economic policy modeling by integrating advanced analytical techniques with empirical validation.

Figure 1
www.frontiersin.org

Figure 1. Conceptual framework illustrating how machine translation performance influences public health outcomes through enhanced language accessibility and improved operational efficiency within social security systems.

In Section 3.2, we introduce the theoretical foundation underlying economic policy formulation, focusing on the principles of policy intervention and regulatory frameworks. This provides a rigorous basis for understanding how governments manage macroeconomic variables to achieve specific objectives such as controlling inflation, reducing unemployment, and fostering economic growth. formalizes the problem by defining a mathematical framework that captures the complexities of economic decision-making. We employ symbolic representations to model policy instruments, economic agents, and macroeconomic indicators, ensuring that our approach is both comprehensive and adaptable to various economic contexts. In Section 3.3, we introduce our innovative model, which extends traditional economic policy analysis by incorporating dynamic optimization techniques and statistical learning methodologies. Our model enhances predictive capabilities, allowing for more precise assessment of policy impacts under different economic scenarios. in Section 3.4, we propose a strategic framework that leverages domain-specific knowledge to optimize policy interventions. By designing targeted strategies based on economic theory and empirical data, our approach offers a refined mechanism for policy evaluation and implementation.

3.2 Preliminaries

Economic policy encompasses a broad range of government interventions aimed at regulating economic activity to achieve specific macroeconomic and microeconomic objectives. These interventions include fiscal policies, which influence government spending and taxation; monetary policies, which regulate money supply and interest rates; trade policies, which govern international trade and tariffs; and industrial policies, which promote sectoral growth and innovation. In this section, we provide a formalized representation of the economic policy problem using mathematical and symbolic notation, ensuring a rigorous foundation for subsequent analysis.

Let E represent the overall economy, defined as a system of agents, institutions, and policies. We model E as a dynamic system characterized by a state vector st at time t, which captures key macroeconomic indicators such as GDP (Yt), inflation rate (πt), unemployment rate (ut), and government debt (Dt). The evolution of the economy is governed by a set of policy instruments pt, which are determined by government intervention.

st+1=f(st,pt,et),    (1)

where f is a transition function that captures the structural relationships within the economy, and et represents external shocks, such as global financial crises or supply chain disruptions.

Fiscal policy, denoted as ptF, consists of government expenditure Gt and taxation Tt. The government budget constraint is given by:

Dt+1=(1+rt)Dt+Gt-Tt.    (2)

Monetary policy, denoted as ptM, is implemented through the central bank's control over interest rates rt and money supply Mt. The Taylor rule provides a common framework for setting interest rates:

rt=r*+α(πt-π*)+β(Yt-Y*).    (3)

Trade policy, denoted as ptT, affects import and export dynamics through tariffs τt and quotas. The trade balance equation is:

NXt=Xt-Mt.    (4)

Industrial policy, denoted as ptI, includes subsidies St and research and development (R&D) investments Rt. The effect of industrial policy on sectoral growth can be modeled as:

gtS=ϕSt+γRt.    (5)

3.3 Dynamic Equilibrium Policy Model (DEPM)

In this section, we introduce the Dynamic Equilibrium Policy Model (DEPM), a novel framework for analyzing economic policy decisions (Table 2). The DEPM integrates macroeconomic state dynamics, policy interventions, and agent-based optimization within a unified mathematical structure. Our model extends traditional policy analysis by incorporating dynamic optimization, stochastic shocks, and equilibrium constraints, providing a more robust representation of policy effects in complex economic environments.

Table 2
www.frontiersin.org

Table 2. Summary of Dynamic Equilibrium Policy Model (DEPM) variants.

As shown in Figure 2, illustrates the architecture of the proposed Dynamic Equilibrium Policy Model (DEPM), consisting of four core modules: (a) state dynamics and policy interaction, (b) equilibrium conditions and constraints, (c) stochastic shocks and optimization, and (d) hierarchical feature extraction. Module (a) models the interaction between macroeconomic variables (e.g., GDP, inflation, unemployment) and policy instruments (e.g., fiscal spending, interest rate, tariffs). Module (b) ensures system-wide consistency through market-clearing conditions such as national income identity and labor market equilibrium. Module (c) incorporates Gaussian stochastic shocks to account for real-world uncertainties. Module (d) employs a deep neural architecture based on a Masked Mamba Encoder-Decoder to capture nonlinear temporal dependencies across macroeconomic indicators. This framework enables realistic simulation of dynamic policy interventions.

Figure 2
www.frontiersin.org

Figure 2. The diagram illustrates the key components of the Dynamic Equilibrium Policy Model (DEPM), integrating macroeconomic state dynamics, policy interventions, and stochastic optimization. It includes (a) state dynamics and policy interaction, showing how economic variables evolve under policy influences; (b) equilibrium conditions and constraints, ensuring macroeconomic consistency; (c) stochastic shocks and optimization, modeling uncertainties in policy decisions; and (d) hierarchical feature extraction, leveraging deep learning-based encoders and decoders for complex economic modeling and prediction.

3.3.1 State dynamics and policy interaction

Let the economic state at time t be represented by a vector st, which includes key macroeconomic variables:

st={Yt,πt,ut,Dt,NXt,rt,Mt,Ct,It},    (6)

where Yt is the GDP, πt is the inflation rate, ut is the unemployment rate, Dt is government debt, NXt is net exports, rt is the interest rate, Mt is the money supply, Ct represents private consumption, and It represents investment.

The economy evolves according to the stochastic transition function:

st+1=f(st,pt,et),    (7)

where pt represents the set of policy instruments, and et denotes exogenous shocks such as productivity shocks, financial shocks, or external trade shocks.

The government controls a set of policy levers pt={ptF,ptM,ptT,ptI}, where fiscal policy instruments are denoted by ptF={Gt,Tt}, monetary policy instruments by ptM={rt,Mt}, trade policy instruments by ptT={τt,qt}, and industrial policy instruments by ptI={St,Rt}.

Fiscal policy affects aggregate demand through government spending and taxation. The fiscal rule is given by:

Gt=λ1Yt+λ2ut+λ3Dt+εtF,    (8)

where Gt represents government expenditure at time t, which is influenced by the current GDP (Yt), unemployment rate (ut), and government debt (Dt). The coefficients λ1, λ2, and λ3 represent the responsiveness of government expenditure to changes in these variables. The term εFt represents a fiscal policy shock, capturing external factors that may affect government spending.

Monetary policy is determined using an extended Taylor rule, adjusting the nominal interest rate in response to deviations in inflation and output from their targets:

rt=r*+α(πt-π*)+β(Yt-Y*)+γut+εtM.    (9)

where rt is the nominal interest rate at time t, adjusted based on the target inflation rate (π*), the output gap (Yt-Y*), and the unemployment rate (ut). The coefficients α, β, and γ describe the sensitivity of interest rates to these economic indicators. εMt represents a monetary policy shock, reflecting unexpected changes in policy or external factors.

Trade policy can influence net exports through tariffs and quotas. A simple rule-based trade policy function is:

τt=η1NXt+η2Yt+εtT,    (10)

where τt represents the tariff rate at time t, which is influenced by the net exports (NXt) and GDP (Yt). The coefficients η1 and η2 capture the responsiveness of trade policy to changes in these variables. εTt represents a trade policy shock.

Industrial policy aims to support innovation and sectoral development. The government may subsidize research and strategic industries, modeled as:

St=ρ1It+ρ2Yt+εtI,    (11)

where St is the subsidy rate at time t, influenced by investment (It) and GDP (Yt). The coefficients ρ1 and ρ2 indicate the impact of these variables on industrial policy. εIt represents a shock to industrial policy.

3.3.2 Equilibrium conditions and constraints

Equilibrium conditions ensure consistency across macroeconomic variables, reflecting the interactions among different sectors of the economy. The goods market equilibrium condition ensures that total output equals the sum of consumption, investment, government spending, and net exports:

Yt=Ct+It+Gt+NXt,    (12)

where Ct represents private consumption, It is investment, Gt is government spending, and NXt is net exports, which capture the difference between exports and imports. This identity ensures that all goods produced within the economy are accounted for through different forms of demand.

In the labor market, the unemployment rate is determined by deviations from its natural rate and the effects of real wages. The equilibrium condition in the labor market follows:

ut=ut*+θ(wt-wt*),    (13)

where ut is the actual unemployment rate, ut* is the natural rate of unemployment, wt is the real wage, and wt* is the equilibrium wage level. The parameter θ captures the responsiveness of unemployment to wage deviations. This equation highlights the role of labor market rigidities and wage-setting mechanisms in determining employment levels.

External trade equilibrium is maintained through the balance of payments condition, ensuring that the sum of net exports and the capital account balance is zero:

NXt+KAt=0,    (14)

where KAt denotes the capital account balance, which reflects cross-border capital flows. A surplus in the trade balance (NXt>0) must be offset by a capital account deficit (KAt < 0), and vice versa. This relationship ensures that financial transactions align with trade imbalances, maintaining external stability.

The government formulates optimal policies to maximize a social welfare function that incorporates economic stability and growth. The objective function is defined as:

maxpt𝔼t=0δtU(Yt,ut,πt),    (15)

where pt represents policy instruments, δ is the discount factor reflecting intertemporal preferences, and U(·) is a utility function that depends on output (Yt), unemployment (ut), and inflation (πt). The policymaker aims to achieve an optimal balance between economic growth, labor market conditions, and price stability.

The first-order condition for optimal policy intervention ensures that marginal changes in policy instruments do not generate welfare losses. This condition is given by:

UYtYtpt+Uututpt+Uπtπtpt=0.    (16)

This equation states that the government adjusts its policy instruments such that the weighted sum of marginal utilities with respect to output, unemployment, and inflation is equal to zero. The policy intervention takes into account the trade-offs between different economic objectives and ensures an optimal allocation of resources.

3.3.3 Stochastic shocks and optimization

The Dynamic Economic Policy Model (DEPM) incorporates stochastic shocks to model uncertainty in economic fluctuations, allowing for a more realistic representation of macroeconomic dynamics. Let εt represent a vector of stochastic disturbances:

εt~N(0,Σ),    (17)

where Σ is the covariance matrix that captures the volatility and interdependence of various exogenous shocks, including fiscal policy shocks, monetary policy shocks, and trade disruptions. These disturbances propagate through the system, influencing state variables and observed economic indicators.

As shown in Figure 3 presents the mechanism for processing stochastic disturbances in the DEPM framework. Economic shocks, modeled as multivariate Gaussian noise, are first processed using patch and position embedding techniques to capture localized variations. A Fast Fourier Transform (FFT) is then applied to map signals into the frequency domain, enabling the identification of periodic or cyclical patterns in economic volatility. These transformed signals are integrated into the state transition functions to inform optimal policy computation. This frequency-aware approach allows for a more nuanced response to economic disturbances, particularly in high-frequency or crisis scenarios.

Figure 3
www.frontiersin.org

Figure 3. The diagram illustrates the stochastic shocks and optimization, integrating Fourier-based shock processing and state-space modeling. The framework processes stochastic shocks through patch and position embedding, spatial mixing, and state transition functions, utilizing FFT to analyze frequency-domain impacts on macroeconomic stability and optimal policy decisions.

The state-space representation of the model consists of a transition equation and an observation equation:

st+1=f(st,pt)+εt,    (18)
yt=h(st)+νt.    (19)

Here, st represents the vector of state variables, which includes key macroeconomic indicators such as output, inflation, interest rates, and consumption. The function f(·) governs the law of motion of state variables, capturing the effects of policy variables pt and stochastic shocks εt. The observation equation links the unobserved state variables to observed economic indicators yt, incorporating measurement noise νt.

The DEPM is solved using dynamic programming and numerical simulation techniques. The goal is to determine the optimal policy sequence pt that maximizes the expected discounted utility over time. The Bellman equation characterizing the optimization problem is:

V(st)=maxpt{U(st,pt)+δ𝔼V(st+1)}.    (20)

where V(st) is the value function, U(st, pt) represents the period utility function, and δ∈(0, 1) is the discount factor ensuring time-consistent decision-making. The expectation operator E accounts for uncertainty in future states.

To compute the equilibrium policy, value function iteration is employed. Starting with an initial guess for V(s), the algorithm iterates until convergence, updating the value function at each step. Convergence is achieved when the state variables satisfy:

||st+1-st||<ϵ.    (21)

where ϵ is a pre-defined tolerance level. The iteration process ensures that the system reaches a steady-state equilibrium, where optimal policies effectively mitigate the impact of stochastic shocks and guide the economy toward stability.

3.4 Adaptive Policy Optimization Strategy (APOS)

Building on the Dynamic Equilibrium Policy Model (DEPM) introduced in the previous section, we propose the Adaptive Policy Optimization Strategy (APOS) to refine economic decision-making under uncertainty. The APOS integrates dynamic programming, reinforcement learning, and stochastic control mechanisms to enhance policy adaptability in response to evolving economic conditions. Our strategy ensures that fiscal, monetary, trade, and industrial policies dynamically adjust to economic fluctuations, minimizing adverse shocks while optimizing long-term economic welfare.

As shown in Figure 4 depicts the Adaptive Policy Optimization Strategy (APOS), designed to dynamically adjust policy responses to evolving economic conditions. The first module introduces a policy correction term Δpt to refine initial optimal policies pt* based on real-time economic feedback using a gradient-based adjustment. The second module employs Temporal Difference (TD) Learning to iteratively update the value function V(st), optimizing long-term utility. The third module coordinates multiple autonomous policy agents (fiscal, monetary, trade, industrial) using a Nash bargaining framework. Together, these components ensure both local adaptability and global policy coherence under uncertainty.

Figure 4
www.frontiersin.org

Figure 4. Illustration of the Adaptive Policy Optimization Strategy (APOS) framework, integrating optimization via policy adjustment, reinforcement learning for adaptability, and risk-aware multi-agent coordination. The framework refines policy decisions dynamically using reinforcement learning and stochastic control mechanisms to enhance adaptability and economic stability.

3.4.1 Optimization via policy adjustment

Let pt* denote the optimal policy set obtained from the DEPM framework:

pt*=argmaxpt𝔼t=0δtU(st,pt).    (22)

This policy set represents an optimal decision rule that maximizes the expected cumulative utility under a discount factor δ. However, static policy rules often struggle to accommodate real-time economic fluctuations, necessitating an adaptive mechanism to refine decision-making dynamically.

To address this limitation, the Adaptive Policy Optimization Strategy (APOS) introduces an adjustment term Δpt, modifying the initial policy set:

pt=pt*+Δpt.    (23)

where pt* represents the initial optimal policy set, and Δpt denotes the adjustment made to the policy based on feedback from the economy. This allows the model to respond dynamically to changes in economic conditions.

The policy adjustment process follows a gradient-based learning framework, where updates occur iteratively according to:

Δpt=-ηpL(st,pt).    (24)

where η is the learning rate, controlling the size of policy updates, and ∇pL(st, pt) represents the gradient of the loss function L with respect to the policy parameters. The gradient indicates the direction in which the policy should be adjusted to minimize the economic cost.

To further refine policy adjustments, a regularization term Ω(pt) is incorporated into the loss function, constraining abrupt policy fluctuations:

L(st,pt)=E[C(st,pt)]+λΩ(pt).    (25)

The term E[C(st, pt)] represents the expected cost associated with the current policy pt at state st. The regularization term λΩ(pt) penalizes large fluctuations in policy, helping to maintain stability in the policy updates.

To prevent excessive oscillations in policy adjustments, a momentum term μ is introduced:

Δpt=-ηpL(st,pt)+μΔpt-1.    (26)

The term μΔpt−1 ensures that the policy updates take into account the previous adjustment, smoothing out abrupt changes and preventing instability. The parameter μ controls the momentum, balancing the rate of adaptation and ensuring more gradual transitions.

3.4.2 Reinforcement learning for adaptability

To enhance adaptability, APOS employs reinforcement learning (RL) to refine policy choices based on observed outcomes. Let V(st) represent the value function capturing long-term economic performance:

V(st)=maxpt{U(st,pt)+δ𝔼V(st+1)}.    (27)

Here, U(st, pt) denotes the immediate utility of selecting policy pt in state st, and δ∈(0, 1) is the discount factor that accounts for the importance of future rewards. The expectation is taken over all possible future states st+1, considering the probabilistic transitions influenced by the current policy.

The policy is refined using temporal difference (TD) learning to update the value function iteratively:

V(st)V(st)+α(rt+δV(st+1)-V(st)),    (28)

where α is the learning rate, determining the step size of updates, and rt is the realized economic reward obtained from the transition. This update mechanism allows APOS to adjust policy valuation dynamically, ensuring adaptability to changing economic conditions.

To improve decision-making, APOS employs an action selection mechanism based on an ϵ-greedy strategy:

pt={arg maxpV(st,p),with probability 1ϵ,random policy,with probability ϵ.    (29)

This strategy balances exploitation (selecting the best-known policy) and exploration (choosing a random policy to discover potentially better options). A smaller ϵ prioritizes exploitation, while a larger ϵ encourages broader exploration.

The transition dynamics of the system are modeled using a Markov decision process (MDP), where the probability of transitioning to a new state st+1 depends on the current state st and policy pt:

P(st+1|st,pt)=wtP(st+1|st,pt,wt)P(wt|st,pt).    (30)

Here, wt represents external stochastic factors influencing state transitions, and the total probability is obtained by marginalizing over all possible realizations of wt.

To ensure convergence of the value function, APOS incorporates a dynamic learning rate schedule:

αt=11+βt,    (31)

where β is a decay parameter controlling the rate at which the learning rate decreases over time. This adaptive step size prevents excessive fluctuations in value updates while allowing continued policy improvements.

3.4.3 Risk-Aware multi-agent coordination

To mitigate economic volatility and enhance resilience against external shocks, APOS integrates stochastic control theory to optimize risk-adjusted policies.

As shown in Figure 5 provides a detailed breakdown of the risk-aware multi-agent coordination component within APOS. The architecture begins with encoding the variance of state transitions Rt = Var(st+1|st, pt) into risk representations using deep convolutional operations. These are followed by linear projections of each agent's policy features, enabling inter-agent comparability. Aggregation is performed using attention-based fusion mechanisms, combining individual policy vectors into a unified risk-sensitive strategy. Matrix operations and skip connections preserve semantic consistency while enabling flexible policy refinement. The agents update their strategies iteratively until convergence to a risk-optimized, Nash-consistent policy set.

Figure 5
www.frontiersin.org

Figure 5. The image illustrates the computational architecture for risk-aware multi-agent coordination within the APOS framework, showcasing a pipeline that integrates policy feature projection, representation embedding, depth-wise convolution, and risk-aware aggregation to optimize decision-making under uncertainty. Various operations such as element-wise multiplication, matrix multiplication, and addition are employed to enhance policy robustness while balancing economic performance and risk minimization.

Define Rt as the risk function capturing economic instability:

Rt=Var(st+1|st,pt),    (32)

where st represents the economic state at time t, and pt denotes the set of policy actions taken. The goal is to formulate policies that balance economic performance and risk minimization. A robust policy is derived by solving the optimization problem:

ptrobust=argmaxpt[U(st,pt)-λRt],    (33)

where U(st, pt) represents the utility function reflecting economic performance, and λ is a risk-aversion parameter that dictates the trade-off between stability and growth.

Given the complex interdependencies among fiscal, monetary, trade, and industrial policies, APOS models each policy domain as an autonomous agent Ai. Each agent optimizes its individual objective function Ji, defined as:

Ji(st,pti)=𝔼t=0δtUi(st,pti),    (34)

where δ∈(0, 1) is the discount factor ensuring convergence, and Ui(st,pti) is the utility of agent Ai under policy action pti. Agents must coordinate their actions to avoid conflicts and ensure overall policy coherence.

To achieve coordination, APOS employs a Nash bargaining framework. The equilibrium condition requires that the sum of individual policy gradients satisfies:

iptiJi(st,pti)=0.    (35)

This condition ensures that no agent can unilaterally improve its outcome without affecting others, leading to a balanced policy design. To iteratively reach equilibrium, each agent updates its policy according to:

pti,k+1=pti,k+κ(ptiJi-̄J),    (36)

where ̄J is the mean policy gradient across all agents, and κ is a step-size parameter that controls the adaptation rate. The iterative update continues until convergence is achieved.

Convergence is established when the policy update norm falls below a predefined threshold:

||ptk+1-ptk||<ϵ.    (37)

Here, ϵ is a small positive constant ensuring policy stability. By iteratively adjusting policies while considering risk and multi-agent interactions, APOS effectively stabilizes economic dynamics while optimizing performance.

4 Experimental setup

4.1 Dataset

The MLQA Dataset (37) is a benchmark dataset designed for evaluating cross-lingual question-answering models. It includes question-answer pairs in seven languages: English, Arabic, German, Spanish, Hindi, Vietnamese, and Chinese. The dataset is derived from Wikipedia and is widely used for training and evaluating multilingual natural language processing (NLP) models. MLQA enables research in transfer learning and zero-shot cross-lingual question answering by providing aligned parallel questions and answers across different languages. The FLoRes-200 Dataset (38) (Facebook Low Resource Languages Benchmark) is a dataset developed to evaluate machine translation (MT) models across 200 languages, particularly focusing on low-resource languages. It provides high-quality parallel text data for evaluating translation quality between a diverse set of language pairs. The dataset is used to benchmark MT models, particularly in low-resource settings, and has been instrumental in advancing research in multilingual translation. The OpenSubtitles Dataset (39) is a large-scale corpus of movie and TV subtitles, widely used for training and evaluating machine translation and dialogue systems. It consists of millions of aligned subtitle sentences in multiple languages, making it a valuable resource for multilingual NLP tasks. The dataset is particularly useful for developing conversational AI models, automatic subtitle generation, and translation models that require informal and context-aware language understanding. The BEA Dataset (40) (Building Educational Applications) is a dataset designed for grammatical error correction (GEC) and other educational NLP tasks. It contains English-language texts from non-native speakers and provides annotations for grammatical, lexical, and fluency errors. The dataset is widely used for training and evaluating GEC models, making it an essential resource for automated writing assistance tools and language learning applications.

4.2 Experimental details

In this study, we evaluate our anomaly detection model on four benchmark datasets: MLQA Dataset, FLoRes-200 Dataset, OpenSubtitles Dataset, and BEA Dataset. All experiments are conducted on a workstation equipped with an NVIDIA A100 GPU, 64GB RAM, and an Intel Xeon Silver 4216 CPU. The implementation is based on PyTorch, and we utilize CUDA acceleration to optimize computational efficiency. For training, we employ the Adam optimizer with an initial learning rate of 0.0001, a batch size of 32, and a weight decay of 1e−5. The learning rate is scheduled with a cosine annealing strategy to ensure smooth convergence. The training process runs for 100 epochs, with early stopping applied if the validation loss does not improve for 10 consecutive epochs. We apply data augmentation techniques, including random cropping, flipping, and Gaussian noise injection, to improve generalization. For the MLQA Dataset, we extract frames from videos at 10 fps and resize them to 256 × 256 resolution. The model is trained using a self-supervised approach, where normal samples are used for training, and anomalies are detected as deviations from learned normal representations. We evaluate the model using frame-level AUC (Area Under the Curve) and Equal Error Rate (EER). For the FLoRes-200 Dataset, we employ a patch-based strategy where each image is divided into non-overlapping patches of size 64 × 64. Anomalies are detected at the patch level using feature embeddings extracted from a deep convolutional network. We report pixel-level mean Intersection over Union (mIoU) and AUROC as evaluation metrics. For the OpenSubtitles Dataset, we preprocess the data by normalizing continuous features and applying one-hot encoding to categorical features. We use a deep autoencoder-based model, where normal network traffic is learned, and deviations are identified as intrusions. The model is evaluated using precision, recall, and F1-score. For the BEA Dataset, we utilize a recurrent neural network-based approach with LSTM layers to model temporal dependencies in time-series data. The model is trained in an unsupervised manner using reconstruction errors to detect anomalies. Performance is evaluated using the BEA scoring mechanism, which considers detection timeliness and accuracy. To ensure fair comparisons, we follow standardized evaluation protocols and perform five-fold cross-validation where applicable. Hyperparameters are fine-tuned based on grid search, and statistical significance tests are conducted to confirm the robustness of our results. We also conduct ablation studies to analyze the contribution of each component of our proposed model (Algorithm 1).

Algorithm 1
www.frontiersin.org

Algorithm 1. Training procedure of DEPM model.

4.3 Comparison with SOTA methods

To evaluate the effectiveness of our proposed anomaly detection model, we compare it against several state-of-the-art (SOTA) methods on the MLQA Dataset, FLoRes-200 Dataset, OpenSubtitles Dataset, and BEA Dataset. The comparison includes Transformer, LSTM-Attn, BART, T5, MBART, and NAT. The performance metrics used for evaluation are BLEU, METEOR, ROUGE-L, and TER. Our model consistently outperforms the baseline methods across all datasets, demonstrating superior anomaly detection capability. The quantitative results are presented in Tables 3, 4.

Table 3
www.frontiersin.org

Table 3. Comparison of our method with SOTA methods on MLQA dataset and FLoRes-200 dataset.

Table 4
www.frontiersin.org

Table 4. Comparison of our method with SOTA methods on OpenSubtitles dataset and BEA dataset.

In Figure 6, From the results, our method achieves the highest scores in BLEU, METEOR, and ROUGE-L, while obtaining the lowest TER, indicating more accurate anomaly detection. on the MLQA Dataset, our approach attains a BLEU score of 42.85, which surpasses the closest competitor, MBART, by a margin of 2.73. Similarly, on the FLoRes-200 Dataset, our model obtains a BLEU score of 44.67, demonstrating an improvement of 2.54 over MBART. The superior performance is attributed to our model's ability to capture complex spatial-temporal dependencies in video-based anomaly detection and localize fine-grained defects in high-resolution industrial images. The substantial improvements in ROUGE-L and METEOR further highlight the robustness of our method in feature extraction and anomaly localization. In Figure 7, For network-based anomaly detection, our approach achieves a BLEU score of 43.41 on the OpenSubtitles Dataset, outperforming MBART by 2.28. Similarly, for time-series anomaly detection on the BEA Dataset, our model attains the highest BLEU score of 44.12, reflecting its ability to model long-term dependencies and capture rare patterns indicative of anomalies. The improvement in METEOR and ROUGE-L metrics further supports the effectiveness of our feature learning strategy, which integrates self-supervised learning and domain adaptation techniques to enhance generalization across diverse datasets. The reduction in TER across all datasets indicates that our approach produces fewer incorrect detections compared to previous methods. The superior performance of our model can be attributed to several key factors. First, our method incorporates a hybrid feature extraction framework that combines convolutional and transformer-based architectures to effectively capture spatial and temporal dependencies. Second, the model benefits from a self-supervised pretraining strategy that enhances its ability to learn normal patterns, thereby improving anomaly detection accuracy. Third, our approach employs adaptive thresholding mechanisms that dynamically adjust detection sensitivity based on data distribution, leading to better generalization across datasets. extensive hyperparameter tuning and fine-grained feature selection contribute to the model's robustness in diverse anomaly detection scenarios.

Figure 6
www.frontiersin.org

Figure 6. Comparison of our method with SOTA methods on MLQA dataset and FLoRes-200 dataset.

Figure 7
www.frontiersin.org

Figure 7. Comparison of our method with SOTA methods on OpenSubtitles dataset and BEA dataset.

Our model not only achieves significant improvements over traditional SOTA models in terms of metric scores, but it also offers substantial advantages in real-world policy environments. One of the key strengths of our approach is its ability to dynamically adjust policies in response to changing economic conditions. This adaptability is crucial in practical settings, allowing policymakers to react effectively to shifts in global markets or unforeseen public health crises. By integrating machine translation, our model ensures that social security benefits and healthcare services are accessible to non-native speakers, especially in multilingual societies. This enhances equity in the distribution of resources and contributes to improved public health outcomes, particularly for marginalized groups who may otherwise face barriers to accessing essential services. Moreover, our model's ability to handle large-scale datasets makes it well-suited for a variety of policy scenarios, from local to national levels. The use of machine learning and deep learning techniques allows for faster, more efficient decision-making, reducing the time and computational resources required for policy analysis. In addition to its operational efficiency, our model provides a long-term framework for assessing the sustainability of social security policies. By forecasting the impacts of different interventions on public health outcomes, policymakers can make informed decisions that balance fiscal sustainability with improved public health. These practical advantages demonstrate the real-world applicability of our approach in optimizing social security systems and fostering long-term economic stability and welfare.

4.4 Ablation study

To analyze the contribution of different components in our proposed method, we conduct an ablation study by progressively removing key modules and evaluating the performance on all four datasets. The ablation settings include: w/o State Dynamics and Policy Interaction, where State Dynamics and Policy Interaction is removed; w/o Equilibrium Conditions and Constraints, where Equilibrium Conditions and Constraints is removed; and w/o Optimization via Policy Adjustment, where Optimization via Policy Adjustment is removed. The performance of these ablated models is compared against our complete method, and the results are reported in Tables 5 and 6.

Table 5
www.frontiersin.org

Table 5. Ablation study results on our method across MLQA dataset and FLoRes-200 dataset.

Table 6
www.frontiersin.org

Table 6. Ablation study results on our method across OpenSubtitles dataset and BEA dataset.

In Figure 8, we observe that removing any of the key components leads to a decline in performance across the MLQA Dataset and the FLoRes-200 Dataset. when State Dynamics and Policy Interaction is removed, BLEU drops from 42.85 to 39.12 on UCSD, and from 44.67 to 41.67 on MVTec, indicating its critical role in capturing important features. Similarly, w/o Equilibrium Conditions and Constraints results in a more significant degradation, suggesting that Equilibrium Conditions and Constraints plays a crucial role in improving feature robustness. Optimization via Policy Adjustment also contributes to overall performance, as removing it causes a drop in all metrics, although the effect is slightly less pronounced than the removal of Equilibrium Conditions and Constraints. A similar trend is observed in Figure 9 for the OpenSubtitles Dataset and BEA Dataset. The removal of State Dynamics and Policy Interaction results in a noticeable performance decline, with BLEU decreasing from 43.41 to 40.21 on KDD and from 44.12 to 42.12 on BEA. The effect of removing Equilibrium Conditions and Constraints is even more significant, particularly in METEOR and ROUGE-L scores, reinforcing the importance of this module in network intrusion detection and time-series anomaly detection tasks. Optimization via Policy Adjustment also contributes to the final results, as demonstrated by the decrease in BLEU and METEOR when removed.

Figure 8
www.frontiersin.org

Figure 8. Ablation study results on our method across MLQA dataset and FLoRes-200 dataset.

Figure 9
www.frontiersin.org

Figure 9. Ablation study results on our method across OpenSubtitles dataset and BEA dataset.

5 Conclusions and future work

In this study, we investigate the intricate relationship between social security systems and public health outcomes from an economic perspective. Traditional economic models often fail to capture the dynamic interplay between social security mechanisms and health indicators, which limits the ability of policymakers to develop effective interventions. To address this gap, we introduce an advanced economic policy modeling framework that integrates dynamic optimization and statistical learning methodologies. Our model conceptualizes social security as a dynamic system, incorporating key macroeconomic variables such as government expenditures, health benefits, and labor market conditions. Through the application of dynamic equilibrium modeling and empirical validation, we demonstrate that well-optimized social security policies can lead to significant improvements in public health outcomes while maintaining economic stability. The experimental results confirm the efficacy of our approach, showing that our model provides more precise assessments of policy interventions under diverse economic conditions. These findings offer valuable insights for policymakers aiming to enhance public health through strategic economic policies.

While the proposed framework demonstrates strong performance and practical relevance in optimizing social security systems and improving public health outcomes, several limitations remain. First, the reliance on historical administrative and economic data may introduce biases into the model, particularly if the datasets are incomplete or unrepresentative of certain populations. This could affect the generalizability of policy recommendations, especially in rapidly changing or resource-constrained environments. In addition, the implementation of AI-driven decision-making in public health and social policy raises important ethical concerns. Issues of fairness may emerge if the model unintentionally reinforces existing inequalities, such as under-serving marginalized linguistic or socioeconomic groups. Moreover, the use of complex algorithms–particularly deep neural networks–can reduce transparency and make it difficult for policymakers and the public to understand the basis for certain decisions. This “black box” nature poses challenges for accountability and public trust. Addressing these concerns requires careful attention to model interpretability, bias mitigation, and the development of regulatory and ethical oversight mechanisms when deploying AI-based systems in real-world governance contexts.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

SN: Writing – original draft, Writing – review & editing, Methodology, Supervision, Project administration, Validation, Resources, Visualization. WL: Writing – original draft, Writing – review & editing, Data curation, Conceptualization, Formal analysis, Investigation, Funding acquisition, Software. LL: Writing – review & editing, Visualization, Supervision, Funding acquisition.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This work was supported by Project fifteen of Zhejiang Education Department: the construction and implementation of the personnel training model of “Industry, education and innovation” for Finance and economics majors based on the OBE concept, Project number:jg20230407 and Zhejiang Province Department of Education Cooperation Project: based on the “Four-chain coordination” of higher vocational colleges creative integration of collaborative education model construction and practice, Project number: 428.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Gen AI was used in the creation of this manuscript.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Xu H, Sharaf A, Chen Y, Tan W, Shen L, Durme BV, et al. Contrastive preference optimization: pushing the boundaries of LLM performance in machine translation. arXiv [Preprint]. (2024). doi: 10.48550/arXiv.2401.08417

Crossref Full Text | Google Scholar

2. Liu Y, Gu J, Goyal N, Li X, Edunov S, Ghazvininejad M, et al. Multilingual denoising pre-training for neural machine translation. Trans. Assoc. Comput. Linguist. (2020) 8:726–42. doi: 10.1162/tacl_a_00343

Crossref Full Text | Google Scholar

3. Zhang B, Haddow B, Birch A. Prompting large language model for machine translation: a case study. In: Proceedings of the 40th International Conference on Machine Learning. Vol. 202. (2023). p. 41092–110. Available online at: https://proceedings.mlr.press/v202/zhang23m.html

Google Scholar

4. Peng K, Ding L, Zhong Q, Shen L, Liu X, Zhang M, et al. Towards making the most of ChatGPT for machine translation. arXiv [Preprint]. (2023). doi: 10.48550/arXiv.2401.05176

Crossref Full Text | Google Scholar

5. Zhu W, Liu H, Dong Q, Xu J, Kong L, Chen J, et al. Multilingual machine translation with large language models: empirical results and analysis. In: NAACL-HLT. Mexico City, Mexico: Association for Computational Linguistics (2023).

Google Scholar

6. Kocmi T, Bawden R, Bojar O, Dvorkovich A, Federmann C, Fishel M, et al. Findings of the 2023 conference on machine translation (WMT23). In: Proceedings of the Seventh Conference on Machine Translation (WMT). Abu Dhabi: Association for Computational Linguistics (2023). p. 1–45. Available online at: https://aclanthology.org/2022.wmt-1.1/

Google Scholar

7. Moslem Y, Haque R, Way A. Adaptive machine translation with large language models. In: European Association for Machine Translation Conferences/Workshops. Tampere: European Association for Machine Translation (2023).

Google Scholar

8. Goyal N, Gao C, Chaudhary V, Chen PJ, Wenzek G, Ju D, et al. The flores-101 evaluation benchmark for low-resource and multilingual machine translation. In: Transactions of the Association for Computational Linguistics. Cambridge, MA: MIT Press (2021).

Google Scholar

9. Freitag M, Foster GF, Grangier D, Ratnakar V, Tan Q, Macherey W. Experts, errors, and context: a large-scale study of human evaluation for machine translation. In: Transactions of the Association for Computational Linguistics. Cambridge, MA: MIT Press (2021).

Google Scholar

10. García X, Bansal Y, Cherry C, Foster GF, Krikun M, Feng F, et al. The unreasonable effectiveness of few-shot learning for machine translation. In: Proceedings of the 40th International Conference on Machine Learning. Vol. 202. (2023). p. 10867–78. Available online at: https://proceedings.mlr.press/v202/garcia23a.html

Google Scholar

11. Jiang N, Lutellier T, Tan L. CURE: code-aware neural machine translation for automatic program repair. In: 2021 IEEE/ACM 43rd International Conference on Software Engineering (ICSE), Madrid (2021). p. 1161–73. doi: 10.1109/ICSE43902.2021.00107

Crossref Full Text | Google Scholar

12. Fan A, Bhosale S, Schwenk H, Ma Z, El-Kishky A, Goyal S, et al. Beyond English-centric multilingual machine translation. J Mach Learn Res. (2021) 22:1–48. Available online at: http://www.jmlr.org/papers/v22/20-1307.html

Google Scholar

13. Kocmi T, Bawden R, Bojar O, Dvorkovich A, Federmann C, Fishel M, et al. Findings of the 2022 conference on machine translation (WMT22). In: Conference on Machine Translation. Abu Dhabi: Association for Computational Linguistics (2022).

Google Scholar

14. Agrawal S, Zhou C, Lewis M, Zettlemoyer L, Ghazvininejad M. In-context examples selection for machine translation. In: Annual Meeting of the Association for Computational Linguistics. Albuquerque: Association for Computational Linguistics (2022).

Google Scholar

15. Zhu J, Xia Y, Wu L, He D, Qin T, Zhou W, et al. Incorporating BERT into neural machine translation. arXiv [Preprint]. (2020). doi: 10.48550/arXiv.2002.06823

Crossref Full Text | Google Scholar

16. Li M, Huang PYB, Chang X, Hu J, Yang Y, Hauptmann A. Video pivoting unsupervised multi-modal machine translation. IEEE Trans Pattern Anal Mach Intell. (2022) 45:3918–32. doi: 10.1109/TPAMI.2022.3181116

PubMed Abstract | Crossref Full Text | Google Scholar

17. Xiao Y, Wu L, Guo J, Li J, Zhang M, Qin T, et al. A survey on non-autoregressive generation for neural machine translation and beyond. IEEE Trans Pattern Anal Mach Intell. (2022) 45:11407–27. doi: 10.1109/TPAMI.2023.3277122

PubMed Abstract | Crossref Full Text | Google Scholar

18. Arenas AG, Toral A. Creativity in translation: machine translation as a constraint for literary texts. Transl Spaces. (2022) 11:184–212. doi: 10.1075/ts.21025.gue

Crossref Full Text | Google Scholar

19. Khandelwal U, Fan A, Jurafsky D, Zettlemoyer L, Lewis M. Nearest Neighbor Machine Translation. In: International Conference on Learning Representations. (2020).

Google Scholar

20. Pan X, Wang M, Wu L, Li L. Contrastive learning for many-to-many multilingual neural machine translation. arXiv [Preprint]. (2021). doi: 10.48550/arXiv.2105.09501

Crossref Full Text | Google Scholar

21. Li D, Xiao H, Ma S, Zhang J. Health benefits of air quality improvement: empirical research based on medical insurance reimbursement data. Front Public Health. (2022) 10:855457. doi: 10.3389/fpubh.2022.855457

PubMed Abstract | Crossref Full Text | Google Scholar

22. Akhbardeh F, Arkhangorodsky A, Biesialska M, Bojar O, Chatterjee R, Chaudhary V, et al. Findings of the 2021 conference on machine translation (WMT21). In: Conference on Machine Translation. (2021). Available online at: https://cris.fbk.eu/handle/11582/330742

Google Scholar

23. Kocmi T, Federmann C, Grundkiewicz R, Junczys-Dowmunt M, Matsushita H, Menezes A. To ship or not to ship: an extensive evaluation of automatic metrics for machine translation. arXiv [Preprint]. (2021). doi: 10.48550/arXiv.2107.10821

Crossref Full Text | Google Scholar

24. Li D, Gao M, Hou W, Song M, Chen J, A. modified and improved method to measure economy-wide carbon rebound effects based on the PDA-MMI approach. Energy Policy. (2020) 147:111862. doi: 10.1016/j.enpol.2020.111862

Crossref Full Text | Google Scholar

25. Savoldi B, Gaido M, Bentivogli L, Negri M, Turchi M. Gender Bias in Machine Translation. Transactions of the Association for Computational Linguistics. Cambridge, MA: MIT Press (2021). doi: 10.1162/tacl_a_00401

Crossref Full Text | Google Scholar

26. Raunak V, Menezes A, Junczys-Dowmunt M. The curious case of hallucinations in neural machine translation. arXiv [Preprint]. (2021). doi: 10.48550/arXiv.2104.06683

Crossref Full Text | Google Scholar

27. Ranathunga S, Lee E, Skenduli M, Shekhar R, Alam M, Kaur R. Neural machine translation for low-resource languages: a survey. In: ACM Computing Surveys. (2021).

Google Scholar

28. Haddow B, Bawden R, Barone AVM, Helcl J, Birch A. Survey of low-resource machine translation. In: Computational Linguistics. Cambridge, MA: MIT Press (2021).

Google Scholar

29. Zheng X, Zhang Z, Guo J, Huang S, Chen B, Luo W, et al. Adaptive nearest neighbor machine translation. arXiv [Preprint]. (2021). doi: 10.48550/arXiv.2105.13022

Crossref Full Text | Google Scholar

30. Li D, Xiao H, Ding J, Ma S. Impact of performance contest on local transformation and development in China: empirical study of the National Civilized City program. Growth Change. (2022) 53:559–92. doi: 10.1111/grow.12598

Crossref Full Text | Google Scholar

31. Li Y, Lin D, Gong X, Fu D, Zhao L, Chen W, et al. Inter-relationships of depression and anxiety symptoms among widowed and non-widowed older adults: findings from the Chinese Longitudinal Healthy Longevity Survey (CLHLS) based on network analysis and propensity score matching. Front Public Health. (2025) 13:1495284. doi: 10.3389/fpubh.2025.1495284

PubMed Abstract | Crossref Full Text | Google Scholar

32. Cai D, Wang Y, Li H, Lam W, Liu L. Neural machine translation with monolingual translation memory. arXiv [Preprint]. (2021). doi: 10.48550/arXiv.2105.11269

Crossref Full Text | Google Scholar

33. Qian L, Zhou H, Bao Y, Wang M, Qiu L, Zhang W, et al. Glancing transformer for non-autoregressive neural machine translation. arXiv [Preprint]. (2020). doi: 10.48550/arXiv.2008.07905

Crossref Full Text | Google Scholar

34. Phillimore J, Fu L, Jones L, Lessard-Phillips L, Tatem B. “They just left me”: People seeking asylum, mental and physical health, and structural violence in the UK's institutional accommodation. Front Public Health. (2025) 13:1454548. doi: 10.3389/fpubh.2025.1454548

PubMed Abstract | Crossref Full Text | Google Scholar

35. Rivera-Trigueros I. Machine translation systems and quality assessment: a systematic review. Lang Resour Eval. (2021) 56:593–619. doi: 10.1007/s10579-021-09537-5

Crossref Full Text | Google Scholar

36. Puiu S, Udristioiu MT, Zăvăleanu M. A multivariate analysis of the impact of knowledge and relationships on perceptions about aging among generation Z-a starting point for public health strategies. Front Public Health. (2025) 13:1522078. doi: 10.3389/fpubh.2025.1522078

PubMed Abstract | Crossref Full Text | Google Scholar

37. AlMahmoud RH, Alian M. The effect of clustering algorithms on question answering. Expert Syst Appl. (2024) 243:122959. doi: 10.1016/j.eswa.2023.122959

Crossref Full Text | Google Scholar

38. Hatvani P, Yang ZG. Training embedding models for Hungarian. In: 2024 IEEE 3rd Conference on Information Technology and Data Science (CITDS). Debrecen (2024). p. 1–6. doi: 10.1109/CITDS62610.2024.10791389

Crossref Full Text | Google Scholar

39. Arnob NMK, Faiyaz A, Fuad MM, Al Masud SMR, Das B, Mridha M. IndicDialogue: a dataset of subtitles in 10 Indic languages for Indic language modeling. Data in Brief . (2024) 55:110690. doi: 10.1016/j.dib.2024.110690

PubMed Abstract | Crossref Full Text | Google Scholar

40. He L. Self-attention convolutional long short term memory for intelligent detection system in English translation errors. In: 2024 International Conference on Intelligent Algorithms for Computational Intelligence Systems (IACIS). Hassan: IEEE (2024). p. 1–4.

Google Scholar

41. Pu Q, Xi Z, Yin S, Zhao Z, Zhao L. Advantages of transformer and its application for medical image segmentation: a survey. Biomed Eng Online. (2024) 23:14. doi: 10.1186/s12938-024-01212-4

PubMed Abstract | Crossref Full Text | Google Scholar

42. Ali NF, Albastaki N, Belkacem AN, Elfadel IM, Atef M. A low-complexity combined encoder-LSTM-attention networks for EEG-based depression detection. IEEE Access. (2024) 12:129390–403 doi: 10.1109/ACCESS.2024.3436895

Crossref Full Text | Google Scholar

43. Yang J, Lv X, Xu X, Chen X. Heterogeneous treatment effects of digital transformation on firms' innovation: based on BART method. Pacific-Basin Finance J. (2024) 85:102372. doi: 10.1016/j.pacfin.2024.102372

Crossref Full Text | Google Scholar

44. Piau M, Lotufo R, Nogueira R. ptt5-v2: A closer look at continued pretraining of t5 models for the Portuguese language. In: Brazilian Conference on Intelligent Systems. Cham: Springer (2024). p. 324–338.

Google Scholar

45. Cao K, Cheng W, Hao Y, Gan Y, Gao R, Zhu J, et al. DMSeqNet-mBART: a state-of-the-art adaptive-DropMessage enhanced mBART architecture for superior Chinese short news text summarization. Expert Syst Appl. (2024) 257:125095. doi: 10.1016/j.eswa.2024.125095

Crossref Full Text | Google Scholar

46. Smits-van der Nat M, van der Wilt F, Meeter M, van der Veen C. The value of pretend play for social competence in early childhood: a meta-analysis. Educ Psychol Rev. (2024) 36:46. doi: 10.1007/s10648-024-09884-z

Crossref Full Text | Google Scholar

Keywords: social security systems, public health, economic policy modeling, dynamic optimization, statistical learning

Citation: Niu S, Li W and Li-Li (2025) The impact of social security systems on public health outcomes: an economic perspective on machine translation applications. Front. Public Health 13:1597381. doi: 10.3389/fpubh.2025.1597381

Received: 21 March 2025; Accepted: 03 June 2025;
Published: 10 July 2025.

Edited by:

Ding Li, Southwestern University of Finance and Economics, China

Reviewed by:

Mohammed Habes, Yarmouk University, Jordan
Cintya Lanchimba, Escuela Politécnica Nacional, Ecuador

Copyright © 2025 Niu, Li and Li-Li. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Weihua Li, ZGVsbGlsaWN1cDd4QG91dGxvb2suY29t

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.