- School of Informatics, Yunnan Normal University, Kunming, China
In recent years, the integration of artificial intelligence (AI) with drug discovery has become a promising frontier in biomedical research. However, as artificial intelligence systems become increasingly complex, ensuring their alignment with human values and goals becomes essential. Specifically, combining artificial intelligence systems with human values is crucial for reducing potential risks in the field of drug discovery and maximizing social benefits. This article explores the concepts and challenges related to alignment with artificial intelligence in the context of drug discovery, emphasizing on human-centered approaches to AI development and deployment. We further investigated popular technology frameworks designed for human-centered AI alignment, aimed at improving the robustness and interpretability of AI models. We provide some insights into the challenges of human-centered AI alignment, which represents a significant advancement in addressing robustness and interpretability, thus taking a step forward in the field of AI alignment research. Finally, we discuss strategies for systematically integrating human values into AI-driven drug discovery systems. This article aims to emphasize the importance of AI alignment as a foundational principle in the field of drug discovery and advocate the perspective that “AI alignment is all your need for future drug discovery”.
1 Introduction
The annals of human civilization are replete with narratives of endeavors to combat afflictions and calamities. Essential to this effort are pharmaceuticals, which constitute the principal ways for mitigating diseases, having been progressively refined through human empirical inquiry and application. Nonetheless, conventional drug development methods still face significant challenges, including high costs, lengthy timelines, and meager success rates. In fact, the diffcult road from drug conceptualization to market outcomes typically spans ten years, with expenditures approaching 2.6 billion dollars. In addition, the success rate of candidate drugs transitioning from trial phases to market availability scarcely exceeds 10% (Avorn, 2015; DiMasi et al., 2016). Hence, artificial intelligence has significantly shortened the drug development trajectory, provided novel avenues for the advancement of drug discovery (Wu et al., 2018; Xia et al., 2023; Li et al., 2022a, 2021b, 2023).
However, with the continuous improvement of artificial intelligence (AI) system capabilities, related risks are also increasing (Ji et al., 2024). Emerging trends, such as the proliferation of intelligent agents based on large language models and the development of generative AI technology, have the potential to achieve a form of universal artificial intelligence where systems may attain or even surpass human-level intelligence in specific domains. Although these advancements foreshadow potential benefits such as automation, increased efficiency, and accelerated technological progress, they also bring significant risks, including security vulnerabilities, biases, and societal inequalities. Moreover, there are concerns about the potential impact of large-scale deployment of superhuman artificial intelligence systems. Notably, contemporary large language models exhibit significant biases in terms of gender, sexual identity, and immigration status, which exacerbates pre-existing social inequities. Furthermore, the negative behaviors observed in these models, such as the propagation of inaccurate responses, flattery, and deceit, tend to escalate with the amplification of model size, which raises ethical considerations regarding the deployment of advanced artificial intelligence systems. Concurrently, the development of generative language models has sparked discussions around the control and governance mechanisms, which is required for effective management such systems (Arora, 2024).
In the domain of drug discovery (Urbina et al., 2022), if the output provided by large language models is proven to be erroneous or fallacious, it has the potential to misguide researchers and trigger erroneous determinations throughout the entire drug discovery process, which, in turn, may lead to resource waste, deviation from the intended research trajectory, and the inadvertent introduction of unsafe or inefficacious drugs into the market. Large language models may be susceptible to manipulation or exploitation as a channel for the promotion of specific drugs or therapeutic modalities, regardless of their empirical foundation or clinical efficacy. Such instances of flattery and deceit have the propensity to compromise patient welfare, which will lead to the adoption of superfluous or ineffectual treatment methods. The sensitivity of large language models to data-driven biases can exacerbate bias in research outcomes and drug recommendations, which could exacerbate existing healthcare inequalities. In scenarios where large language models provide illegal or morally questionable recommendations, such as the promotion of prohibited substances or improper drug utilization, which will arise legal culpability and moral quandaries, thereby questioning the integrity and credibility of the healthcare industry. Hence, the harmful behaviors exhibited by large language models indicate deleterious impacts on research, clinical practice, and patient welfare (Urbina et al., 2022; Vijayan et al., 2022).
In the context of AI-driven drug discovery, human values refer to fundamental principles such as fairness, transparency, accountability, and respect for human well-being, which are key guiding principles. Embedding these values into artificial intelligence systems is not only a moral requirement, which is also a practical necessary condition to ensure that drug discovery results are trustworthy, beneficial to society, and meet the needs of patients. For example, fairness is crucial in preventing bias in data analysis and candidate prioritization. Transparency improves the interpretability of predictive models. This multidimensional perspective emphasizes that combining artificial intelligence systems with human values can greatly impact the credibility and acceptance of society. Han et al. (2022) further emphasized the need for a deep understanding of the consistency between artificial intelligence and human values, providing important insights on how to achieve this consistency in the biomedical field. Therefore, the integration of human-centered artificial intelligence (HCAI) into drug discovery offers several potential benefits (Mbatha et al., 2023). By aligning AI systems with human values, there is a greater focus on ensuring that drug candidates identified by AI are safe and effective for human use, which could reduce adverse effects and improve the success rate of clinical trials (Wang et al., 2023). Human-centered AI can help ensure that drug discovery efforts are inclusive and meet the needs of different populations (Zheng et al., 2023). Integrating human values into AI systems can ensure that ethical considerations are central to the drug discovery process. By prioritizing human-centered approaches, AI systems can mitigate potential biases in the data or algorithms used in drug discovery, which leads to more fair and unbiased decision-making processes, ultimately improving the fairness and reliability of drug discovery efforts. Human-centered AI alignment can simplify the drug discovery process by prioritizing drug candidate drugs that are most likely to meet the needs and preferences of patients and clinical doctors, which can lead to faster development timelines and more efficient resources allocation in drug discovery research.
The endeavor to achieve AI alignment depict the key basic trajectory toward the attainment of human-centered AI. AI alignment (Ji et al., 2024) is predicated on the imperative to engender AI systems that align their behavior with human intentions and values. This pursuit includes four overall objectives, namely Robustness, Interpretability, Controllability and Ethicality (RICE) (Ji et al., 2024). AI alignment has the potential to fundamentally change the drug discovery process by improving efficiency, safety, ethical considerations, and fairness. By aligning AI systems with human intentions and values, researchers can harness the transformative power of AI to address unmet medical needs and improve patient outcomes in a responsible and equitable manner. Ji et al. believe that as the capabilities of artificial intelligence systems continue to increase, the risk of alignment failure is also increasing. Mitigating the extinction risk brought by artificial intelligence should become an important global priority. They first proposed four key objectives for AI alignment, namely RICE, and divided AI alignment into two key components: forward alignment and backward alignment (Ji et al., 2024). The difference in this article's approach is that we provide a review of the alignment problem of artificial intelligence in the field of drug discovery, with a focus on robust and interpretable technical architectures.
To push the boundaries of drug discovery into human-centered AI alignment as shaping a positive future of drug discovery, we argue that aligning AI systems with human values is essential to mitigate potential risks and maximize societal benefits in the domain of drug discovery. In this paper, initially, we scrutinized the challenges inherent in AI alignment. Subsequently, we survey popular technical frameworks designed for human-centered AI alignment and give some insights to aim at enhancing the robustness and interpretability of artificial intelligence models, that are the two most critical objectives of AI alignment. Finally, we propose relevant strategies to incorporate human values into AI systems for drug discovery. In summary, the key contributions of our work are as follows:
• To the best of our knowledge, this work represents the first comprehensive survey of technical frameworks specifically designed for human-centered AI alignment, with a focus on AI technologies in the field of drug discovery.
• We give some insights into the challenge of AI alignment, aimed at enhancing the robustness and interpretability of artificial intelligence models based on molecule-related downstream tasks, which describes our preliminary findings concerning AI alignment.
• We propose relevant strategies to incorporate human values into the design and implementation of AI systems for drug discovery.
2 Challenges in AI alignment
Foundation models, such as GPT-3 (Brown et al., 2020) and BERT (Devlin et al., 2019), can advance the development of science by augmenting research methods with foundation models and generative AI. If explicit guidelines are formulated for the utilization of foundation models and generative AI, the potential benefits they offer to science and scholarly inquiry outweigh the associated risks (Rossi et al., 2024). The endeavor to construct artificial intelligence systems that harmonize with human values and intentions poses a significant challenge. Presently, there lacks a universally accepted standard for gauging alignment. Leike et al. (2018) have described the intelligent agent alignment problem and expanded its scope to encompass super artificial intelligence systems (OpenAI, 2023b). Meanwhile, the RICE principle proposed by Ji et al. (2024) centers on discerning and accommodating human intentions, which points the four cardinal objectives of AI alignment: Robustness, Interpretability, Controllability and Ethicality. These four basic objectives form a unified whole, referred to as the RICE principles, which present distinct challenges in the context of AI alignment.
• Robustness: The robustness of AI systems represents their capacity to maintain stability and dependability amid diverse uncertainties, disruptions, or adversarial attacks (Dietterich, 2017). In essence, a robust AI system excels at maintaining proficient performance across varied environments and situations, while exhibiting robust adaptability to alterations or perturbations in input data. A robust AI system exhibits minimal performance degradation or susceptibility to failure when confronted with variations in external conditions. Such robustness constitutes a pivotal attribute requisite for ensuring the dependable operation of AI systems in real-world settings and for effectively addressing various challenges. To enhance the robustness of AI systems, researchers typically employ a variety of methods, including but not limited to augmenting the diversity of training data, implementing data augmentation techniques, improving algorithms and model architectures for enhanced resilience, and designing tailored testing and evaluation protocols. By iteratively enhancing the robustness of the system, it becomes better equipped to navigate complex real-world scenarios, thereby enhancing the efficacy and reliability of AI systems in different application domains. Aligned systems always consistently maintain robustness throughout its entire lifecycle (Russell, 2019).
• Interpretability: The interpretability of AI systems involves their ability to provide clear and transparent explanations or reasoning, thereby facilitating user comprehension of the system's operational framework, decision-making mechanism, and underlying principles for recommendations. A well interpretable AI system can explain its behavior and provide reasonable reasons for its decisions, rather than merely providing results. In many application scenarios, people not only need accurate predictions or decisions from AI systems but also need to understand the mechanisms that support these results and the basic rationale that drive such decisions. Interpretability is crucial in cultivating user's confidence in the system, augmenting its credibility, and aiding in the identification of potential biases or discrepancies therein. Furthermore, the interpretability of AI systems helps clarify the decision-making logic for users and facilitates system developers to identify and correct potential defects, thereby enhancing the robustness and reliability of the system. Therefore, interpretability has emerged as a pivotal research approach within the realm of artificial intelligence, attracting widespread attention and applicability in different practical areas (Lipton, 2016; Guidotti et al., 2018; Doshi-Velez and Kim, 2017).
• Controllability: The controllability of AI systems refers to the ability of humans to effectively manage and control the behavior and decision-making of the system, ensuring that the system's actions and decision-making processes are always supervised and constrained by humans (Soares et al., 2015; Hadfield-Menell et al., 2017). Controllability focuses more on real-time human intervention and adjustment of the system to ensure that it meets human expectations and needs during operation, while avoiding adverse or unexpected consequences. With the increasing development of AI technology, more and more research has expressed concerns about the controllability of these powerful systems (Critch and Krueger, 2020). So, the controllability of AI systems encompasses the capacity to fulfill human expectations and objectives while mitigating potential risks and uncertainties (Bowman et al., 2022).
• Ethicality: The ethical behavior of AI systems is related to their capacity to comply with ethical norms and values during the design, development, deployment, and utilization phases, which includes ensuring that the behavior and decisions of the system comply with ethical norms, upholding individual rights and dignity, and exhibiting accountability toward societal and public interests (Floridi and Sanders, 2004; Jobin et al., 2019). The ethical behavior of AI systems involves diverse aspects, including safeguarding privacy, fostering fairness, promoting transparency, and embracing accountability. Given the profound impact that AI systems may exert on individuals, society, and the environment, the consideration of ethical imperatives is increasingly deemed imperative (Mittelstadt et al., 2016). Neglecting moral considerations during the development and deployment of AI systems may lead to adverse effects and social challenges (Taddeo and Floridi, 2018). Thus, for AI systems, their ethical behavior is considered a key determinant in maintaining their positive societal impact, avoiding harm and unfairness.
3 Implementation strategies to AI alignment
3.1 Mainstream methods of AI alignment for drug discovery
The key technologies utilized in the AI alignment process include reinforcement learning with human feedback (RLHF) (Knox and Stone, 2009), out-of-distribution (OOD) generalization (Ji et al., 2022; Zhuang et al., 2023; Eissa et al., 2024; Tossou et al., 2024) and OOD detection (Wang et al., 2025; Shen et al., 2024; He et al., 2024; Theunissen et al., 2025; Liu et al., 2025) techniques and visualization methods (Li et al., 2021b,a). RLHF is a subfield of reinforcement learning (RL) that incorporates human feedback into the learning process to guide and improve the performance of RL agents, whose goal is to train agents to perform specific tasks, and humans typically provide evaluation feedback or guidance to agents in the form of reward signals, criticisms, preferences, or demonstrations. This kind of human feedback helps RL agents learn more efficiently and effectively, especially in challenging domains where designing reward functions is crucial. OOD techniques (Eissa et al., 2024; Li et al., 2025; Antonluk et al., 2025) are designed to tackle the ubiquitous problem of distribution shift, wherein the distribution observed during training is different from that encountered during testing. OOD generalization refers to the ability of a model to maintain reasonable predictive ability during the testing phase for OOD data, which is new distribution samples unseen during training but are relevant to the task. It aims to learn the basic laws and transferable representations of data, rather than relying solely on the statistical characteristics of training data, in order to have stronger robustness and adaptability in open environments. OOD detection aims to automatically distinguish between input samples from the In Distribution (ID) or OOD during the model inference phase, whose core goal is to prevent the model from making high confidence erroneous predictions on unseen and distributed data, thereby improving the reliability and safety of the model in practical applications. Visualization (Li et al., 2021b, 2023, 2021a) helps to increase the interpretability of the model, which is important in understanding why the model makes such decisions and inferences. Table 1 shows the implementation strategies and mainstream methods of AI alignment for drug discovery, as well as their relationship with RICE principles. In the next section, we will introduce a technical framework and methodology aimed at studying the robustness and interpretability of neural network models, specifically addressing the first two challenges outlined in the RICE framework. This effort aims to demonstrate our conceptual approach and feasible solutions for achieving AI alignment.
Table 1. The implementation strategies and mainstream methods of AI alignment for drug discovery, and the implementation objectives of the RICE principles covered by each strategy and method.
3.2 Technical framework for exploring robustness and interpretability
The fundamental concept that support AI alignment involves directing our attention toward AI technologies as a means to address the four challenges mentioned above. Figure 1 illustrates the design of our framework and methodology for addressing the four challenges in AI alignment. For each specific challenge, we have come up with the most effective approaches to response it. Specifically, in order to improve the reliability of neural network models, we suggest implementing several techniques that can enhance model robustness. Especially, we recommend focusing on AI technologies that can help models identify and detect to “out of distribution data”, meaning that models not only perform well on training data, but also work properly when encountering new samples that are different from the training data. Additionally, we suggest integrating the methods of “residual vector quantization” and “invariant substructure” into AI model. The advantage of doing so is that AI model can identify and capture key features that are consistent exist across different data, and make the model perform better and easier to understand, thereby improving its overall performance and interpretability. When addressing the challenge of interpretability in AI models, we propose a range of visualization techniques. These methods can greatly aid in understanding the inner workings of the models. Specifically, we suggest visualizing the model flowchart and molecular 3D structure, which can provide a clearer understanding of the model's architecture and the data format processed by model. Additionally, visualizing the feature space can help clarify how the model processes and distinguishes various inputs. Furthermore, by visualizing the contribution results of the target task, it is possible to gain insights into which features or components of the input data are most influential in driving the model's predictions. Then, we survey and give some insight into novel AI algorithms to address the challenges of robustness and interpretability in AI alignment for drug discovery. In recent years, graph neural networks have demonstrated impressive performance. Consequently, in this paper, we represent drug molecules using a graph structure. Within this framework, atoms are depicted as nodes, and bonds as edges. Given this representation, we aim to explore the robustness and interpretability of tasks related to molecules in non-Euclidean space.
Figure 1. A technical framework and methodology for addressing the four challenges in AI alignment. (a) The out-of-distribution (OOD) module is tailored for robustness, with the capability to acquire invariance and resilience against distributional shifts. (b) The visualization module is specifically crafted to enhance interpretability, offering functionalities such as 3D structural visualization, feature space visualization, and contribution visualization. (c) The controllable module is engineered to facilitate controllability, affording opportunities for human intervention. (d) The module delineating social values designed for ethicality that adheres to global moral standards and upholds human values.
3.2.1 Suggested approaches for robustness
Neural network models are always built on the assumption of independent and identically distributed (i.i.d) across training and testing data. In the field of drug discovery, when distribution shift occurs, such as molecular scaffolds (Wu et al., 2018), size (Ji et al., 2022), label noise changing on the training and testing sets, or when the i.i.d assumption is not valid, the performance of the model will be poor. As is well known, in the virtual screening process of hit recognition, the prediction model always trains on a determined target. The COVID-19 event resulted in a new target with an unprecedented data distribution, leading to a significant decrease in the performance of the prediction model when applied to this new target. OOD learning in this field focuses on handling scenarios where training and testing data present different distributions, aiming to alleviate performance degradation and improve model robustness (Zhuang et al., 2023; Muandet et al., 2013). In light of the present limitations of unbiased learning in capturing distributional shifts pertaining to both labels and feature spaces, the proposed approach endeavors to examine the interplay between unbiased learning on graphs and OOD detection within a unified latent discrete space. Subsequently, it aims to introduce a method for unbiased distributional representation of graph data and OOD detection guided by environmental variables. Figure 2 shows the overview of suggested unbiased learning and out of distribution detection method.
Figure 2. The suggested unbiased learning and out of distribution detection method. It includes two tasks: how to better represent data during the training phase (unbiased learning) and how to better utilize data during the testing phase (OOD detection).
This method aims to utilize a binary environmental variable e to model both in-distribution (ID) and out-of-distribution data, where e = 1 denotes in-distribution data and e = 0 signifies biased or out-of-distribution data. Define labeled training set and test set , where Ntrain and Ntest represent the number of samples in the training and test sets, respectively. Let be a graph with adjacency matrix and initial node features X = {xv} for , where is nodes set and is edges set. Our objective is to develop a predictive model fη:G → ŷ capable of accurately forecasting labels within . Let as the loss function on sample (G, y) under environment e, then the empirical risk minimization (ERM) (Cherkassky, 1997) can be defined as
where is a constant when e = 0 holds, p denotes the probability distribution and satisfies the following condition: p(G, y|e) = p(G|e)p(y|G, e). Furthermore, to identify OOD samples within the test set, it is imperative to design an OOD detector dect, including predictive model fη, score function score, threshold χ. The detector dect can be defined as:
Through this approach, the model engages in unbiased learning during the training phase and additionally performs out-of-distribution detection during the testing phase of downstream tasks. This strategy effectively mitigates the negative impact of out-of-distribution samples on the model's performance, consequently enhancing its efficacy and robustness. Hence, the proposed method exemplifies our strategic approach and investigational efforts toward bolstering the robustness of the model, constituting a modest stride in the direction of fortifying its resilience. In the field of drug discovery, in terms of OOD generalization, CODI (Eissa et al., 2024) is a contextual OOD integration method designed to generates synthetic data by incorporating unrepresented sources of variation observed in real-world applications into a given molecular fingerprint dataset. By augmenting the dataset with OOD variance, CODI enhances the ability of machine learning models to generalize to samples beyond the original training data, thereby reducing the reliance on extensive experimental data collection. Tossou et al. present a rigorous method (Tossou et al., 2024) to investigate molecular OOD generalization in the field of drug discovery. This method uses covariate changes to quantitatively quantify the distribution changes of sample distance to the training set encountered during actual deployment, which can lead to performance degradation of up to 60% and uncertainty calibration degradation of up to 40%. In terms of OOD detection, PGR-MOOD (Shen et al., 2024) is a molecular OOD detection method by using an auxiliary diffusion model, which compares similarities between input molecules and reconstructed graphs. Due to the generative bias toward reconstructing in-distribution training samples, the similarity scores of OOD molecules will be much lower to facilitate detection. GDDA (He et al., 2024) is a novel two-phase method of graph disentangled diffusion augmentation, aimed to disentangle graph representations into semantic factors and style factors by using a distribution-shift-controlled score-based diffusion model. Theunissen et al. evaluate six OOD detection methods to demonstrate OOD detection performance in both synthetical and real-world application settings (Theunissen et al., 2025), specifically in the context of single-cell transcriptomics annotation. Liu et al. proposed TS-DAR (Liu et al., 2025), a transition state identification method based on dispersion and variational principle regularized neural networks. TS-DAR is a deep learning framework inspired by OOD detection in trustworthy artificial intelligence, aimed at understanding protein conformational changes. Unlike traditional Molecular Dynamics (MD) simulations, TS-DAR leverages deep learning techniques to identify transition states, offering a novel approach to studying dynamic molecular processes.
3.2.2 Suggested approaches for interpretability
It is widely acknowledged that neural networks are often referred to as black box models primarily due to the inherent challenge in intuitively understanding and elucidating their internal operating mechanisms. Neural networks typically comprise a multitude of neurons and layers, learning patterns and features of input data through intricate weight adjustments during training. Given the intricate internal architecture of neural networks, elucidating the precise relationship between individual neurons and weights proves challenging, thereby impeding the intuitive explanation of neural network decision-making processes. Despite neural networks' capability to undergo training on extensive datasets and yield highly accurate predictions, their decision-making processes often remain opaque and defy straightforward explanation. In the realm of drug discovery, the interpretability of models assumes paramount importance, as it fosters patient trust in medical diagnoses and facilitates the provision of comprehensive disease treatment by healthcare professionals.
Drug molecules are always represented as graph structures. Rotation invariance-based 3DMol-Net (Li et al., 2021b) model demonstrates the interpretability by learning three-dimensional (3D) soft relation and K-nearest neighbors (KNNs) relation in 3D space, subsequently constructing 3D graph Laplacian, then building rotation-invariant map (RIM) with attention mechanism. This process can be expressed as follows:
where and Dsr denote the soft relationship matrix and its degree matrix, respectively. Similarly, and Dknn denote the KNN relationship matrix in 3D space and its degree matrix, respectively. and and represent Laplacian matrices under KNN relationships and soft relationships, respectively. Finally, the proposed 3D graph Laplacian matrix Lap(3D) can be obtained using Equation 5 with a hyperparameter α. As shown in Figure 3, Lap(3D) is initialized by KNN Laplacian and refined by soft Laplacian . Then define RIM as a function of rotation-invariant map, Vcoord as 3D coordinates vector for each atom. The final rotation invariance-based geometric representation Fri_geom for molecules can be calculated as:
Furthermore, attention attmn can be formulated as:
where m is the target atom, n is the neighbor node, and hm, hn represent the state vectors of atom m and n, respectively. W denotes trainbale weight matrix. LeakeRelu denotes nonlinear activation function.
The 3DMol-Net leverages adaptive graph convolutional networks to proficiently acquire the 3D molecular representations, showcasing commendable efficacy in predicting molecular properties. Moreover, the proposed model exhibits notable interpretability in discerning and explicating the predicted outcomes. As shown in Figure 4, hydrophilic groups exhibit a more pronounced influence on predicting water solubility in the ESOL dataset, with their contribution to molecular features playing a decisive role in the final prediction. Furthermore, the group most pertinent to predicting the activity of inhibiting HIV replication demonstrates a heightened contribution to graph-level representation. Additionally, 3D structural visualization greatly improves the interpretability of AI models within the context of drug discovery. The 3D structural attributes of small molecules and proteins play a crucial role in determining their biochemical functions and activity predictions. These 3D characteristics predominantly dictate both the properties of drugs and the binding characteristics of their respective targets. Therefore, Li et al. (2021a) conducted visualization of 3D voxel representations of molecules within 3D space, with distinct colors denoting various atoms, as shown in Figure 5. This visualization technique affords a more intuitive comprehension of the three-dimensional microstructure of molecules, thereby exerting a discernible influence on the prediction of drug-related interactions, such as drug-drug interaction (DDI) and compound protein interaction (CPI). In addition, iCAN (Weckbecker et al., 2024) aims to covercome the constraints of machine learning models that typically rely on structured and rigid input formats. It encodes the neighborhoods of carbon atoms using a counting array, enhancing the effectiveness of the generated representations for machine learning tasks. By producing interpretable molecular encodings, iCAN method facilitates the comparison of molecular neighborhoods, the detection of recurring patterns, and the visualization of important features through heat maps. MolLoG (Feng et al., 2024) is a molecular deep interpretability method that establishes a bridge between local features and global representations, aiming to enhance the prediction of drug-target interactions. MolLoG comprises local feature encoders (LFE) and global interactive learning (GIL) modules, offering biologically relevant interpretations for the predictions generated by black-box models.
Figure 4. Visualizing the contribution of atoms to molecule-level predictive tasks, that is shown using jet colormap with range [0.25, 1.0]. Red signifies the most significant contribution, blue denotes the least contribution, while yellow and green lie in the intermediate range.
Figure 5. Three represents for the same drug molecule. (a) 2D molecular structure. (b) 3D molecular structure. (c) 3D voxel structure automatically modeled in 3D space using python language.
As for protein 3D representation, since the 3D conformation of a protein dictates its function, the 3D structure of proteins is foundational to understanding their role in biological systems. The specific arrangement of amino acids determines how a protein interacts with other molecules, including substrates and inhibitors. Consequently, insights into protein structure can reveal mechanisms of action and inform the design of new drugs. Furthermore, the stability and dynamics of protein structures are integral to their functionality. Changes in a protein's conformation can significantly affect its activity and interactions. Thus, elucidating these structural nuances is essential for predicting how proteins behave under various conditions. Following GEOM-CVAE (Li et al., 2024), the 3D mesh structure of protein surfaces can be visualized in Figure 6, which encompasses abundant geometric information crucial for deriving an effective protein characterization. The different colors in the mesh represent different surface features. The process of mesh simplification is illustrated from right to left. The simplification of protein 3D mesh can also be regarded as the graph sampling and graph pooling in graph neural network. In tasks associated with AI-Generated Content (AIGC), such as drug design and the prediction of compound-protein binding pockets, this visualization method and underlying representation techniques demonstrate commendable robustness and interpretability.
Figure 6. Visualization of the 3D mesh structure of protein surfaces. The first line depicts the original folding of the protein 5EOY, visualized using PyMOL. The second line illustrates the modeling of the 5EOY surface utilizing three-dimensional (3D) mesh, which encompasses abundant geometric information crucial for deriving an effective protein characterization. From left to right in the second line, the mesh resolution progressively increases, accompanied by an increasing number of vertices in the 3D mesh. The different colors in the mesh represent different surface features. In AIGC-related tasks, such as drug design and the prediction of compound-protein binding pockets, this visualization method and underlying representation techniques demonstrate commendable robustness and interpretability.
3.3 Strategies for incorporating human values
AI systems have penetrated into various aspects of our lives and careers, that are bringing us numerous conveniences. It is worth noting that advanced AI models, such as AIGC large language systems, possess the ability to independently decompose complex tasks into manageable subtasks and execute decisions without human intervention. The emergence of AIGC has brought substantial advantages, especially in improving productivity, tailoring services according to personal preferences, cultivating creativity and breakthroughs, and promoting industrial progress. However, on the other hand, these are countless inherent potential risks associated with AI systems. Empirical research emphasizes the potential capacity of AI systems to pose a threat to global security (Turchin and Denkenberger, 2020; Ji et al., 2024). For instance, the initial iterations of the GPT-4 model, as identified by OpenAI, exhibited a series of dangerous behaviors, including spreading misinformation, manipulating public emotions, and even formulating novel biochemical substances (OpenAI, 2023a). Furthermore, investigations by Urbina et al. (2022) emphasize the potential health risks posed by AI systems in domains such as drug discovery and synthetic biology. The AIGC models tailored for drug design have generated an astonishing 40,000 toxic molecules, whose synthesis and introduction into the human body could potentially trigger significant disasters.
In addition, artificial intelligence has also raised issues such as employment and economic inequality, privacy breaches and security risks, algorithmic bias and discrimination, and ethical concerns (Taddeo and Floridi, 2018; Jobin et al., 2019; Floridi and Cowls, 2019). In the absence of adequate regulation and governance, artificial intelligence systems have the potential to pose catastrophic risks to humanity, that may even endanger human survival (Yu et al., 2018). We must maintain a profound sense of responsibility toward AI alignment. Here are several strategies for integrating human values:
• Technology Integration: The task of AI developers and researchers is to consider AI alignment issues right from the beginning of AI system design. This requires careful design of artificial intelligence algorithms, models, and training protocols, with a focus on prioritizing security, transparency, and consistency with human values. Here are some examples of how human values can be applied in AI design, which will help to bridge theory with practice. For instance, fairness can be applied to AI-driven drug discovery by incorporating fairness metrics to mitigate bias in molecular property prediction tasks, thereby reducing potential biases generated by imbalanced training data (Chen et al., 2024; Salmi et al., 2024). Transparency may be achieved by employing interpretable graph neural networks or attention mechanisms in candidate drug design, which allow researchers and clinicians to understand how candidate molecules are prioritized (Li et al., 2021b; Fang et al., 2023). Accountability can be strengthened through the integration of bias detection and monitoring protocols during model training and validation, which ensures that systematic errors are identified early and addressed (Antonluk et al., 2025; Liu et al., 2024). Together, these practices demonstrate how to embed human values into the workflow of artificial intelligence, that enables drug discovery systems to be not only technically effective but also ethically aligned and socially responsible.
• Policy Regulation: Governments play a crucial role in formulating and managing the regulations for the development and deployment of artificial intelligence. By developong responsible AI policies, guidelines, ethical frameworks, interdisciplinary collaboration and ethical oversight and accountability, we ensure that artificial intelligence systems uphold human values and adhere to ethical standards. For example, the European Union's AI Act proposes a comprehensive legal framework that classifies AI applications based on risk and mandates requirements such as transparency, human oversight, and accountability for high-risk systems, which include applications in healthcare (European Commission, 2021). Similarly, the U.S. Food and Drug Administration (FDA) has released guidelines on AI/ML-based Software as a Medical Device (SaMD), which emphasize continuous learning, performance monitoring, and transparency to ensure patient safety and ethical compliance (U.S. Food and Drug Administration, 2021). These measures demonstrate how regulatory measures can effectively safeguard the fairness, reliability, and societal trust in the application of AI for drug discovery.
• Public Engagement: Collaboration on a global scale is crucial for raising public awareness about the challenges and impacts of AI alignment. It can facilitate mutual supervision and ensures that different perspectives are taken into account throughout the alignment process by encouraging public participation in discussions on artificial intelligence ethics. Global cooperation and public consultation are crucial for ensuring AI alignment in reflecting different perspectives and social values. The OECD Principles on Artificial Intelligence and other international initiatives provide examples of how multi stakeholder governance can promote responsible AI practices (OECD Legal Instruments, 2019). Similarly, the European Union's Artificial Intelligence Act explains how transparent participation mechanisms can incorporate public opinion into policy frameworks (European Commission, 2021). The participatory design approaches further emphasize that involving end-users and affected communities in the early stages of the process can enhance inclusivity and legitimacy (Vines et al., 2013).
4 Global developments in AI alignment for drug discovery
Countries have made certain progress and achievements in aligning AI with the field of drug discovery. The United States has always been at the forefront of drug discovery, with its AI aligned strategies mainly focused on funding research, policy-making, and regulation. For example, the National Institutes of Health (NIH) and the Food and Drug Administration (FDA) have funded many AI projects for drug discovery and developed relevant policy guidance (Uddin et al., 2025; National Institutes of Health (NIH), 2023). In addition, some large pharmaceutical companies in the United States, including Pfizer and Merck, are actively exploring the application of AI in drug discovery (XtalPi, 2025; Merck, 2023). The Chinese government regards healthcare as one of the important development areas and promotes the development of AI in the field of drug discovery through funding research projects, formulating policies, and strengthening international cooperation. Recently, China has proposed the concept of “new quality productive forces”, which are advanced productive forces led by innovation and in line with the new development concept, ultimately achieving harmonious coexistence between humans and nature (The Central People's Government of the People's Republic of China, 2024). The European Commission has funded many AI projects for drug discovery through institutions such as the European Innovation Commission, and has proposed strategic goals to promote pharmaceutical innovation and drug discovery (European Innovation Council, 2022). Other countries, such as Canada, the United Kingdom, Germany, etc., have also conducted some research and practice on AI alignment in the field of drug discovery. Some research institutions and pharmaceutical companies are also actively exploring the application of AI in drug discovery (UK Biobank, 2025). In contrast, the United States emphasizes regulatory leadership and industry adoption, while China emphasizes state driven innovation and integration into broader socio-economic strategies, and the European Union prioritizes cross-border cooperation and ethical governance frameworks. These different methods have demonstrated their respective advantages (Blanco-González et al., 2023). Future research and policies should integrate the effectiveness of these different strategies and promote global cooperation to address ethical and practical challenges, which will enable responsible and influential artificial intelligence to play a role in drug discovery.
5 Conclusion
Drawing on insights from the fields of computer science, and pharmacology, the paper explores the potential benefits of human-centered AI alignment in drug discovery, such as enhanced safety, effectiveness, and accessibility of pharmaceutical interventions. By prioritizing human values and societal well-being, AI-driven drug discovery programs can better meet the needs and preferences of patients, clinicians, and other stakeholders. Subsequently, the paper focuses on the challenges faced by artificial intelligence alignment and measures to address these challenges, incorporating human values into the design and implementation of drug discovery. AI alignment not only emphasizes the integration of ethical principles, stakeholder engagement, and interdisciplinary collaboration throughout the entire AI development lifecycle, but also recommends the use of robustness, transparent and interpretable AI models which incorporate different perspectives in algorithmic decision-making, and establish sustained ethical oversight and accountability mechanisms.
Looking ahead, we should conduct further research on how to achieve deep coupling between artificial intelligence and human values throughout the entire drug discovery process. On the one hand, it is necessary to explore new interdisciplinary collaboration models that integrate ethics, clinical medicine, pharmacology, and artificial intelligence algorithm research more closely to promote the integration and innovation of interdisciplinary knowledge. On the other hand, we should strengthen research on the interpretability and causal inference ability of artificial intelligence models, which not only provide efficient prediction results, but also provide traceable scientific basis for drug mechanism research and clinical decision-making. In addition, we should also focus on studying issues of fairness and universality, ensuring that artificial intelligence systems can maintain stable performance and fair results when facing different populations, disease types, and medical environments. At the practical level, it is recommended to establish a long-term ethical supervision and evaluation framework, which combined with dynamic regulatory mechanisms and open science principles, to ensure the safe, controllable, and transparent development of artificial intelligence technology. Through continuous exploration in the above directions, it is expected to promote the landing and popularization of human-centered artificial intelligence in drug discovery, thereby promoting the sustainable development of the healthcare industry.
In conclusion, we advocates for a paradigm shift toward human-centered AI alignment in drug discovery. Researchers, policy makers, and industry stakeholders should prioritize ethical considerations and societal impact when developing and applying AI systems. Only in this way can the application of artificial intelligence in biomedical research truly serve human well-being. By aligning AI with human values, it can not only play a huge role in promoting the development of healthcare, but also ensure that the benefits brought by technology can be fairly distributed. This can also make the entire process more ethical. All in all, AI alignment is all your need for future drug discovery.
Author contributions
CL: Conceptualization, Methodology, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing, Project administration, Funding acquisition.
Funding
The author declares that financial support was received for the research and/or publication of this article. This work was supported by the National Natural Science Foundation of China (Grant No.62472370 and 62262072) and the Yunnan Provincial Philosophy and Social Science Planning Social Think Tank Project (Grant No.SHZK2024204).
Acknowledgments
We would like to thank the reviewers for their helpful comments.
Conflict of interest
The author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author declares that no Gen AI was used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Antonluk, E., Zaman, S., Ben-Nun, T., Li, P., Diffenderfer, J., Demlrcl, B., et al. (2025). BOOM: Benchmarking out-of-distribution molecular property predictions of machine learning models. arXiv [preprint] arXiv:2505.01912. doi: 10.48550/arXiv.2505.01912
Arora, P. (2024). Creative data justice: a decolonial and indigenous framework to assess creativity and artificial intelligence. Inform. Commun. Soc. 1–17. doi: 10.1080/1369118X.2024.2420041
Avorn, J. (2015). The $2.6 billion pill-methodologic and policy considerations. New Eng. J. Med. 372, 1877–1879. doi: 10.1056/NEJMp1500848
Blanco-González, A., Cabezón, A., Seco-González, A., Conde-Torres, D., Antelo-Riveiro, P., Pineiro, A., et al. (2023). The role of ai in drug discovery: challenges, opportunities, and strategies. Pharmaceuticals 16:891. doi: 10.3390/ph16060891
Bowman, S., Hyun, J., Perez, E., Chen, E., Pettit, C., Heiner, S., et al. (2022). Measuring progress on scalable oversight for large language models. arXiv [preprint] arXiv:2211.03540. doi: 10.48550/arXiv.2211.03540
Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J., Dhariwal, P., et al. (2020). “Language models are few-shot learners,” in Advances in Neural Information Processing Systems, 1877–1901.
Chen, W., Yang, K., Yu, Z., Shi, Y., and Chen, C. L. P. (2024). A survey on imbalanced learning: latest research, applications and future directions. Artif. Intellig. Rev. 57. doi: 10.5772/intechopen.107726
Chen, Y., Zhang, Y., Bian, Y., Yang, H., Ma, K., Xie, B., et al. (2022). “Learning causally invariant representations for out-of-distribution generalization on graphs,” in Proceedings of the 36th International Conference on Neural Information Processing Systems (NeurIPS) Red Hook, NY: Curran Associates Inc.57 Morehouse Lane. 22131–22148.
Cherkassky, V. (1997). The nature of statistical learning theory. IEEE Trans. Neural Netw. 8:1564. doi: 10.1109/TNN.1997.641482
Critch, A., and Krueger, D. (2020). AI research considerations for human existential safety (arches). arXiv [preprint] arXiv:2006.04948. doi: 10.48550/arXiv.2006.04948
Devlin, J., Chang, M., Lee, K., and Toutanova, K. (2019). “BERT: Pre-training of deep bidirectional transformers for language understanding,” in Proceedings of NAACL-HLT, 4171–4186.
Dietterich, T. (2017). Steps toward robust artificial intelligence. AI Magazine 38, 3–24. doi: 10.1609/aimag.v38i3.2756
DiMasi, J., Grabowski, H., and Hansen, R. (2016). Innovation in the pharmaceutical industry: new estimates of R&D costs. J. Health Econ. 47, 20–23. doi: 10.1016/j.jhealeco.2016.01.012
Doshi-Velez, F., and Kim, B. (2017). Towards a rigorous science of interpretable machine learning. arXiv [preprint] arXiv:1702.08608. doi: 10.48550/arXiv.1702.08608
Eissa, T., Huber, M., Obermayer-Pietsch, B., Linkohr, B., Peters, A., Fleischmanna, F., et al. (2024). CODI: enhancing machine learning-based molecular profiling through contextual out-of-distribution integration. PNAS Nexus 3, 449–464. doi: 10.1093/pnasnexus/pgae449
Erman, E., and Furendal, M. (2022). Artificial intelligence and the political legitimacy of global governance. Polit. Stud. 72, 421–441. doi: 10.1177/00323217221126665
European Commission (2021). Proposal for a Regulation Laying Down Harmonised Rules on Artificial Intelligence (Artificial Intelligence Act) Brussels.
Fan, S., Wang, X., Mo, Y., Shi, C., and Tang, J. (2022). “Debiasing graph neural networks via learning disentangled causal substructure,” in Proceedings of the 36th International Conference on Neural Information Processing Systems (NeurIPS) Red Hook, NY: Curran Associates Inc.57 Morehouse Lane.
Fang, Y., Zhang, Q., Zhang, N., Chen, Z., Zhuang, X., Shao, X., et al. (2023). Knowledge graph-enhanced molecular contrastive learning with functional prompt. Nature Mach. Intellig. 5, 542–553. doi: 10.1038/s42256-023-00654-0
Feng, B., Zhang, Y., Zhou, X., Wang, J., and Feng, Y. (2024). MolLoG: A molecular level interpretability model bridging local to global for predicting drug target interactions. J. Chem. Inf. Model. 64, 4348–4358. doi: 10.1021/acs.jcim.4c00171
Floridi, L., and Cowls, J. (2019). A unified framework of five principles for AI in society. Harvard Data Sci. Rev. 1, 2–15. doi: 10.1162/99608f92.8cd550d1
Floridi, L., and Sanders, J. (2004). On the morality of artificial agents. Minds Mach. 14, 349–379. doi: 10.1023/B:MIND.0000035461.63578.9d
Guidotti, R., Monreale, A., Ruggieri, S., Turini, F., Giannotti, F., and Pedreschi, D. (2018). A survey of methods for explaining black box models. ACM Comp. Surv. 51, 1–42. doi: 10.1145/3236009
Hadfield-Menell, D., Dragan, A., Abbeel, P., and Russell, S. (2017). “The off-switch game,” in Workshops at the Thirty-First AAAI Conference on Artificial Intelligence. San Francisco, CA: Association for the Advancement of Artificial Intelligence (AAAI).
Han, S., Kelly, E., Nikou, S., and Svee, E. (2022). Aligning artificial intelligence with human values: reflections from a phenomenological perspective. AI Soc. 37, 1383–1395. doi: 10.1007/s00146-021-01247-4
He, Z., Zhao, C., Shao, M., Lin, Y., Li, D., and Tian, Q. (2024). GDDA: semantic ood detection on graphs under covariate shift via score-based diffusion models. arXiv [preprint] arXiv:2410.17526. doi: 10.48550/arXiv.2410.17526
Hu, W., Liu, B., Gomes, J., Zitnik, M., Liang, P., Pande, V., et al. (2020). “Strategies for pre-training graph neural networks,” in International Conference on Learning Representations (ICLR) Addis Ababa: International Conference on Learning Representations (ICLR).
Ji, J., Qiu, T., Chen, B., Zhang, B., Lou, H., Wang, K., et al. (2024). AI alignment: a comprehensive survey. arXiv [preprint] arXiv.2310.19852. doi: 10.48550/arXiv.2310.19852
Ji, Y., Zhang, L., Wu, J., Wu, B., Huang, L., Xu, T., et al. (2022). DrugOOD: out-of-distribution (OOD) dataset curator and benchmark for ai-aided drug discovery - a focus on affinity prediction problems with noise annotations. arXiv [preprint] arXiv.2201.09637. doi: 10.48550/arXiv.2201.09637
Jobin, A., Ienca, M., and Vayena, E. (2019). The global landscape of AI ethics guidelines. Nat. Mach. Intellig. 1, 389–399. doi: 10.1038/s42256-019-0088-2
Kerry, C., Meltzer, J., and Renda, A. (2021). Strengthening International Cooperation on AI, Progress Report Brookings Institution.
Knox, W., and Stone, P. (2009). “Interactively shaping agents via human reinforcement: the tamer framework,” in Proceedings of the Fifth International Conference on Knowledge Capture (California: Redondo Beach), 9–16.
Leike, J., Krueger, D., Everitt, T., Martic, M., Maini, V., and Legg, S. (2018). Scalable agent alignment via reward modeling: a research direction. arXiv [preprint] arXiv:1811.07871. doi: 10.48550/arXiv.1811.07871
Li, C., Feng, J., Liu, S., and Yao, J. (2022a). A novel molecular representation learning for molecular property prediction with a multiple SMILES-based augmentation. Comput. Intell. Neurosci. 2022. doi: 10.1155/2022/8464452
Li, C., Wang, J., Niu, Z., Yao, J., and Zeng, X. (2021a). A spatial-temporal gated attention module for molecular property prediction based on molecular geometry. Brief. Bioinformat. 22. doi: 10.1093/bib/bbab078
Li, C., Wei, W., Li, J., Yao, J., Zeng, X., and Lv, Z. (2021b). 3DMol-Net: learn 3D molecular representation using adaptive graph convolutional network based on rotation invariance. IEEE J. Biomed. Health Inform. 26, 5044–5054. doi: 10.1109/JBHI.2021.3089162
Li, C., Yao, J., Su, J., Liu, Z., Zeng, X., and Huang, C. (2023). “LagNet: deep lagrangian mechanics for plug-and-play molecular representation learning,” in The 37th AAAI Conference on Artificial Intelligence (AAAI) (Washington, DC: AAAI Press) 37, 5169–5177.
Li, C., Yao, J., Wei, W., Niu, Z., Zeng, X., Li, J., et al. (2024). Geometry-based molecular generation with deep constrained variational autoencoder. IEEE Trans. Neural Netw. Learn. Syst. 35, 4852–4861. doi: 10.1109/TNNLS.2022.3147790
Li, H., Wang, X., Zhang, Z., and Zhu, W. (2022b). OOD-GNN: out-of-distribution generalized graph neural network. IEEE Trans. Knowl. Data Eng. 35, 7328–7340. doi: 10.1109/TKDE.2022.3193725
Li, H., Zhang, Z., Wang, X., and Zhu, W. (2022c). “Learning invariant graph representations for out-of-distribution generalization,” in Proceedings of the 36th International Conference on Neural Information Processing Systems (NeurIPS) New Orleans, LA: Neural Information Processing Systems Foundation, Inc. (NeurIPS).
Li, Z., Wu, Q., Nie, F., and Yan, J. (2022d). “GraphDE: a generative framework for debiased learning and out-of-distribution detection on graphs,” in Proceedings of the 36th International Conference on Neural Information Processing Systems (NeurIPS). New Orleans, LA: Neural Information Processing Systems Foundation, Inc. (NeurIPS).
Li, Z., Xu, Z., Cai, R., Yang, Z., Yan, Y., Hao, Z., et al. (2025). Identifying semantic component for robust molecular property prediction. IEEE Trans. Pattern Analy. Mach. Intellig. 47, 3796–3814. doi: 10.1109/TPAMI.2025.3598461
Lipton, Z. (2016). The mythos of model interpretability. arXiv [preprint] arXiv:1606.03490. doi: 10.48550/arXiv.1606.03490
Liu, B., Boysen, J., Unarta, I., Du, X., Li, Y., and Huang, X. (2025). Exploring transition states of protein conformational changes via out-of-distribution detection in the hyperspherical latent space. Nat. Commun. 16. doi: 10.1038/s41467-024-55228-4
Liu, G., Zhao, T., Xu, J., Luo, T., and Jiang, M. (2022a). “Graph rationalization with environment-based augmentations,” in Proceeding of the 28th ACM Conference on Knowledge Discovery and Data Mining (KDD) (New York, NY: Association for Computing Machinery), 1069–1078.
Liu, H., Hu, B., Wang, X., Shi, C., Zhang, Z., and Zhou, J. (2022b). “Confidence may cheat: Self-training on graph neural networks under distribution shift,” in The Web Conference New York, NY: Association for Computing Machinery. 1248–1258.
Liu, Q., Chan, R., and Yu, R. (2024). “The efficacy of pre-training in chemical graph out-of-distribution generalization,” in Proceedings of the 41st International Conference on Machine Learning (ICML) Vienna.
Lu, Y., Wang, C., Wang, Z., Zhang, X., Zhou, G., and Li, C. (2025). Semi-supervised learning-based virtual adversarial training on graph for molecular property prediction. Alexandria Eng. J. 115, 491–500. doi: 10.1016/j.aej.2024.11.110
Mbatha, S., Mulaudzi, T., Mbita, Z., Adeola, H., Batra, J., Blenman, K., et al. (2023). Artificial Intelligence-Enhanced Drug Discovery and the Achievement of Next-Generation Human-Centered Health System. Cham: Springer Nature Switzerland, 155–177.
Merck (2023). Merck Enters Two Strategic Collaborations to Strengthen AI-Driven Drug Discovery Darmstadt. Available online at: https://www.merckgroup.com/en/news/two-ai-partnerships-in-healthcare-20-09-2023.html
Miao, S., Liu, M., and Li, P. (2022). “Interpretable and generalizable graph learning via stochastic attention mechanism,” in Proceedings of the 39th International Conference on Machine Learning (ICML) (Baltimore, MD), 15524–15543.
Mittelstadt, B., Allo, P., Taddeo, M., Wachter, S., and Floridi, L. (2016). The ethics of algorithms: Mapping the debate. Big Data Soc. 3. doi: 10.1177/2053951716679679
Muandet, K., Balduzzi, D., and Scholkopf, B. (2013). “Domain generalization via invariant feature representation,” in Proceedings of the 30th International Conference on Machine Learning (ICML) Atlanta, GA.
National Institutes of Health (NIH) (2023). NIH Launches Bridge2AI Program to Expand the Use of Artificial Intelligence in Biomedical and Behavioral Research. Available online at: https://www.nih.gov/news-events/news-releases/nih-launches-bridge2ai-program-expand-use-artificial-intelligence-biomedical-behavioral-research
OECD Legal Instruments (2019). OECD AI Principles Overview. Available online at: https://oecd.ai/en/ai-principles
OpenAI (2023a). GPT-4 technical report. arXiv [preprint] arXiv.2303.08774. doi: 10.48550/arXiv.2303.08774
Rossi, S., Rossi, M., Mukkamala, R., Thatcher, J., and Dwivedi, Y. (2024). Augmenting research methods with foundation models and generative AI. Int. J. Inform. Managem. 77:102749. doi: 10.1016/j.ijinfomgt.2023.102749
Russell, S. (2019). Human Compatible: Artificial Intelligence and the Problem of Control. New York City: Viking.
Sadeghi, A., Ma, M., Li, B., and Giannakis, G. (2021). Distributionally robust semi-supervised learning over graphs. arXiv [preprint] arXiv:2110.10582. doi: 10.48550/arXiv.2110.10582
Salmi, M., Atif, D., Oliva, D., Abraham, A., and Ventura, S. (2024). Handling imbalanced medical datasets: review of a decade of research. Artif. Intellig. Rev. 57:273. doi: 10.1007/s10462-024-10884-2
Shen, X., Wang, Y., Zhou, K., Pan, S., and Wang, X. (2024). “Optimizing ood detection in molecular graphs: A novel approach with diffusion models,” in Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining (New York: ACM), 2640–2650.
Soares, N., Fallenstein, B., Yudkowsky, E., and Armstrong, S. (2015). “Corrigibility,” in Workshops at the Twenty-Ninth AAAI Conference on Artificial Intelligence Austin, TX.
Taddeo, M., and Floridi, L. (2018). How AI can be a force for good. Science 361, 751–752. doi: 10.1126/science.aat5991
Tallberg, J., Erman, E., Furendal, M., Geith, J., Klamberg, M., and Lundgren, M. (2023). The Global Governance of Artificial Intelligence: Next Steps for Empirical and Normative Research. 25, viad040. doi: 10.1093/isr/viad040
The Central People's Government of the People's Republic of China (2024). 2024 Government Work Report - Accelerating the Development of New Quality Productivity. Available online at: https://www.gov.cn/zhengce/jiedu/tujie/202403/content_6936388.html.
Theunissen, L., Mortier, T., Saeys, Y., and Waegeman, W. (2025). Evaluation of out-of-distribution detection methods for data shifts in single-cell transcriptomics. bioRxiv. doi: 10.1101/2025.01.24.634709
Tossou, P., Wognum, C., Craig, M., Mary, H., and Noutahi, E. (2024). Real-world molecular out-of-distribution: specification and investigation. J. Chem. Inf. Model. 64, 697–711. doi: 10.1021/acs.jcim.3c01774
Touvron, H., Martin, L., Stone, K., Albert, P., Almahairi, A., Babaei, Y., et al. (2023). Llama 2: Open foundation and fine-tuned chat models. arXiv [preprint] arXiv.2307.09288. doi: 10.48550/arXiv.2307.09288
Turchin, A., and Denkenberger, D. (2020). Classification of global catastrophic risks connected with artificial intelligence. AI Soc. 35, 147–163. doi: 10.1007/s00146-018-0845-5
Uddin, M., Sovon, M., Mondal, S., Ahmed, S., Al-Mizan, M., Aktar, F., et al. (2025). Artificial intelligence and machine learning in pharmaceutical sciences: Unpacking regulatory guidance, opportunities, and challenges for safe and effective drug development. Int. J. Med. Inform. doi: 10.2139/ssrn.5505018
UK Biobank (2025). AI-Driven Integration of Multi-Modal Data for Biomarker Discovery in Early Mental Disorder Diagnosis and Prediction. Available online at: https://www.ukbiobank.ac.uk/projects/ai-driven-integration-of-multi-modal-data-for-biomarker-discovery-in-early-mental-disorder-diagnosis-and-prediction/
Urbina, F., Lentzos, F., Invernizzi, C., and Ekins, S. (2022). Dual use of artificial-intelligence-powered drug discovery. Nat. Mach. Intellig. 4, 189–191. doi: 10.1038/s42256-022-00465-9
US Food and Drug Administration (2021). Artificial Intelligence/Machine Learning (AI/ML)-Based Software as a Medical Device (SAMD) Action Plan. Available online at: https://www.fda.gov/media/145022/download
Vijayan, R., Kihlberg, J., Cross, J., and Poongavanam, V. (2022). Enhancing preclinical drug discovery with artificial intelligence. Drug Discov. Today 27, 967–984. doi: 10.1016/j.drudis.2021.11.023
Vines, J., Clarke, R., Wright, P., McCarthy, J., and Olivier, P. (2013). “Configuring participation: On how we involve people in design,” in Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI) (New York, NY: Association for Computing Machinery), 249–438.
Wang, H., Vaze, S., and Han, K. (2025). Dissecting out-of-distribution detection and open-set recognition: a critical analysis of methods and benchmarks. J. Comp. Vision 133, 1326–1351. doi: 10.1007/s11263-024-02222-4
Wang, L., Zhang, Z., Wang, D., Cao, W., Zhou, X., Zhang, P., et al. (2023). Human-centered design and evaluation of AI-empowered clinical decision support systems: a systematic review. Front. Comp. Sci. 5:1187299. doi: 10.3389/fcomp.2023.1187299
Weckbecker, M., Anžel, A., Yang, Z., and Hattab, G. (2024). Interpretable molecular encodings and representations for machine learning tasks. Comput. Struct. Biotechnol. J. 23, 2326–2336. doi: 10.1016/j.csbj.2024.05.035
Wu, M., Pan, S., Zhu, X., Zhou, C., and Pan, L. (2019). “Domain-adversarial graph neural networks for text classification,” in IEEE International Conference on Data Mining (ICDM) (Beijing: IEEE), 648–657.
Wu, Y., Wang, X., Zhang, A., He, X., and Chua, T. (2022). “Discovering invariant rationales for graph neural networks,” in Proceedings of the 10th International Conference on Learning Representations (ICLR) International Conference on Learning Representations (ICLR).
Wu, Z., Ramsundar, B., Feinberg, E., Gomes, J., Geniesse, C., Pappu, A., et al. (2018). MoleculeNet: a benchmark for molecular machine learning. Chem. Sci. 9, 513–530. doi: 10.1039/C7SC02664A
Xia, J., Zhao, C., Hu, B., Gao, Z., Tan, C., Liu, Y., et al. (2023). “Mole-BERT: Rethingking pre-training graph neural networks for molecules,” in Proceedings of the 11th International Conference on Learning Representations (ICLR) Kigali: International Conference on Learning Representations (ICLR).
XtalPi (2025). Xtalpi and Pfizer Expand Strategic Collaboration to Advance AI-Driven Drug Discovery and Materials Science Simulations. Available online at: https://www.prnewswire.com/news-releases/xtalpi-and-pfizer-expand-strategic-collaboration-to-advance-ai-driven-drug-discovery-and-materials-science-simulations-302494033.html
Yang, N., Zeng, K., Wu, Q., Jia, X., and Yan, J. (2022). “Learning substructure invariance for out-of-distribution molecular representations,” in Proceedings of the 36th International Conference on Neural Information Processing Systems (NeurIPS) New Orleans, LA: Neural Information Processing Systems Foundation, Inc. (NeurIPS).
Yehudai, G., Fetaya, E., Meirom, E., Chechik, G., and Maron, H. (2021). “From local structures to size generalization in graph neural networks,” in International Conference on Machine Learning (ICML) (International Conference on Machine Learning (ICML)), 11975–11986.
Yu, H., Shen, Z., Miao, C., Leung, C., Lesser, V., and Yang, Q. (2018). “Building ethics into artificial intelligence,” in Proceedings of the 27th International Joint Conference on Artificial Intelligence (IJCAI) (Stockholm: International Joint Conference on Artificial Intelligence (IJCAI)), 5527–5533.
Zheng, Y., Rowell, B., Chen, Q., Kim, J., Kontar, R., Yang, X., et al. (2023). Designing human-centered AI to prevent medication dispensing errors: Focus group study with pharmacists. JMIR Format. Res. 7:e51921. doi: 10.2196/51921
Zhuang, X., Zhang, Q., Ding, K., Bian, Y., Wang, X., Lv, J., et al. (2023). “Learning invariant molecular representation in latent discrete space,” in Proceedings of the 37th International Conference on Neural Information Processing Systems (NeurIPS) New Orleans, LA: Neural Information Processing Systems Foundation, Inc. (NeurIPS).
Keywords: AI alignment, drug discovery, human values, generative AI, robustness, interpretability
Citation: Li C (2025) AI alignment is all your need for future drug discovery. Front. Artif. Intell. 8:1668794. doi: 10.3389/frai.2025.1668794
Received: 18 July 2025; Accepted: 13 October 2025;
Published: 03 November 2025.
Edited by:
Stefan Platz, AstraZeneca, United KingdomReviewed by:
Georgia Livieri, Cyprus University of Technology, CyprusZiye Wu, Guizhou University of Finance and Economics, China
Copyright © 2025 Li. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Chunyan Li, bGNoeUB5bm51LmVkdS5jbg==