Moving Past Species Classifications for Risk-Based Approaches to Food Safety: Salmonella as a Case Study

Classification of bacterial pathogens into species has traditionally been an essential prerequisite for many aspects of infection control and food safety. Consequently, control measures, regulations, industry specifications, and other guidance documents for foodborne pathogens typically specify a pathogen species (or genus) as target. For example, current Codex Alimentarius guidelines recommend absence of Salmonella in two 375 g samples for infant formula (FAO, 2020). For other ready-to-eat (RTE) foods, absence in smaller samples size (e.g., 25 g) is typically expected. However, there is increasing evidence for Salmonella, as well as other foodborne pathogens, that subgroups, serovars, and/or clonal groups within a given species may differ considerably in their ability to cause human disease. Importantly, there is emerging evidence that use of “sub-species” (or “sub-genus”) classification for defining food safety hazards may allow for more effective management of foodborne pathogens. The genus Salmonella, a pathogen where all members of the genus are considered a food safety hazard, represents one key foodborne pathogen where there is increasing evidence that subtypes differ considerably in their likelihood of causing human disease and where science-based subtype-specific and risk-based approaches of control may have a positive public health impact. There is clear evidence that certain Salmonella subtypes (e.g., certain subtypes within serovars Kentucky and Cerro) show reduced human virulence, while being frequently isolated from food animal species (specifically poultry and cattle) (Rodriguez-Rivera et al., 2014; Haley et al., 2016). Additionally, under the current approach of considering all Salmonella equally likely to cause human disease, limited progress has been made worldwide to reduce foodborne salmonellosis cases (Havelaar et al., 2015; Tack et al., 2020). In some situations, there are counterproductive incentives to prioritize control of highly frequent Salmonella strains with limited human public health relevance (e.g., through poultry vaccines targeting S. Kentucky) over control of less frequent Salmonella strains commonly linked to human salmonellosis cases. We outline a new paradigm for incorporating strain-specific assessment of human public health significance into regulations and control strategies, which we propose should be used to incentivize control of Salmonella subtypes with the greatest public health impact, with a focus on pre-harvest control in meat and poultry. This paradigm could subsequently be applied to other foodborne pathogens, such as Listeria monocytogenes.


INTRODUCTION
Classification of bacterial pathogens into species has traditionally been an essential prerequisite for many aspects of infection control and food safety. Consequently, control measures, regulations, industry specifications, and other guidance documents for foodborne pathogens typically specify a pathogen species (or genus) as target. For example, current Codex Alimentarius guidelines recommend absence of Salmonella in two 375 g samples for infant formula (FAO, 2020). For other ready-to-eat (RTE) foods, absence in smaller samples size (e.g., 25 g) is typically expected. However, there is increasing evidence for Salmonella, as well as other foodborne pathogens, that subgroups, serovars, and/or clonal groups within a given species may differ considerably in their ability to cause human disease. Importantly, there is emerging evidence that use of "sub-species" (or "sub-genus") classification for defining food safety hazards may allow for more effective management of foodborne pathogens. The genus Salmonella, a pathogen where all members of the genus are considered a food safety hazard, represents one key foodborne pathogen where there is increasing evidence that subtypes differ considerably in their likelihood of causing human disease and where science-based subtype-specific and risk-based approaches of control may have a positive public health impact. There is clear evidence that certain Salmonella subtypes (e.g., certain subtypes within serovars Kentucky and Cerro) show reduced human virulence, while being frequently isolated from food animal species (specifically poultry and cattle) (Rodriguez-Rivera et al., 2014;Haley et al., 2016). Additionally, under the current approach of considering all Salmonella equally likely to cause human disease, limited progress has been made worldwide to reduce foodborne salmonellosis cases (Havelaar et al., 2015;Tack et al., 2020). In some situations, there are counterproductive incentives to prioritize control of highly frequent Salmonella strains with limited human public health relevance (e.g., through poultry vaccines targeting S. Kentucky) over control of less frequent Salmonella strains commonly linked to human salmonellosis cases. We outline a new paradigm for incorporating strain-specific assessment of human public health significance into regulations and control strategies, which we propose should be used to incentivize control of Salmonella subtypes with the greatest public health impact, with a focus on pre-harvest control in meat and poultry. This paradigm could subsequently be applied to other foodborne pathogens, such as Listeria monocytogenes.

SALMONELLA IS NOT A "ONE SIZE FITS ALL" PATHOGEN
The genus Salmonella consists of two species, Salmonella enterica and Salmonella bongori; S. enterica can be further divided into six recognized subspecies, the most clinically relevant of which is S. enterica subsp. enterica (Worley et al., 2018), a pathogen of warm-blooded mammals including humans. Within S. enterica subsp. enterica (hereafter referred to as "S."), there are >1,500 serovars (Issenhuth-Jeanjean et al., 2014), which differ in their ability to cause human disease (Brenner et al., 2000). Typhoidal (i.e., S. Typhi) and paratyphoidal (e.g., S. Paratyphi A) serovars cause a systemic illness characterized by fever, headache, malaise, abdominal pain, and rash, while non-typhoidal serovars (e.g., S. Typhimurium) typically cause a self-limiting gastroenteritis. Additionally, S. Typhi and S. Paratyphi A are considered to be human-restricted (i.e., they only cause disease in and are spread by humans) serovars and are typically transmitted through food and/or water contaminated with human feces due to poor sanitation. Non-typhoidal serovars on the other hand are transmitted through both direct animal contact and food (as well as fomites) with contamination often linked to animal hosts and foods derived from warm-blooded animals (e.g., raw meat and poultry, raw eggs, raw milk and dairy products products) (Gal-Mor et al., 2014). However, a variety of foods have been sources of non-typhoidal Salmonella cases and outbreaks, including fruits and vegetables as well as dry products (e.g., spices, chocolate) and non-typhoidal Salmonella have been shown to also survive for extended times in extra-host environments such as soil and even processing plant environments (Waldner et al., 2012;Jechalke et al., 2019). Because typhoidal and paratyphoidal serovars typically cause systemic disease, they have traditionally been considered a higher public health risk than non-typhoidal serovars. Non-typhoidal serovars, despite differing in their ability to cause disease in humans, are considered an equal public health risk. With this current approach, efforts to decrease Salmonella human clinical cases in the United States have essentially failed (Tack et al., 2020). While Healthy People (2020) targeted a reduction in Salmonella incidence from 15.0 cases to 11.4 cases per 100,000 people by 2020, Salmonella incidence reached 17.1 cases per 100,000 people in 2019 (Tack et al., 2020). Similarly, at a global level, Salmonella continues to be the foodborne pathogen with the greatest public health impact based on disabilityadjusted life years (Havelaar et al., 2015).
While regulations in many countries consider all nontyphoidal Salmonella (NTS) serovars a hazard and equal public health risk, there is strong scientific evidence that NTS serovars and clonal groups differ in the public health risk they represent, as supported by (i) presence or absence of virulence genes that encode fully functional virulence factors; (ii) phenotypic data (e.g., in tissue culture or animal models); and (iii) epidemiological evidence (e.g., under-representation of certain serovars or clonal groups among human clinical cases relative to their presence in food, raw materials, or animals). Salmonella serovars or clonal groups may also show strong geographical associations which may need to be taken into account when assessing epidemiological evidence for hypo-or hypervirulence. Instances where all three lines of evidence suggest reduced virulence in specific serovars or clonal groups represent a particularly strong case for classifying a group as a reduced public health risk. However, even two lines of evidence [e.g., (i) and (iii)] may be sufficient in some cases to identify groups that differ in the public health risk they represent. This general approach is essentially what has been used to identify Escherichia coli serovars and groups (e.g., Enterohemorrhagic E. coli [EHEC]) that represent a particular public health risk. Importantly, as detailed below for NTS, sufficient data appear to be already available for some NTS to allow for science-based classification of virulence groups.
It is well-established that Salmonella serovars and clonal groups within a serovar can differ greatly in the virulence factors they encode. One important example being differences in gene content on Salmonella Pathogenicity Islands (SPIs), which are large islets of genes that contribute to a particular virulence phenotype (Marcus et al., 2000). Currently, there are 24 recognized SPIs (Cheng et al., 2019); data on presence/absence of these SPIs and genes in these SPIs, as well as data on the functional conservation of genes in these SPIs (e.g., absence of premature stop codons), can be used to assess the virulence potential and public health risks associated with Salmonella serovars or clonal groups. For example, SPI-7, which encodes the Vi polysaccharide (Helena, 2008), is found in multiple serovars associated with invasive human disease, including serovars Typhi (Faucher et al., 2006;Lee et al., 2006), Paratyphi C (Daniels et al., 1989), and Dublin (Pickard et al., 2003). On the other hand, S. Cerro, a serovar rarely associated with human disease, is missing several genes found in SPIs 10, 12, and 13 (Rodriguez-Rivera et al., 2014), and contains premature stop codons in genes involved in pathogenesis, phsA and sopA (Rodriguez-Rivera et al., 2014;Kovac et al., 2017b). Existing evidence thus suggests that genomic information can help to assess both (i) the likelihood that a given Salmonella serovar will cause human disease and (ii) the likelihood of causing more severe or mild human disease.
Additionally, epidemiological evidence suggests that some serovars show reduced likelihood of causing human disease. This type of evidence is particularly strong for meat and poultry associated Salmonella, where large datasets on serovars present in raw meat and poultry are available. For example, in the US, S. Cerro represents the most common serovar isolated from cattle without clinical signs of salmonellosis (National Salmonella Surveillance, 2018b), but this serovar only accounted for 38 culture-confirmed cases of human salmonellosis in the US in 2016 (National Salmonella Surveillance, 2018a). Similarly, S. Kentucky was the most frequently isolated serovar isolated from US broiler chickens between 1998 and 2014 (USDA FSIS, 2014), but this serovar is also isolated from very few cases of human clinical illnesses in the US (e.g., 63 in 2016) (National Salmonella Surveillance, 2018b). Hence, regulations and control efforts that consider all NTS serovars and clonal groups an equal public health risk may negatively impact the ability to effectively control serovars that represent the largest public health burden. For example, a framework where all NTS are considered an equal public health risk would encourage vaccination of live animals (e.g., poultry) against those serovars most commonly found in a given population, even if those serovars are characterized by mutations that confer reduced human virulence and are thus unlikely to cause human disease. This is a particular concern as vaccination of live animals against a large number of Salmonella serovars is not possible and as most polyvalent animal Salmonella vaccines typically target at most three serovars (Berghaus et al., 2011).

CURRENT REGULATORY AND INDUSTRY APPROACHES TO CONTROL SALMONELLA IN FOODS
Salmonella is a pathogen of concern in a wide range of products, from infant formula, to Ready-To-Eat (RTE) foods (e.g., spices, chocolate) to produce, raw meat and poultry. Regulatory approaches to control this pathogen reflect the diversity of products where Salmonella is a concern. While absence of Salmonella in specified sample sizes (e.g., one 25 g sample) is typically the limit for RTE foods, many countries do not have clear regulatory limits for Salmonella in raw meat and poultry. For example, the US has a Salmonella verification program, which was created in 1996 through the Pathogen Reduction, Hazard Analysis and Critical Control Point (PR/HACCP) Systems Final Rule (USDA FSIS, 2019). The purpose of this program is to verify that establishments that process raw meat and poultry have implemented proper process control strategies to prevent Salmonella contamination. The current system specifies a maximum acceptable number of % positive samples within a moving 52-week window with a different threshold for each USDA FSIS-regulated product. While the USDA FSIS performance standards target all Salmonella spp. and serovars, regulatory programs targeting specific Salmonella serovars also exist. For example, the US National Poultry Improvement Plan includes a focus on control of two specific Salmonella serovars: S. Enteritidis and S. Gallinarum biovars Pullorum and Gallinarum (Shivaprasad, 2000; USDA APHIS, 2020); the focus on S. Enteritidis specifically addresses the serovar that causes the largest number of poultry and egg associated salmonellosis cases in the US. A similar focus on control of S. Enteritidis (and other serovars of particular human health concern) is embedded in regulations in other countries, such as European Commission Regulations (EC) No. 646/2007(12 June 2007) and 584/2008(20 June 2008, which target reduction of S. Enteritidis and S. Typhimurium prevalence in broilers and turkeys, respectively. Reflecting the diversity of foods associated with human salmonellosis cases, control strategies for different products and commodities also vary. For RTE products, effective control can be achieved through implementation of HACCP or FSMAcompliant food safety plans with appropriate preventive controls. In RTE products, detection of Salmonella at low levels (1 Salmonella cell in 25 or 375 g) would indicate a failure of the HACCP or food safety plan; hence regulatory standards that target detection of Salmonella spp. are warranted for RTE products. However, control of Salmonella in raw meat and poultry, is considerably more challenging due to the lack of a kill step (a processing step that substantially reduces target pathogen levels [e.g., by 5 log]) and involves a combination of strategies, including some that target all Salmonella (and in most cases a larger group, such as all Enterobacteriaceae) as well as some that by design or necessity can only target specific Salmonella serovars or clonal groups. For example, vaccination of poultry against Salmonella, while a highly effective control strategy, can typically only target a few Salmonella serovars. Supporting the efficacy of vaccinations, Dórea et al. showed that compared to a farm without a vaccination strategy, a farm vaccinating breeder chickens against serovars Typhimurium, Berta, and Kentucky displayed significantly lower prevalence of Salmonella in the ceca and reproductive tracts of hens and their progeny, as well as significantly lower prevalence of Salmonella in environmental samples (Dórea et al., 2010). This study however also highlights one of the challenges of using vaccines to control Salmonella in meat and poultry; if regulations and performance standards target all Salmonella, then the vaccine used would typically be selected to target the most common serovars present in a given animal species or in a given operation and not necessarily the serovars representing the largest public health risk. Consistent with this, the Dórea et al. study, which was conducted in the US, included serovar Kentucky as one of the three vaccine strains, most likely since this serovar is common in poultry; however, as detailed above, in the US, the most prevalent S. Kentucky sequence type, ST152, is rarely associated with human cases (Haley et al., 2016) and inclusion of another serovar would likely have a larger public health benefit.
The poultry industry currently relies on two types of vaccines to vaccinate flocks against Salmonella: autogenous and live attenuated. Autogenous vaccines are derived from Salmonella isolated from infected birds and are used in the flock the Salmonella was isolated from. These vaccines must be redesigned every 2 years and administration of these vaccines is often labor-intensive as birds must be given them multiple times. Thus, autogenous vaccines are typically used for a short amount of time and only in cases in which no other alternatives exist. Live attenuated vaccines only protect against certain serovars, but are administered through less labor-intensive methods, including drinking water or sprays. However, they require approval by the USDA, which can often take up to 7 years (Eskin, 2020). With limited options available, farmers must focus vaccination efforts on vaccine options that are available to them or those serovars most prevalent in their flocks, often resulting in vaccination against serovars with limited public health relevance. To maximize the benefits of public health risk-based classification of Salmonella, vaccination options thus need to expand to facilitate vaccination against serovars with high public health relevance. This illustrates why a regulatory approach and framework that incentivizes control of Salmonella serovars that represent a higher public health risk would likely have significant public health benefits, particularly with regard to human salmonellosis cases attributed to raw meat and poultry. In addition to facilitating vaccine deployment that emphasizes positive public health outcomes, a public health and serovar focused approach to control Salmonella in raw meat and poultry could also facilitate and accelerate innovation with regard to development of effective serovar-specific control strategies that can be applied at the farm and raw product level (e.g., phage-based control strategies).  (Worley et al., 2018); serovars that are known to be polyphyletic are shown in blue.

A FRAMEWORK FOR SEROVAR AND CLONAL GROUP SPECIFIC AND RISK-BASED SALMONELLA CONTROL STRATEGIES IN RAW MEAT AND POULTRY
As detailed above, there is clear evidence that new approaches are needed to improve our ability to control Salmonella transmission through foods. Improved control of Salmonella in raw meat and poultry in particular would likely have a significant public health impact, as these products are a frequent source of human Salmonella infections (IFSAC, 2020). However, with the frequent presence of Salmonella in poultry and meat animals, risk-based approaches that target prevention and control will be essential to make public health gains. Possible risk-based strategies would be to focus control efforts on animal populations (e.g., herds, flocks) that have a high Salmonella prevalence and that contribute high Salmonella loads and/or to focus control efforts on the specific animals shedding high levels of Salmonella (so called "supershedders"); a focus on supershedders can however be difficult as these are rare and hence, difficult to identify. Another opportunity to shift toward a risk-based approach to Salmonella control is to develop and implement a framework that facilitates interventions focusing on Salmonella serovars that represent the largest public health risk and a reduced focus on serovars that represent a low public health risk. Importantly, a strategy that considers all Salmonella an equal public health risk (which, with few exceptions, is the de facto status quo in many countries around the world) leads to the unintended consequence that resources and efforts are focused on control of the most prevalent serovars and clonal groups. This is a particular concern for meat and poultry, where, at least in some countries and as detailed above, some of the most frequently isolated Salmonella (e.g., serovars Cerro and Kentucky ST152 in the US) show clear evidence for reduced human virulence. Hence, we propose that a shift toward public health prioritization of meat and poultry associated Salmonella with an initial focus on strategies that define virulence attenuated Salmonella and reduce unintentional incentives to focus on intervention efforts (e.g., vaccines) that target these Salmonella groups will improve our ability to reduce the public health burden of human salmonellosis.
A key challenge is how to develop an effective and sciencebased approach to classify Salmonella serovars based on their human public health risk. Current approaches for public health classification of foodborne pathogens typically take a binary approach and classify bacterial groups (genera, species, or serovars) as a public health hazard (or not). An alternative approach, which we propose for Salmonella control in raw meat and poultry, would be to develop and implement a quantitative and risk-based approach to public health classification. This could be particularly valuable to allow for modification of current performance standards (such as the USDA FSIS performance standard detailed above), where a numerical "virulence modifier" could be used to adjust the contributions of samples positive for virulence attenuated Salmonella (Figure 1A). Defining sciencebased numerical modifiers (or "adjustment factors") will not be simple and likely introduces additional challenges. We propose that a stepwise approach to defining adjustment factors for different Salmonella serovars or clonal groups, focusing first on common serovars where there is scientific support for reduced ability to cause human disease (e.g., serovar Cerro, serovar Kentucky ST152) (Haley et al., 2016;National Salmonella Surveillance, 2018b), represents the most likely path to success. While one can envision numerous possible strategies to define adjustment factors, we will briefly outline two possible approaches. As detailed above, initial high confidence identification and characterization of virulence attenuated Salmonella serovars or clonal groups is essential; ideally this classification should be based on a convergence of (i) presence/absence of well-characterized virulence-associated genes, (ii) phenotypic evidence from cell culture and/or animal models, and (iii) epidemiological evidence. Synthesis of these data will reduce the impact of potential biases (e.g., sampling or sequencing biases) that may result in falsely predicting virulence attenuation of some serovars or clonal groups.
One approach that can be used to define an adjustment factor would be based on the relative prevalence of a given serovar or clonal group among the source population (e.g., poultry) and human clinical cases; a similar approach has been used to classify L. monocytogenes clonal complexes into hypoand hypervirulent clones (Maury et al., 2016). Conceptually, this approach quantifies over-and underrepresentation of pathogen groups (e.g., serovars) among human cases relative to exposure; for example, a factor of 0.1 would indicate that a given group is 10 times less likely to cause human disease relative to the "average" Salmonella. Consequently, this number could be directly used as an adjustment factor ( Figure 1A). Possible challenges and drawbacks of this approach include the importance of representative sampling of different sources (to allow for a reliable estimate of this factor) as well as other possible biases in the data used to calculate this factor. A more involved, but likely more robust approach to define an adjustment factor, would be to define the probability of different Salmonella to cause human disease, for example using the r value defined for a given serovar or clonal group. The r value represents the probability of a single pathogen cell causing illness (Chen et al., 2006); the log 10 r value for the genus Salmonella has been estimated to be around 2.6 (i.e., 400 cells) (FAO/WHO, 2002). A possible approach for calculating r values for different strains or subgroups of a foodborne pathogen has previously been reported for L. monocytogenes (Chen et al., 2011); mean log 10 r values were estimated to be 8.1 and 10.7 for L. monocytogenes subtypes with genes encoding a full-length and a truncated version of the key virulence gene inlA, respectively. A similar approach could be used to initially define r values for all Salmonella and for putative virulence attenuated serovars or clonal groups, which could subsequently be used to calculate an adjustment factor. For example, if the r value for all Salmonella was 400, while the r value for S. Cerro was estimated to be 4,000, then an adjustment factor of 0.1 may be appropriate for use with S. Cerro.
Importantly, a risk-based system that accounts for differences in the ability of Salmonella serovars or clonal groups to cause human disease, requires effective surveillance and monitoring strategies that can identify changes in virulence and emergence of clonal groups with unique hypo-or hypervirulence phenotypes. The widespread use of whole genome sequencing (WGS) for foodborne disease surveillance in an increasing number of countries represents an effective tool for "virulence surveillance, " which is essential for reliable and safe implementation of a riskbased approach to Salmonella control as described above. The NCBI Pathogen Detection browser (ncbi.nlm.nih.gov/pathogen) integrates WGS information for human clinical, animal, food, and environmental isolates and as of March 24, 2021, includes WGS data for >350,000 Salmonella isolates, which are classified into single nucleotide polymorphism (SNP) clusters; an isolate is grouped into a given SNP cluster if it differs by <50 SNPs from at least one isolate already in the cluster. NCBI Pathogen Detection thus provides an essential infrastructure for implementation of modern, risk-based precision food safety approaches. Importantly, this dataset can be used to define both potential hypo-and hypervirulent serovars and clonal groups, by identifying groups with over and under-representation of human isolates (Figure 1B). These WGS data can also be used to identify genomic features and mutations that are likely to be associated with attenuated human virulence. With this strategy, representative and reliable WGS data for human, food, and animal isolates are essential. Currently, CDC, USDA FSIS, and FDA all perform WGS on Salmonella isolates obtained through their respective surveillance systems. For example, USDA FSIS uses WGS to characterize Salmonella isolates from product samples, National Antimicrobial Resistance Monitoring System (NARMS) cecal samples, and USDA Agriculture Marketing Service samples (USDA FSIS, 2020). CDC and state health department partners perform WGS on all human Salmonella isolates obtained in the US. Continued routine collection of WGS data is essential to ensure that the system proposed here works effectively and appropriately. In addition, increased funding and continued efforts to standardize WGS approaches across regulatory and public health agencies will be extremely valuable for this type of approach. On the other hand, it is important to note that the food industry is still in early phases of implementing the use of WGS data. Currently, the food industry uses WGS data for root cause analyses in the case of pathogen contamination events and for identification of virulence factors or antimicrobial resistance genes in a pathogen, but it is rarely used for routine surveillance (Jagadeesan et al., 2019). As such, industry WGS data is currently limited and may be potentially biased (e.g., WGS may be performed predominantly for strains that cause clinical disease).
Virulence and public health risk associated with a given clonal group (e.g., a SNP cluster) and a serovar is not static; use of NCBI Pathogen Detection (or a similar dataset or repository) to monitor for evidence of changes in virulence potential will be essential. Examples of an indication for increased virulence potential of a clonal group previously considered to represent a reduced public health would be if (i) isolates of a given clonal group newly added to a database are found to contain mutations or mobile genetic elements (e.g., plasmids and phages) that are predicted to restore virulence (Miller et al., 2020;Tassinari et al., 2020) or if (ii) addition of human clinical isolates significantly increases the number of human isolates in this clonal group (indicating an increased frequency of human cases by a clonal group previously assumed to be "hypovirulent"). Additionally, monitoring WGS and other data for indications of an increased prevalence of a given serovar or clonal group can provide important data on the changing ecology and transmission of Salmonella. For example, a campaign to eliminate S. Gallinarum in chickens during the early 20th century resulted in other serovars, like S. Enteritidis, exploiting a newly unoccupied niche and causing it to become a major cause of human salmonellosis (Callaway et al., 2008). As such, increased monitoring of WGS data will also allow for faster detection of similar phenomena and thus, faster response to prevent a similar outcome. Finally, virulence capacity can also vary over time or based on geographical location with specific hypoor hypervirulent Salmonella serovars or clonal groups being associated with a specific region. Increased monitoring of WGS data will be useful in quickly identifying so-called "geographic hotspots" where hypo-or hypervirulent clonal groups or serovars have increased in prevalence. For example, S. Kentucky appears to include clonal groups that differ in the ability to cause human disease with some regions characterized by circulation of fully virulent strains (e.g., ST198 in Europe), while other regions may be characterized by circulation of strains that show reduced likelihood of causing human disease (e.g., ST152 in the US) (Haley et al., 2016). Importantly, these examples illustrate that implementation of food safety systems that consider Salmonella virulence differences require a sophisticated food safety and public heath infrastructure.

CONCLUSIONS
While hazard-based approaches (e.g., HACCP) have been credited to have played an important role in reducing the public health burden associated with bacterial foodborne illnesses, there is increasing evidence that continuous food safety improvements will require improved implementation of riskbased and precision food safety approaches (Kovac et al., 2017a). While risk-based food safety systems can consider a number of different factors that affect risk (e.g., pathogen loads in food, ability of a food to support pathogen growth, host susceptibility), differences in the ability of a given pathogen group (e.g., species, genus) to cause human disease are rarely explicitly considered in regulations, performance standards, or food safety strategies. With the emergence of new tools and data (e.g., large WGS datasets for foodborne pathogen isolates from diverse sources), it is now feasible to develop food safety strategies that consider differences in ability to cause human disease, among members of a given species or genus that currently may be considered a "carte blanche" pathogen. As detailed here, the genus Salmonella represents a pathogen group where stepwise implementation of a new approach that considers the public health risk associated with different serovars and clonal groups may provide for a considerably improved ability to prevent human Salmonella infections linked to raw meat and poultry, by focusing directed interventions (e.g., vaccination) on those Salmonella most likely to cause human infections.

AUTHOR CONTRIBUTIONS
This opinion piece was conceptualized by RC, RO, and MW. AC and MW wrote, reviewed, and edited the original and final drafts. RO and RC reviewed and edited the final drafts. Funding was secured by MW. All authors contributed to the article and approved the submitted version.

FUNDING
This work was funded by Pew Charitable Trusts contract number 34063. RC was supported by USDA 2020-67034-31905.