- Centro de Investigación en Computación, Instituto Politécnico Nacional, Mexico City, México
Classification is a central task in machine learning, underpinning applications in domains such as finance, medicine, engineering, information technology, and biology. However, machine learning pattern classification can become a complex or even inexplicable task for current robust models due to the complexity of objective datasets, which is why there is a strong interest in achieving high classification performance. On the other hand, in particular cases, there is a need to achieve such performance while maintaining a certain level of explainability in the operation and decisions of classification algorithms, which can become complex. For this reason, an algorithm is proposed that is robust, simple, highly explainable, and applicable to datasets primarily in medicine with complex class imbalance. The main contribution of this research is a novel machine learning classification algorithm based on binary string similarity that is competitive, simple, interpretable, and transparent, as it is clear why a pattern is classified into a given class. Therefore, a comparative study of the performance of the best-known state-of-the-art classification algorithms and the proposed model is presented. The experimental results demonstrate the benefits of the proposal in this research work, which were validated through statistical hypothesis tests to assess significant performance differences.
1 Introduction
This study proposes a novel algorithm for pattern classification. The proposed algorithm is primarily based on binary string similarity and is called the N-Similarity Binary Classifier (n-SBC), as it uses the Hamming string similarity method and a binary-value encoder called the reflected binary code (RBC) or Gray code. The creation, design, implementation, and application of n-SBC support the solution-finding process for the problem represented by the supervised case in pattern classification.
Humans can recognize objects, actions, and everyday elements (patterns); however, what is simple for humans can be a very complex problem for a computational algorithm. The discipline that includes in its field of study the modeling and programming of automatic object and action recognition tasks is Pattern Recognition (PR) (Sarker, 2021). There are four basic tasks of PR: classification, regression, recovery, and clustering (De Sa, 2012; Rane et al., 2024). The first three are located in the supervised learning paradigm, while the last one is the emblematic task of the unsupervised learning paradigm (Janani and Vijayarani, 2019). In this study, emphasis will be placed on the first task of the supervised learning paradigm: intelligent pattern classification, or machine learning for pattern classification. In the state of the art, a wide variety of conceptual bases provide theoretical support for the task, such as Bayesian classifiers and distance-based models like kNN. Based on decision trees (C4.5 or Random Forest), based on and inspired by the neurons of the human brain (Multilayer Perceptron), or based on optimization of analytical functions, such as support vector machines (Bhargavi and Jyothi, 2009; Cover and Hart, 1967; McCulloch and Pitts, 1943; Quinlan, 1990; Rosenblatt, 1958; LeCun et al., 2015; Cortes, 1995). As important assistants in the development of PR and related disciplines, there are dataset repositories (Dua and Graff, 2019) and certain platforms where some research groups make available to users valuable computational tools, as well as implementations of algorithms and methods; such is the case of WEKA (Hall et al., 2009) and KEEL (Bhargavi and Jyothi, 2009; Cover and Hart, 1967; McCulloch and Pitts, 1943; Quinlan, 1990; Rosenblatt, 1958; LeCun et al., 2015; Cortes, 1995), two of the most useful, famous and popular platforms.
Before 1997, when the No Free Lunch Theorem (De Sa, 2012; Duda et al., 2001) was published, a large number of research groups were trying to find the best classifier; however, this theorem resulted in researchers concluding that this search is futile, since there is no intelligent pattern classifier that is the best in all cases. Therefore, researchers' efforts are currently directed toward finding alternatives to improve the performance of pattern classifiers, recognizing that there is no best one. One of the main recent achievements has been the development of a new pattern classification paradigm, Minimalist Machine Learning (MML) (Yáñez-Márquez, 2020). It is in this context that the central proposition of this work arises.
Recently, significant efforts have been devoted to finding alternatives to improve the performance of intelligent pattern classifiers, recognizing that there is no single best approach. In these research processes, a wide variety of tools and methodologies developed over the decades are used. Thus, one recurring theme in the generated algorithms is the use of associative models. These associative models are not designed for intelligent pattern classification but rather for pattern retrieval; however, if the designer adequately represents the output patterns, they can perform the classification task correctly. The first associative model recorded is the Lernmatrix, created in 1961 by Steinbuch (1961), followed by the associative model called the Correlograph, whose creation and publication occurred 8 years later (Willshaw et al., 1969). The year 1972 saw the birth of one of the best-known associative models: the Linear Associator, which emerged as the fusion of two independent models (Kohonen, 1972; Anderson, 1972); from then on, a considerable number of associative models have been generated in the world with successful applications in various areas of human activity (Hopfield, 1982; Talib, 2018; Ibrahim and Abdulazeez, 2021). It is pertinent to note that research on the subject is ongoing (Hoffmann, 2019; Nozari et al., 2024; Zhu et al., 2024; Bian and Priyadarshi, 2024).
Early detection of diseases has increased its relevance in recent years due to the various benefits that have a beneficial impact on public health, such as increasing the chances of survival in patients suffering from severe respiratory diseases (Vayadande, 2024; Rasool et al., 2023) and achieving a better recovery thanks to detection at an early stage of the disease. Research focused on pre-diagnosis of respiratory diseases has recently gained momentum worldwide, with widespread interest in improving early detection. Currently, both invasive and non-invasive methods are applied. However, lately, the use of machine learning classification algorithms for disease diagnosis has become an increasingly important area of research globally due to their ease of implementation and accessibility (Rana and Bhushan, 2023). This has caused frequent research in the literature on the development of novel specialized models for the medical pre-diagnosis of all types of diseases (Kumar et al., 2023; Ahsan et al., 2022).
In this paper, elements of associative models have been taken in order to create and design the main algorithm of the proposal, but in addition to these elements, the concept of string similarity has been used, as well as the Hamming distance and a binary pattern encoder, the reflected binary code (RBC or Gray code). The rest of this paper is organized as follows: Section 2 details the related works. Section 3 describes the novel proposal algorithm, with detailed examples of its operations in training and classification phases. In Section 4, the experimental phase and results are presented, and, finally, in Section 5, the conclusions and future research are included.
2 Related research
As discussed above, assuming the existence of a universally optimal pattern classification algorithm is unmotivated due to the no free lunch theorem, forcing researchers in machine learning-related areas to focus on improving the performance of existing models and thereby reducing classification error (Bui et al., 2020; Shehadeh et al., 2021; Misra and Yadav, 2020). Alternatively, some studies propose entirely novel machine learning models for pattern classification, with the aim of exploring new possibilities, as demonstrated by Amygdalos et al. (2023) and Hissou et al. (2023). Similarly, researchers have pioneered the development of new algorithms based on associative memories, including those by Moreno-Ibarra et al. (2021), (Yang and Ding 2020), and Luna-Ortiz et al. (2023).
Section 2 is divided into three parts. Section 2.1 describes the Hamming Distance algorithm, a fundamental concept for our novel pattern classification algorithm. Section 2.2 explores the RBC (Reflected Binary Code), another crucial element of our novel method for converting the original dataset into binary strings. Finally, Section 2.3 provides an overview of the state-of-the-art machine learning algorithms, including both well-known models and associative memories used for classification tasks, as well as a deep dive into current research on Hamming Distance and RBC in machine learning.
2.1 Hamming distance
The Hamming distance, the most used metric with binary strings and a natural similarity measure on binary codes, can be computed with just a few machine instructions per comparison (Pappalardo et al., 2009). The computational effort required to calculate the Hamming distance linearly depends on the size of the string, and it is often used to quantify the extent to which two bit-strings of the same dimension differ (Norouzi et al., 2012; Bookstein et al., 2002).
The distance is defined as the minimum number of errors that could transform a pattern A into a pattern B, i.e., it measures the minimum number of values that must be changed to transform a string into another target string (Zhang et al., 2013).
Another way to define it could be the number of positions at which the corresponding bits are different, that is, express it as the following (Gaitanis et al., 1993):
where Ai and Bi are the bits at the i-th position of the respective strings. And the subtraction refers to the XOR logic gate operation. The use of the Hamming distance has many applications, the most relevant being in coding theory, the electronics field, and term clustering (Norouzi et al., 2012). It has been shown that one can perform exact nearest-neighbor search in Hamming space significantly faster than linear search, achieving sublinear run times.
2.2 Reflected Binary Code (RBC)
The Gray encoder, also known as Reflected Binary Code (RBC), was invented by Frank Gray in 1953 in a Bell Telephone Laboratories patent (Agrell et al., 2004; Doran, 2007; Goodall, 1951). It is a binary numbering system in which the main property is that two adjacent values differ by only a single digit. For example, value 2 differs from values 1 and 3 in RBC by a single digit. Table 1 is an illustrative example.
In this case, unlike the classic binary encoder, the bit difference between an adjacent decimal value is only one digit. In this sense, this advantage helps preserve similarity between neighboring patterns, unlike standard binary encoding, which can cause adjacent values to differ across multiple bits, creating more complex relationships between close patterns. Therefore, this helps and supports the performance of our proposed classifier, as explained in Section 3, since it is based on string binary simultaneities.
This system binary code is commonly used to refer to any single distance. Its unique characteristics make it very useful across different domains, especially for error correction, position encoders, genetic algorithms, and digital communication (Agrell et al., 2004; Bhat and Savage, 1996).
To obtain a binary string using RBC, it can be done as follows: First, convert the decimal value to classic binary code, and subsequently convert from binary code to RBC, applying XOR (exclusive OR) to each bit with the right bit, excluding the most significant bit. For example, let us say we want to convert the number 5 into RBC. First, the binary value of 5 is 101, and the MSB in this case is 1. Now, applying the XOR operation, starting from the right to the left but going on the right side, taking the second bit (0) and applying XOR with the first bit (1), the result is 1; then taking the third bit (1) and applying XOR with the second bit (0), the result is 1; and finally, concatenating the MSB as the first bit of the resulting string after the XOR operations; therefore, the RBC of the number 5 is 111 (Bhat and Savage, 1996).
Before converting decimal values to binary strings using the RBC method, the dataset values are preprocessed: the minimum value per feature is computed; if required, the decimals are truncated to 2 decimal places; and finally, the values are rounded to integers. This aims to obtain only positive integer values.
To illustrate the conversion to integer values and truncation, the following example is provided. Consider a continuous numeric feature:
In this case, to obtain only positive numbers, the minimum value is the sum of all the values of the feature array, which in this case is −0.11, obtaining the following result:
Then, it is truncated to two decimals only:
Subsequently, all the values of the feature are escalated to integer values, such as
Finally, using these feature values, the RBC binary string is computed. Table 2 shows an example of how the binary codes look after RBC encoding.
2.3 Pattern classification algorithms
In the current state of the art, many machine learning algorithms focus on classification tasks. Some of them are based on distance, such as the kNN (k-nearest neighbors) model (Zhang, 2021), while others are based on optimization, such as SVM (support vector machines) (Abdullah and Abdulazeez, 2021). Others are based on decision trees (Costa and Pedreira, 2023), such as C4.5, or bagging approaches such as the random forest algorithm. In more recent literature, models are inspired by biological concepts, such as the human brain. For instance, the multilayer perceptron (an artificial neural network) falls into this category. Currently, the most widely used are deep learning models (Sharifani and Amini, 2023), which are neural networks with many layers and additional specialized preprocessing stages, such as CNNs (convolutional neural networks) for image processing and transformers and embedding approaches for natural language processing tasks (Galli et al., 2024).
2.3.1 Associative memories
An associative memory M is a pattern-input/output system whose primary purpose is to learn to correctly retrieve complete patterns from inputs that may be corrupted by several sources of noise. Can be expressed as x → M → y. The input and output patterns are represented by the column vectors x and y, respectively. Each input is associated with a corresponding output pattern; such an association is expressed as (x, y). Memory M is represented by a matrix. This matrix is formed from a finite set of previously known associations, known as the fundamental set (considered as the learning stage).
Finally, the retrieval process (which could be known as the classification stage if the designer made adequate changes) consists of performing the memory M with the given steps for that phase, with the aim of finding enough conditions to obtain the fundamental output pattern y from the pattern x (Nozari et al., 2024).
In the state of the art, there are pioneers of associative memory with their original purpose as retrieval machines, such as Steinbuch's Lernmatrix and Linear Associator (Steinbuch, 1961; Nozari et al., 2024). In these models, learning is typically implemented by updating a memory matrix with a set of rules. For example, in a Lernmatrix, each association (x, y) contributes an update of the form M ← M + yx. During inference, the unknown pattern x is projected through the learned memory, y = Mx, and a non-linear function (e.g., thresholding) produces the retrieved output pattern. If class labels are encoded as output vectors (e.g., one-hot) and the decision is taken from y by a winner-take-all rule, the same associative mechanism can be used as a classifier. This idea is exploited in modern associative classifiers (Velazquez-Rodriguez et al., 2020), which extend the classical Lernmatrix with a novel mathematical transformation that makes the matrix updates and recall rule suitable for supervised pattern classification rather than only for pattern completion.
Associative memories are relevant because the proposed n-SBC classifier was inspired by them. In n-SBC, training patterns are stored as rows in a kind of memory matrix. Then, given a test or unknown pattern, similarity is computed as the bitwise overlap under bipolar coding, which amounts to an affine transformation of an inner product, unlike classic associative memories that learn from projections. Thus, our novel model is conceptually linked to associative memories but implements a different representation (RBC codes) and a Hamming-based decision rule tailored to pattern classification.
2.3.2 Hamming distance and RCB in pattern classification
After an extensive documentary search, it was found that, throughout history, there have been very few attempts to create intelligent pattern classification algorithms based on the Hamming distance. Regarding the RBC code, no impactful work has been found; therefore, this proposal uses the Hamming distance and the RBC code simultaneously within the same pattern classifier algorithm. At the same time, the proposal's novelty and originality are ensured. The closest is a work using RBC codes, published in 2017 (Šarkovskis et al., 2017), the authors Šarkovskis, Jeršovs, Kolosovs and Grabs describe the functionality of a real-time classifier useful for the computation of statistical parameters of data streams, the detection of symbols of different modulation types and other applications where the fastest possible association of a sample of input signals with one of the predefined categories is required.
While explicit RBC and Hamming classifiers are rare in the literature, some frameworks and methodologies encode data into binary codes and perform comparisons, mainly in hashing and ECOC-style multiclass reduction. In the ECOC (error-correcting output codes) methodology, each class is assigned a binary codeword, and a bank of binary base learners (e.g., C4.5, SVM) is trained, one per code column, as described by Dietterich and Bakiri (1994). Then, to perform classification, the column outputs are concatenated, and the label is chosen as the nearest class codeword, typically via the Hamming distance. Thus, ECOC is an ensemble framework, not a classifier per se: it improves the underlying learners but does not replace them (Dietterich and Bakiri, 1994).
On the other hand, in learning-to-hash or supervised hashing, features are transformed into compact binary codes by a learned encoder, and classification is commonly implemented via Hamming space (e.g., k-NN or ranking codes); this establishes that distance is an effective similarity for large-scale prediction when inputs are binary encoded (Norouzi et al., 2012). Thus, supervised hashing is not a classifier per se but a representation-learning method whose downstream machine-learning models operate on the learned bits. This approach offers fast lookups once trained, but it introduces training complexity, and results depend on the codebook.
Our contribution to n-SBC differs from these strands in two quantitative ways. First, unlike supervised hashing, which learns codebooks and then delegates prediction to a classic classifier (e.g., k-NN), n-SBC uses a deterministic RBC mapping per feature and classifies by Hamming distance, removing the encoder learning stage while preserving fast bit-wise comparisons. In supervised hashing, performance depends on the learned encoder; in n-SBC, performance hinges on the RBC representation and Hamming aggregation. Second, whereas ECOC emphasizes maximizing inter-class Hamming separations between class codewords and requires training a bank of binary base learners (e.g., C4.5, SVM), thus acting as a framework rather than a classifier per se, n-SBC treats the entire RBC binary string instance as the object of comparison, performing instance Hamming matching rather than decoding to a fixed class codeword. In short, both models are enabling methods that rely on baseline models (C4.5, SVM, k-NN, etc.), whereas n-SBC is the classifier itself. Together with our operational unification of RBC and Hamming, these distinctions place n-SBC at a different point in the design space (Xiao et al., 2022). A summary of the main differences is shown in Table 3.
To make these differences concrete, a compact comparison table covering code construction and decision rule, along with a small ablation replacing RBC with a fixed-width standard binary encoding to isolate RBC's contribution. As discussed, RBC improves n-SBC because adjacent codes differ by only one bit, preserving similarity between neighboring values; in contrast, standard binary encodings may flip multiple bits between consecutive values, distorting local neighborhoods and weakening bit-wise interpretability.
Therefore, unlike ECOC and supervised hashing, and unlike classical k-NN in feature space, n-SBC is an associative classifier whose decision rule operates directly in Hamming space.
3 Our proposal model
In this section, the main idea of the N-Similarity Binary Classifier (n-SBC) algorithm is explained, along with its operation; the learning phase and, finally, the classification phase of the proposed algorithm are addressed. The proposed algorithm is primarily based on the Hamming string similarity method and the reflected binary code (RBC) encoder, also known as the Gray Code, both of which are fundamental components of the model. The purpose of this study is to improve the performance of associative approach classifiers across several medical datasets to enhance disease detection.
To address the issue of missing values and categorical data, our proposed method requires preprocessing the dataset to address this complexity beforehand. To handle missing values, the classic imputation method was applied, replacing missing values with the mean for numerical data and the mode for categorical data. This resulted in datasets without missing values when present. Finally, the categorical variables were converted using the classic label encoding method, which assigns each category a unique numeric value.
Then, the RBC method is applied to the entire dataset. In this case, every feature of the input patterns xμ is converted to their equivalent binary RBC code to obtain a p-dimensional binary string, where p represents the maximum length of the largest converted value, denoted as .
In order to obtain a single binary string, we concatenate each transformed feature together, expressed as follows:
Let us assume that there is a dataset D, divided into two subsets: L and T, for learning and testing, respectively.
3.1 Learning phase for the proposed approach
The learning phase of the n-SBC model has only one step. It consists of creating a memory matrix, denoted by M, which contains every transposed binary string pattern of the learning dataset L, generated previously by applying the RBC code to each pattern. Finally, on the matrix M, each element corresponds to the entire binary string representation of bμ, expressed as follows:
3.2 Classification phase for the proposed approach
The classification phase of the n-SBC model has four stages; the first is the calculation of the Hamming Distance between the unknown pattern xω to each pattern of the dataset L. To calculate it, first let us assume that the unknown pattern has already undergone the RBC transformation, yielding bω. Therefore, the Hamming distance, H(bω, bμ), represents the number of positions at which the corresponding bits are different. The above is expressed as follows:
where u is the dimensionality of the patterns. represent the j-th elements of the pattern bω and the training dataset pattern bμ.
This first step resulted in a vector distance, denoted as Z, which contains the result of the subtraction of the cardinality per dataset, denoted as u, with the computed Hamming distance of each pattern for the dataset L to the interested pattern bω. The dimension of Z is equivalent to the cardinality of the dataset L, we can represent it with the following expression:
The second stage of the classification phase consists of handling the generated vector Zω to determine the class.
First, let C be the set of all the classes, such that: C = {k1, k2, …, kc}, where c is the number of classes. Then, let us introduce Ki to denote the number of patterns present within the i-th class, expressed as Ki = |ki|, ∀i ∈ {1, …, c}. Now, we determine the smallest pattern count across all the classes, termed Kmin, such as follows:
Subsequently, for any integer n satisfying 1 ≤ n ≤ Kmin, we extract the n-th largest component from the vector Zω of each class, represented as . The hyperparameter n controls how many of the largest components are aggregated, and different values of n correspond to different versions of the classifier. Finally, a vector yn is created by applying a sum to the selected n-th largest components, therefore, yn is calculated by the following expression:
where in this case s represents the number of samples of each i-class in the dataset. The third step consists of assigning to the unknown pattern xω his corresponding class yω. For that, we update the vector yn with the following rule:
Finally, the fourth stage consists of calculating the predicted class of the unknown pattern xω using the one-hot vector created in stage three. Therefore, the class is assigned based on the position of the hot value, which indicates the predicted class, because each row of the vector corresponds to a class in the dataset. Meeting the following expression .
One of the advantages of the n-SBC is that it aggregates only the top similar n components per class, so additional majority of samples do not grow a class's evidence unboundedly. Besides, the RBC encoding preserves similarity structure (adjacent numeric values differ by one bit), so compact minority clusters remain coherent in Hamming space and can dominate the selected top n samples. Consequently, decisions are driven by local match quality rather than by class prevalence, thereby mitigating the typical bias toward the majority class. This can enhance model performance with imbalanced complexity data.
Regarding the scope of applicability, n-SBC tends to perform well when classes exhibit locally coherent neighborhoods in feature space and when similarity is meaningfully captured by RBC. It may underperform when features are highly non-monotonic or noisy, when classes strongly overlap, or when B is inflated by many irrelevant bits.
3.3 Example of the train and classification phase for n-SBC
Below, a simplified example of the operation process of the learning and classification phases of our proposed classifier, the n-SBC model, is presented in detail. The patterns used for this practical example are detailed, where x1 and x2 belong to class A, while patterns x3, x4 and x5 belong to class B.
After applying the reflected binary code (RBC) and in order to maintain a column vector when concatenating the binary strings obtained, the following patterns result:
Following the Equation 2, the matrix M is created, which contains every transposed binary string representation pattern that will be handled in the classification phase, in this case, is expressed as follows:
At this point, the learning phase is complete. We have all the binary strings from the learning dataset ready to manipulate and proceed with the following steps for inference. Then, for the classification phase, the vector Zω is created based on the Hamming distances of each pattern and the unknown pattern xω. However, before obtaining the distance matrix, we define xω as follows:
Therefore, the Zω is denoted as follows. In this case, u = 6 because the dimensionality of each pattern is six.
At this stage, we must define the value of n, which in this example we define as n = 2. Having established the necessary parameters, we instantiate the vector yn following Equation 11.
where the n largest components for each class are summed to create the column vector yn. These components correspond to positions 1 and 2 in class A, and to positions 3, 4, and 5 in class B. Therefore, since the components with a higher Zω vector, which means they have greater similarity to the unknown pattern, suggest that they are similar to those samples that belong to class B. This information can be used to clarify the model's explainability. Finally, based on the rule defined previously in Equation 12, we update the vector yn obtaining yω.
In this example, due to the result of the one-hot encoding vector, we can see the value of 1 in the second position, indicating that the pattern xω belongs to the second-class B.
To understand the explainability and the reason why the unknown pattern xω was classified as class B. Considering the unknown pattern xω after RBC conversion is: {110010}. Table 4 illustrates the samples and features that influenced the decision of the n-SBC model.
Since n = 2, the two closest samples from each class are selected, which means that bω is classified as class B because they are very similar to patterns b4, and b5, which belong to class B. Moreover, since in the string bω and bi, in this case, each feature of the dataset is represented by 3 bits of the vector; it can be observed that the pattern bω is similar to bω because it matches with the second feature, and it is similar to b5 because it matches the first feature in totality. In this way, we can understand why the model decided to classify this pattern into its corresponding class.
4 Results and discussion
In this part, we present the detailed analysis of the experimental stage of our proposed algorithm against well-known state-of-the-art classification models. Subsection 4.1 describes the dataset selected in the experimental stage. Subsection 4.2 explains the validation method used, while 4.3 describes the performance measures. Subsection 4.4 shows the results obtained using the experimental methods and metrics described, and subsection 4.5 discusses the statistical significance results comparison.
4.1 Datasets
For the experimental phase of the present work, 20 datasets were selected, each representing a variety of diseases, with a focus on chronic conditions.
These data sets were mainly obtained from three widely known repositories: the KEEL repository (available at https://sci2s.ugr.es/keel/index.php), the UCI Machine Learning repository (accessible at https://archive.ics.uci.edu/ml/index.php), and the Kaggle repository (found at https://www.kaggle.com/datasets). To facilitate a deeper understanding, a complete description of each selected data set has been compiled. This compilation is summarized in Table 5, which provides information on the dataset's features, including the nature of the diseases it represents, the data structure, and the class imbalance index.
The imbalance index for each dataset was calculated as follows:
In the following, we provide a brief description of the selected datasets.
Appendicitis: This dataset was collected at https://www.kaggle.com/datasets/timrie/appendicitis from the Kaggle repository. The dataset comprises seven medical measures for 106 patients, with classes indicating whether each patient has appendicitis. Kaggle Snapshot: appendicitis/timrie, downloaded 2024-09-18.
Exasens COPD: This data set aims (based on demographic information from saliva) to classify patients into four classes according to their membership: chronic obstructive pulmonary disease (COPD), asthma, respiratory infections, and completely healthy patients. The dataset was collected from the UCI Machine Learning Repository at https://archive.ics.uci.edu/ml/datasets/Exasens. Downloaded 2024-06-14.
Acute Inflammations D1 and Acute Inflammations D2: These datasets are from a study aimed at detecting two urinary system diseases. Both datasets were obtained from the UCI Machine Learning repository at https://archive.ics.uci.edu/ml/datasets/Acute+Inflammations. Downloaded 2024-03-14.
ACPs Lung Cancer: This dataset was obtained from the UCI repository at https://archive.ics.uci.edu/ml/datasets/Anticancer+peptides, which contains information on peptides (amino acid codes) and their anticancer activity in lung cancer cell lines. Downloaded 2024-03-14.
Vertical Column: This dataset aims to detect if a patient has some vertebral column disease. It was recovered from the UCI Machine Learning repository at http://archive.ics.uci.edu/ml/datasets/vertebral+column. In vertebral Column 2C, the classes Disk Hernia and Spondylolisthesis were merged into a single class, labeled Abnormal. Downloaded 2024-03-14.
Contraceptive: This dataset was collected from the UCI Machine Learning repository at http://archive.ics.uci.edu/dataset/30/contraceptive+method+choice. It is used to predict the current contraceptive method from demographic and socioeconomic information. Downloaded 2024-03-14.
Cryotherapy: This dataset was collected from the UCI Machine Learning repository at https://archive.ics.uci.edu/dataset/429/cryotherapy+dataset, which contains treatment outcomes for 90 patients who underwent cryotherapy. It has two classes: successful and unsuccessful. Downloaded 2024-03-14.
Dermatology: This dataset was obtained from the UCI Machine Learning repository at https://archive.ics.uci.edu/dataset/33/dermatology, whose main aim is to determine the type of Eryhemato-Squamous Disease based on 34 patient attributes. Downloaded 2024-03-14.
Hepatitis: This dataset aims to detect hepatitis using simple tabular data from patients, most of whom have categorical data. Furthermore, the dataset has two classes and was collected from the UCI Machine Learning repository at http://archive.ics.uci.edu/dataset/46/hepatitis. Downloaded 2024-02-03.
Mammographic Masses: This dataset aims to distinguish between benign and malignant mammographic masses using BI-RADS attributes and patient age. The dataset was collected from https://archive.ics.uci.edu/dataset/161/mammographic+mass, in the UCI Machine Learning repository. Downloaded 2024-01-21.
Wisconsin: This dataset was collected from the UCI Machine Learning repository at https://archive.ics.uci.edu/ml/datasets/breast+cancer+wisconsin+(diagnostic), which describes cases from a study conducted at the University of Wisconsin Hospitals in Madison involving patients who had undergone surgery for breast cancer. The classification task is to determine if the detected tumor is benign or malignant. Downloaded 2024-02-24.
HCC Survival: This dataset was obtained from https://archive.ics.uci.edu/dataset/423/hcc+survival, in the UCI Machine Learning repository. It contains real clinical data from 165 patients diagnosed with HCC, with the aim of predicting 1-year survival after diagnosis. Downloaded 2024-03-28.
Autism adolescent and Child: These datasets were collected from the UCI Machine Learning repository at https://archive.ics.uci.edu/dataset/420/autistic+spectrum+disorder+screening+data+for+adolescent and https://archive.ics.uci.edu/dataset/419/autistic+spectrum+disorder+screening+data+for+children, respectively. The idea of both datasets is to detect Autistic Spectrum Disorder. Downloaded 2024-04-02.
Survey Lung Cancer: The classification task in this dataset is to determine whether a given patient has lung cancer, based on variables collected via a survey. The set was obtained from the Kaggle repository at https://www.kaggle.com/mysarahmadbhat/lung-cancer. Kaggle Snapshot: Lung Cancer/Mysar Ahmad Bhat, downloaded 2024-04-13.
Breast Cancer Coimbra: This dataset was collected from https://archive.ics.uci.edu/dataset/451/breast+cancer+coimbra, in the UCI Machine Learning Repository. The dataset comprises clinical features from 64 patients. Downloaded 2024-02-22.
Saheart: This dataset aims to detect patients with heart diseases but was built for Stanford University and was collected at https://web.stanford.edu/~hastie/ElemStatLearn//datasets/SAheart.data. Downloaded 2024-02-24.
Cirrhosis: This dataset comprises 17 clinical features for predicting patient survival in patients with liver cirrhosis, collected from the UCI Machine Learning repository at https://archive.ics.uci.edu/dataset/878/cirrhosis+patient+survival+prediction+dataset-1. Downloaded 2024-02-22.
Multiple Sclerosis: The classification task in this dataset is to detect multiple sclerosis using patient information, such as personal data, symptoms, and metrics from medical tests. The dataset was collected from https://www.kaggle.com/datasets/desalegngeb/conversion-predictors-of-cis-to-multiple-sclerosis/data, the Kaggle repository. Kaggle Snapshot: Multiple Sclerosis Disease/A Legacy Grandmaster!, downloaded 2024-04-11.
4.2 Validation methods
In this section, we describe the validation method used in the experimentation stage. To obtain reliable results when measuring classifier performance during the experimentation stage, it is necessary to have previously implemented a validation method that divides the original dataset into two sets: a test set and a learning set.
One of the most widely used methods is k-fold cross-validation, which randomly divides the original set into k equal-sized subsets (folds), using one fold as the test set and the rest as the training set. This process is repeated k times in order to use all folds at least once as test sets (Wong, 2015; Sarker, 2021). On the other hand, there is a stratified version of this validation method, called stratified k-fold cross-validation, which is highly recommended for data sets with class imbalance, since it attempts to preserve approximate class proportions within each fold. In this way, the test sets created in each iteration present as much as possible the class distribution of the original set, which helped mitigate errors caused by class bias (Derrac et al., 2015; Nakatsu, 2020). Figure 1 shows the operations of the stratified k-fold cross-validation when k = 5.
Given the class-imbalanced datasets used in the current study, stratified k-fold cross-validation with k = 10 has been employed to maintain approximately equal proportions of patterns per class across folds.
4.3 Performance measures
The evaluation of classifier performance is a crucial area of interest in specialized literature. The most popular and naturally simple way to measure performance is to use the accuracy metric, which calculates the percentage of patterns in the test set that are correctly classified; that is, it counts the total number of correctly classified patterns with respect to the total number of patterns. However, there is a way to more completely represent the results of the classifier's performance, which is called a confusion matrix, as shown in Figure 2, which consists of four possible cases within a two-class classification problem (García et al., 2010a), where each cell in the confusion matrix represents TP (true positive), TN (true negative), FP (false positive), and FN (false negative).
As mentioned above, one of the most popular metrics for measuring classifier performance is accuracy. In the case of bi-class problems, and using the confusion matrix as a basis, the metric can be expressed as in the equation:
However, more robust metrics have emerged in the literature to mitigate the limitations of the accuracy metric, which is not suitable for class-imbalanced datasets, a common data complexity mainly found in medical datasets. This data complexity harms the evaluation of the classifier's performance, yielding metrics that do not truly reflect the algorithm's capacity (López et al., 2013).
First, the sensitivity metric will be described, which measures the probability that the classifier returns a positive result when the instance is a true positive. The sensitivity metric can be expressed as follows (García et al., 2010b).
On the other hand, there is another crucial metric, the counterpart of the sensitivity metric: the specificity metric. This metric estimates the probability that the classifier will return a negative result when the instance is actually negative (García et al., 2010b).
There are different metrics for different purposes, such as the area under the ROC curve (AUC), precision, F1 score, and balanced accuracy (BA), but the majority of them are calculated from the confusion matrix (García et al., 2010b). Because the datasets selected for this study exhibit class imbalance, it was decided to use the Balanced Accuracy (BA) performance metric, which is recommended for such cases (López et al., 2013; García et al., 2010b). The BA metric is calculated from the performance metrics Sensitivity and Specificity, which represent the average of both measures.
On the other hand, the value of BA in multi-class datasets, for k classes, is calculated as follows:
where Ti is the number of patterns correctly classified in class i, and Ni represents the total number of patterns within the dataset of class i.
Example. Figure 3 shows a confusion matrix for an unbalanced dataset, with 170 patterns in class A and 30 in class B. Therefore, the similar dataset has a very severe class imbalance; its imbalance index is IR = 5.6.
For class i ∈ {A, B, C} sensitivity (Ti/N):
In this example, the Balance Accuracy (BA) value of the confusion matrix is as follows:
4.4 Time complexity analysis
Table 6 compares the time complexities of the classification algorithms used in the present study and of the proposed n-SBC model.
Notation. nsv: Number of support vectors in SVM; T: Number of trees in Random Forest; H: Number of hidden units in MLP; I: Number of epochs in MLP; |L|: Total number of patterns (instances) in the training dataset; C: Number of classes in the data set; Xω: Unknown pattern (test) to be classified; d: number of features; bi: RBC bit-length of feature i; (total bits per encoded pattern); B: Length of each pattern in the binary string generated by the RBC encoder.
4.5 Classification results
Table 7 compares the performance of the proposed algorithm with that of different classifiers across the 20 datasets described earlier. The algorithms used for comparison were run in Weka version 3.8.2, using the tool's default hyperparameters. The results of the n-SBC algorithm were obtained using MATLAB R2021b with a random seed of 1. For the experimental process, we evaluated two pre-specified SBC variants with n ∈ {3, 5}. These values were pre-selected once from a preliminary sweep n ∈ {1, 2, 3, 4, 5} using training-only validation and were then held fixed across all datasets. To ensure a fair comparison, the study does not cherry-pick the best n values in the classification results; instead, those are explicitly excluded from the Friedman and Holm statistical tests to avoid inflating the number of comparisons.
No preprocessing other than handling missing values and converting categorical values to numeric was applied, as explained in Section 3. No samples were removed from the datasets, nor were synthetic samples added; they also kept the original sizes and format.
The proposed algorithm achieved competitive performance across nine of the twenty-one datasets. For example, it performed well on Acute Inflammation d1 and d2, ACP lung cancer, contraceptive use, mammographic masses, HCC survival, breast cancer Coimbra, cirrhosis, and multiple sclerosis.
Furthermore, Table 7 shows some cases where the classifiers achieved 1 on the balanced accuracy metric. This indicates that it was perfect, i.e., the classifier made zero errors. Thus, if we count the frequency at which classifiers obtained these cases, our proposed model was one of the highest-performing BA in both versions (3-SBC and 5-SBC), receiving it in 3 out of 21 datasets.
Similarly, the algorithm that performed best across the majority of datasets was our proposed 3-SBC model, which was the best in 9 of 20 datasets, followed by our other model, 5-SBC, which was the best in 7 of 20 datasets.
Nevertheless, datasets with high data complexity that obtained inadequate scores were Cirrhosis, Saheart, HCC Survival, and Contraceptive, among which our proposed models achieved the highest performance in 3 out of 5 cases. This happens due to the No Free Lunch theorem; therefore, it is expected that our proposed models will not be the best-performing classifiers across all datasets. This theorem indicates that no classifier is capable of being the best on all types of problems (Wolpert and Macready, 1997; Adam et al., 2019). Furthermore, the classifier with the best performance also performed poorly, such as Saheart, which achieved 0.658 on SMO.
However, in favor of our proposal, it can be noted that, in most cases, the performances of the 3-SBC classifier do not vary overly from the high balanced accuracy values obtained by other classifiers; such is the case of the Survey lung cancer, Dermatology, and Wisconsin datasets in which our proposed model 3-SBC obtained very similar results against the best models in those cases, such as SMO or Naïve Bayes.
4.6 Statistical analysis
Comparing various machine learning algorithms and selecting a final model or algorithm as the winner is a common practice in machine learning, model research, and applications. Models in relation to a set of experiments are evaluated using a validation method, e.g., k-fold cross-validation or leave-one-out cross-validation (a particular case of k-fold cross-validation where k equals the number of instances in the dataset), and the results are directly compared by calculating a performance measure. While this is a simple and somewhat intuitive approach, it is difficult to determine whether a difference is due to the algorithm's real capability or a statistical fluke.
It is crucial to distinguish genuine performance differences from statistical flukes. Therefore, it is necessary to apply statistical hypothesis testing, which addresses this issue by quantifying the probability of observing score differences under the null hypothesis that scores are drawn from the same distribution. Rejection of this null hypothesis indicates that the observed differences are statistically significant, rather than due to chance.
In this context, to conduct a more reliable comparative analysis, it was proposed to use Friedman's test (Friedman, 1937) to determine whether there are significant differences in the yields observed during the experiment.
Table 8 shows the performance obtained by the different classification algorithms proposed. After performing Friedman's statistical test, the null hypothesis was rejected at the 95% confidence level (p-value = 0.000516), indicating statistically significant differences among the classifiers.
The proposed models (5-SBC and 3-SBC) rank first in the Friedman mean rank calculation concerning the remaining seven algorithms, while the k-NN family algorithms rank last in the Friedman mean rank table.
On the other hand, a post-hoc test, the Holm test (Holm, 1979), was applied. The results in Table 9 reject the hypothesis at an adjusted p-value of ≤ 0.05. Therefore, significant performance differences between the two versions of the proposed algorithm and the remaining state-of-the-art algorithms used in the study are demonstrated. In particular, it can be observed that, considering the best algorithm according to the Friedman test, the 5-SBC algorithm, it has significant differences above the 95% confidence level for the 1-NN and 3-NN algorithms; on the other hand, the SVM, MLP, and Naïve Bayes algorithms obtained p-values (although higher than the corrected threshold) that indicate a possible significant difference to the 5-SBC model, which could be interpreted as marginal evidence in exploratory contexts.
After presenting the experiments, it was observed that the proposed algorithm obtained competitive results. This conclusion is supported by statistically significant differences in the n-SBC algorithm's observed performance across two of the seven selected classifiers on the same set of classification datasets.
Consequently, the results corroborate the hypothesis that the proposed n-SBC algorithm is indeed competitive for classification and disease prediction, as the majority of the datasets used focus on detecting different diseases.
5 Conclusion and future research
In this research work, a new proposed model, n-SBC (n Similarity Binary Classifier), was presented, along with an experimental analysis to verify its effectiveness against other state-of-the-art algorithms on datasets related to medicine.
Similarly, the advantages of the proposed model were described, as were its simplicity, explainability, and its ability to address imbalance, a data complexity that is very common in the literature.
The detailed results in this research, presented in Section 4, highlight the capacity of the proposed algorithm, specifically the version 5-SBC, due to its competitive performance compared to other popular classification algorithms in the literature. Similarly, this research explores and presents in detail a new approach that uses the similarity between binary strings as the basis for a machine learning model while maintaining simplicity and effectiveness. Above all, the proposed novel algorithm promotes the research and application of explainable AI, which is of great contribution to specific areas, such as health or finance.
The proposed model has a limitation in handling pattern cardinality, as it converts patterns to binary strings, which can increase computational complexity during classification. Therefore, as future research, it is proposed to develop a method or pursue a completely new approach that can solve the problem identified in the proposed model while maintaining its simplicity, explainability, and performance. It is under consideration for implementation and demonstrates a novel similarity measure that can improve the model's performance while maintaining the algorithm's simplicity and explainability. On the other hand, an important objective is to apply the proposed n-SBC model to image classification tasks, specifically to medical images (x-ray mainly), due to its ease in preprocessing to adapt it and be able to use the n-SBC model with evolutionary algorithms or metaheuristic processes, with the goal of optimizing the model's performance. Another aspect to consider in future studies is analyzing the model's behavior on datasets with outliers to evaluate its robustness and adaptability to more complex, noisy scenarios. Finally, we plan to extend n-SBC by incorporating the three-way decision (3WD) rule based on model margin to explicitly handle classification uncertainty, compare granular 3WD variants, and report risk-coverage improvements on medical data.
Data availability statement
Publicly available datasets were analyzed in this study. This data can be found here: https://archive.ics.uci.edu/datasets.
Author contributions
OV-G: Visualization, Writing – review & editing, Investigation, Writing – original draft. AA-P: Writing – original draft, Writing – review & editing, Formal analysis, Methodology. CY-M: Writing – review & editing, Writing – original draft, Conceptualization, Supervision.
Funding
The author(s) declared that financial support was not received for this work and/or its publication.
Conflict of interest
The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
The author CY-M declared that they were an editorial board member of Frontiers, at the time of submission. This had no impact on the peer review process and the final decision.
Generative AI statement
The author(s) declared that generative AI was not used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/frai.2025.1610856/full#supplementary-material
References
Abdullah, D. M., and Abdulazeez, A. M. (2021). Machine learning applications based on Svm classification a review. Qubahan Acad. J. 1, 81–90. doi: 10.48161/qaj.v1n2a50
Adam, S. P., Alexandropoulos, S-. A. N., Pardalos, P. M., and Vrahatis, M. N. (2019). “No free lunch theorem: a review,” in Approximation Optimization: Algorithms, Complexity Applications (Berlin: Springer), 57–82.
Agrell, E., Lassing, J., Strom, E. G., and Ottosson, T. (2004). On the optimality of the binary reflected gray code. Eee Trans. Inform. Theor. 50, 3170–3182. doi: 10.1109/TIT.2004.838367
Ahsan, M. M., Luna, S. A., and Siddique, Z. (2022). Machine-learning-based disease diagnosis: a comprehensive review. Healthcare. 10:541. doi: 10.3390/healthcare10030541
Amygdalos, I., Müller-Franzes, G., Bednarsch, J., Czigany, Z., Ulmer, T. F., Bruners, P., et al. (2023). Novel machine learning algorithm can identify patients at risk of poor overall survival following curative resection for colorectal liver metastases. J. Hepato-Biliary-Pancreat. Sci. 30, 602–614. doi: 10.1002/jhbp.1249
Anderson, J. A. (1972). A simple neural network generating an interactive memory. Math. Biosci. 14, 197–220. doi: 10.1016/0025-5564(72)90075-2
Bhargavi, P., and Jyothi, S. (2009). Applying naive bayes data mining technique for classification of agricultural land soils. J Int. J. Comput. Sci. Netw. Secur. 9, 117–122.
Bhat, G. S., and Savage, C. D. (1996). Balanced gray codes. Electron. J. Combinat. 3:R25. doi: 10.37236/1249
Bian, K., and Priyadarshi, R. (2024). Machine learning optimization techniques: a survey, classification, challenges, and future research issues. Arch. Comput. Methods Eng. 31, 4209–4233. doi: 10.1007/s11831-024-10110-w
Bookstein, A., Kulyukin, V. A., and Raita, T. (2002). Generalized Hamming Distance. Inf. Retr. Boston. 5, 353–375. doi: 10.1023/A:1020499411651
Bui, D. T., Khosravi, K., Tiefenbacher, J., Nguyen, H., and Kazakis, N. (2020). Improving prediction of water quality indices using novel hybrid machine-learning algorithms. Sci. Total Env. 721:137612. doi: 10.1016/j.scitotenv.2020.137612
Cortes, C. (1995). Support-vector networks. Machine Learn. 20, 273–297. doi: 10.1023/A:1022627411411
Costa, V. G., and Pedreira, C. E. (2023). Recent advances in decision trees: an updated survey. Artif. Intell. Rev. 56, 4765–4800. doi: 10.1007/s10462-022-10275-5
Cover, T., and Hart, P. (1967). Nearest Neighbor Pattern Classification. J IEEE Transac. Inform. Theor. 13, 21–27. doi: 10.1109/TIT.1967.1053964
De Sa, J. M. (2012). Pattern Recognition: Concepts, Methods and Applications. Berlin: Springer Science and Business Media.
Derrac, J., Garcia, S., Sanchez, L., and Herrera, F. (2015). Keel data-mining software tool: data set repository, integration of algorithms and experimental analysis framework. Mult. Valued Logic. Soft. Comput. 17, 255–287.
Dietterich, T. G., and Bakiri, G. (1994). Solving multiclass learning problems via error-correcting output codes. J. Artif. Intell. Res. 2, 263–286. doi: 10.1613/jair.105
Doran, R. W. (2007). The gray code. J. Univ. Comput. Sci. 13, 1573–1597. doi: 10.3217/jucs-013-11-1573
Dua, D., and Graff, C. (2019). Uci machine learning repository [Http://Archive. Ics. Uci. Edu/Ml]. Irvine, Ca: University Of California, School Of Information And Computer Science. IEEE Transac. Pattern Anal. Machine Intell. 1, 1–29.
Duda, R. O., Hart, P. E., and Stork, D. G. (2001). Pattern Classification, Second Edition Edition. Hoboken, NJ: John Willey and Sons Inc.
Friedman, M. (1937). The use of ranks to avoid the assumption of normality implicit in the analysis of variance. J. Am. Stat. Assoc. 32, 675–701. doi: 10.1080/01621459.1937.10503522
Gaitanis, N., Kapogianopoulos, G., and Karras, D. (1993). “Pattern classification using a generalised hamming distance metric,” in Proceedings Of 1993 International Conference On Neural Networks (Ijcnn-93-Nagoya, Japan) (New York, NY: IEEE), 1293–1296.
Galli, C., Cusano, C., Meleti, M., Donos, N., and Calciolari, E. (2024). Topic modeling for faster literature screening using transformer-based embeddings. Metrics 2:2. doi: 10.3390/metrics1010002
García, S., Fernández, A., Luengo, J., and Herrera, F. (2010a). Advanced nonparametric tests for multiple comparisons in the design of experiments in computational intelligence and data mining: experimental analysis of power. Inf. Sci. 180, 2044–2064. doi: 10.1016/j.ins.2009.12.010
García, V., Mollineda, R. A., and Sánchez, J. S. (2010b). “Theoretical analysis of a performance measure for imbalanced data,” in 2010 20th International Conference on Pattern Recognition (New York, NY: IEEE), 617–620. doi: 10.1109/ICPR.2010.156
Goodall, W. (1951). Television by pulse code modulation. Bell Syst. Tech. J. 30, 33–49. doi: 10.1002/j.1538-7305.1951.tb01365.x
Hall, M., Frank, E., Holmes, G., Pfahringer, B., Reutemann, P., Witten, I. H., et al. (2009). The weka data mining software: an update. Acm Sigkdd Explor. Newslett. 11, 10–18. doi: 10.1145/1656274.1656278
Hissou, H., Benkirane, S., Guezzaz, A., Azrour, M., and Beni-Hssane, A. (2023). A novel machine learning approach for solar radiation estimation. Sustainability 15:10609. doi: 10.3390/su151310609
Hoffmann, H. (2019). Sparse associative memory. Neural Comput. 31, 998–1014. doi: 10.1162/neco_a_01181
Holm, S. (1979). A simple sequentially rejective multiple test procedure. Scand. J. Statistics 6, 65–70. doi: 10.2307/4615733
Hopfield, J. J. (1982). Neural networks and physical systems with emergent collective computational abilities. Proc. Nat. Acad. Sci. 79, 2554–2558. doi: 10.1073/pnas.79.8.2554
Ibrahim, I., and Abdulazeez, A. (2021). The role of machine learning algorithms for diagnosing diseases. J. Appl. Sci. Technol. Trends 2, 10–19. doi: 10.38094/jastt20179
Janani, R., and Vijayarani, S. (2019). Text document clustering using spectral clustering algorithm with particle swarm optimization. J. Expert Syst. Appl. 134, 192–200. doi: 10.1016/j.eswa.2019.05.030
Kohonen, T. (1972). Correlation matrix memories. IEEE Transac. Comput. 100, 353–359. doi: 10.1109/TC.1972.5008975
Kumar, Y., Koul, A., Singla, R., and Ijaz, M. F. (2023). Artificial intelligence in disease diagnosis: a systematic literature review, synthesizing framework and future research agenda. J. Ambient Intell. Humanized Comput. 14, 8459–8486. doi: 10.1007/s12652-021-03612-z
LeCun, Y., Bengio, Y., and Hinton, G. (2015). Deep learning. Nature 521, 436–444. doi: 10.1038/nature14539
López, V., Fernández, A., García, S., Palade, V., and Herrera, F. (2013). An insight into classification with imbalanced data: empirical results and current trends on using data intrinsic characteristics. Inf. Sci. 250, 113–141. doi: 10.1016/j.ins.2013.07.007
Luna-Ortiz, I., Aldape-Pérez, M., Uriarte-Arcia, A. V., Rodríguez-Molina, A., Alarcón-Paredes, A., Ventura-Molina, E., et al. (2023). Parkinson's disease detection from voice recordings using associative memories. Healthcare 11:1601. doi: 10.3390/healthcare11111601
McCulloch, W. S., and Pitts, W. (1943). A logical calculus of the ideas immanent in nervous activity. Bull. Math. Biophys. 5, 115–133. doi: 10.1007/BF02478259
Misra, P., and Yadav, A. S. (2020). Improving the classification accuracy using recursive feature elimination with cross-validation. Int. J. Emerg. Technol. 11, 659–665.
Moreno-Ibarra, M-. A., Villuendas-Rey, Y., Lytras, M. D., Yáñez-Márquez, C., and Salgado-Ramírez, J-. C. (2021). Classification of diseases using machine learning algorithms: a comparative study. Mathematics 9:1817. doi: 10.3390/math9151817
Nakatsu, R. T. (2020). An evaluation of four resampling methods used in machine learning classification. IEEE Intell. Syst. 36, 51–57. doi: 10.1109/MIS.2020.2978066
Norouzi, M., Fleet, D. J., and Salakhutdinov, R. R. (2012). “Hamming distance metric learning,” in Advances in Neural Information Processing Systems, vol. 25 (NIPS 2012). New York, NY: Curran Associates, Inc.
Nozari, H., Ghahremani-Nahr, J., and Szmelter-Jarosz, A. (2024). “AI and machine learning for real-world problems,” in Advances in Computers. Amsterdam: Elsevier.
Pappalardo, F., Calonaci, C., Pennisi, M., Mastriani, E., and Motta, S. (2009). “Hamfast: fast hamming distance computation,” in 2009 WRI World Congress on Computer Science and Information Engineering (New York, NY: IEEE), 569–572.
Quinlan, J. R. (1990). Decision trees and decision-making. J. IEEE Transac. Syst. Man Cybernet. 20, 339–346. doi: 10.1109/21.52545
Rana, M., and Bhushan, M. (2023). Machine learning and deep learning approach for medical image analysis: diagnosis to detection. Multimedia Tools Appl. 82, 26731–26769. doi: 10.1007/s11042-022-14305-w
Rane, N., Choudhary, S. P., and Rane, J. (2024). Ensemble deep learning and machine learning: applications, opportunities, challenges, and future directions. Stud. Med. Health Sci. 1, 18–41. doi: 10.48185/smhs.v1i2.1225
Rasool, S., Husnain, A., Saeed, A., Gill, A. Y., and Hussain, H. K. (2023). Harnessing predictive power: exploring the crucial role of machine learning in early disease detection. Jurihum: Jurnal Inovasi Dan Humaniora 1, 302–315.
Rosenblatt, F. (1958). The perceptron: a probabilistic model for information storage and organization in the brain. Psychol. Rev. 65:386. doi: 10.1037/h0042519
Sarker, I. H. (2021). Machine learning: algorithms, real-world applications and research directions. SN Comput. Sci. 2:160. doi: 10.1007/s42979-021-00592-x
Šarkovskis, S., Jeršovs, A., Kolosovs, D., and Grabs, E. (2017). Encoder improvement for simple amplitude fully parallel classifiers based on grey codes. Procedia Eng. 178, 604–614. doi: 10.1016/j.proeng.2017.01.119
Sharifani, K., and Amini, M. (2023). Machine learning and deep learning: a review of methods and applications. World Inform. Technol. Eng. J. 10, 3897–3904. Available online at: https://ssrn.com/abstract=4458723 (Accessed March 10, 2024).
Shehadeh, A., Alshboul, O., Al Mamlook, R. E., and Hamedat, O. (2021). Machine learning models for predicting the residual value of heavy construction equipment: an evaluation of modified decision tree, Lightgbm, And Xgboost Regression. Autom. Constr. 129:103827. doi: 10.1016/j.autcon.2021.103827
Talib, S. A. (2018). New strategies for associative memories. Eng. Technol. J. 36, 207–212. doi: 10.30684/etj.36.2A.13
Vayadande, K. (2024). Innovative approaches for skin disease identification in machine learning: a comprehensive study. Oral. Oncol. Rep. 10:100365. doi: 10.1016/j.oor.2024.100365
Velazquez-Rodriguez, J-. L., Villuendas-Rey, Y., Camacho-Nieto, O., and Yanez-Marquez, C. J. M. (2020). A novel and simple mathematical transform improves the perfomance of lernmatrix in pattern classification. Mathematics 8:732. doi: 10.3390/math8050732
Willshaw, D. J., Buneman, O. P., and Longuet-Higgins, H. C. (1969). Non-holographic associative memory. Nature 222, 960–962. doi: 10.1038/222960a0
Wolpert, D. H., and Macready, W. G. (1997). No free lunch theorems for optimization. IEEE Transac. Evol. Comput. 1, 67–82. doi: 10.1109/4235.585893
Wong, T-. T. (2015). Performance evaluation of classification algorithms by K-fold and leave-one-out cross validation. Pattern Recognit. 48, 2839–2846. doi: 10.1016/j.patcog.2015.03.009
Xiao, W., Luo, Z., and Hu, Q. (2022). A review of research on signal modulation recognition based on deep learning. Electronics 11:2764. doi: 10.3390/electronics11172764
Yáñez-Márquez, C. (2020). Toward the bleaching of the black boxes: minimalist machine learning. IT Prof. 22, 51–56. doi: 10.1109/MITP.2020.2994188
Yang, G., and Ding, F. (2020). Associative memory optimized method on deep neural networks for image classification. Inf. Sci. 533, 108–119. doi: 10.1016/j.ins.2020.05.038
Zhang, L., Zhang, Y., Tang, J., Lu, K., and Tian, Q. (2013). “Binary code ranking with weighted hamming distance,” in Proceedings of the IEEE Conference Computer Vision Pattern Recognition, 1586–1593.
Zhang, S. (2021). Challenges in Knn classification. IEEE Transac. Knowl. Data Eng. 34, 4663–4675. doi: 10.1109/TKDE.2021.3049250
Zhu, F., Ma, S., Cheng, Z., Zhang, X-. Y., Zhang, Z., Liu, C-. L., et al. (2024). Open-world machine learning: a review and new outlooks. arxiv [Preprint] arxiv:2403.01759. Available online at: https://arxiv.org/abs/2403.01759 (Accessed March 4, 2024).
Keywords: binary similarity, classification algorithms, machine learning, medicine dataset, pattern classification, pattern recognition
Citation: Velazquez-Gonzalez O, Alarcón-Paredes A and Yañez-Marquez C (2026) Medical pattern classification using a novel binary similarity approach based on an associative classifier. Front. Artif. Intell. 8:1610856. doi: 10.3389/frai.2025.1610856
Received: 13 April 2025; Revised: 10 December 2025;
Accepted: 11 December 2025; Published: 14 January 2026.
Edited by:
Alaa Eleyan, American University of the Middle East, KuwaitReviewed by:
Reggie Gustilo, De La Salle University, PhilippinesVince Hooper, SPJ Global, United Arab Emirates
Suprativ Saha, JIS University, India
Copyright © 2026 Velazquez-Gonzalez, Alarcón-Paredes and Yañez-Marquez. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Cornelio Yañez-Marquez, Y3lhbmV6QGNpYy5pcG4ubXg=
Antonio Alarcón-Paredes