Impact Factor 2.638 | CiteScore 2.3
More on impact ›

Original Research ARTICLE

Front. Phys., 19 June 2020 | https://doi.org/10.3389/fphy.2020.00196

Predicting Sites of Epitranscriptome Modifications Using Unsupervised Representation Learning Based on Generative Adversarial Networks

  • 1Department of Electrical and Computer Engineering, The University of Texas at San Antonio, San Antonio, TX, United States
  • 2Greehey Children's Cancer Research Institute, University of Texas Health San Antonio, San Antonio, TX, United States
  • 3Department of Population Health Sciences, University of Texas Health San Antonio, San Antonio, TX, United States

Epitranscriptome is an exciting area that studies different types of modifications in transcripts, and the prediction of such modification sites from the transcript sequence is of significant interest. However, the scarcity of positive sites for most modifications imposes critical challenges for training robust algorithms. To circumvent this problem, we propose MR-GAN, a generative adversarial network (GAN)-based model, which is trained in an unsupervised fashion on the entire pre-mRNA sequences to learn a low-dimensional embedding of transcriptomic sequences. MR-GAN was then applied to extract embeddings of the sequences in a training dataset we created for nine epitranscriptome modifications, namely, m6A, m1A, m1G, m2G, m5C, m5U, 2′-O-Me, pseudouridine (Ψ), and dihydrouridine (D), of which the positive samples are very limited. Prediction models were trained based on the embeddings extracted by MR-GAN. We compared the prediction performance with the one-hot encoding of the training sequences and SRAMP, a state-of-the-art m6A site prediction algorithm, and demonstrated that the learned embeddings outperform one-hot encoding by a significant margin for up to 15% improvement. Using MR-GAN, we also investigated the sequence motifs for each modification type and uncovered known motifs as well as new motifs not possible with sequences directly. The results demonstrated that transcriptome features extracted using unsupervised learning could lead to high precision for predicting multiple types of epitranscriptome modifications, even when the data size is small and extremely imbalanced.

Introduction

Epitranscriptome is an exciting emerging area that studies modifications in transcripts. The insurgent interest is largely fueled by the recent discovery of widespread N6-methyl-adenosine (m6A) methylation in mammalian mRNAs [1, 2], which has been shown to play important regulatory roles in every stage of RNA metabolism and be involved in many diseases. Besides m6A, many other types of modifications are found to exist in the eukaryotic transcriptome. While some of them including N1-methyladenosine (m1A), 5-hydroxymethylcytosine (hm5C), 5-methylcytidine (m5C), 2′-O-methylation (2′-O), and pseudouridine (Ψ) are found widespread, other types, such as dihydrouridine (D) and m2G have only a hundred sites discovered thus far. These exciting findings have spurred intense research to identify transcriptome modifications in different cells and to decipher their roles in regulating various biological processes [3].

We consider in this paper the prediction of transcriptome modification sites from transcript sequences. This problem is naturally a supervised learning task, which aims to train predictive models for each type by using labeled positive and negative modification sites. There is a large collection of algorithms for predicting m6A sites from mRNA sequences [410], most notably SRAMP. However, such predictive algorithms for other modifications are still scarce because training robust models for these modification sites faces several challenges. First, training for modification types with scarce labeled samples suffers from significant overfitting, making the model incapable of learning true methylation-specific sequence features from random noise patterns. Second, training data for all modifications suffer significant class imbalance, a common challenge in most of the genomics applications, where only a small percentage of transcriptome nucleotides are true methylation sites and the majority of them are negatively labeled unmodified sites. Unfortunately, traditional supervised learning tends to treat these extremely small numbers of instances as noise and again fails to learn desired methylation-specific sequence features from the positive samples [11]. Third, predicting modification sites of these different types altogether imposes a greater challenge as different types are likely to share similar biological sequence patterns, making them less distinguishable from each other. One viable solution to address these challenges is to take advantage of the vast unlabeled part of the transcriptome sequences with unsupervised representation learning to learn transcriptome-wide sequence features as a whole. We assume that the unlabeled part of the transcriptome could contain unidentified modification sites and/or modification-related functional sites [e.g., RNA-binding protein (RBP)-binding sites]. Therefore, the unsupervised learning of transcriptome sequences may help capture modification-related features, which are difficult to learn otherwise by a supervised approach using only labeled sequences. We intend to learn these features in an unsupervised setting to leverage the supervised learning using labeled data such that the classifiers trained using only a few labeled examples can generalize to predict modification sites robustly. Unsupervised learning methods have recently attracted an influx of research interest, especially in the field of bioinformatics. For example, restricted Boltzmann machines have been applied for unsupervised pre-training of neural networks that were later used to initialize the supervised learning of protein 3D structures [12, 13] and amino acid contacts [14]. Besides, Asgari et al. [15] proposed a low-dimensional representation method for protein sequences, which is inspired by the widely known Word2Vec model of natural language processing. However, none of these methods have been trained to extract features from RNA sequences.

In pursuit of addressing the challenges described above, we propose an unsupervised feature construction approach based on generative adversarial networks (GANs) [16]. Our method, aptly known as MR-GAN, predicts multiple types of modification sites in RNA using GAN-based unsupervised feature learning. This method delves deep into the largely unexplored area of low-dimensional representation of RNA sequences and demonstrates the usefulness of the unsupervised feature learning in handling some of the most difficult problems in the intersection of machine learning and bioinformatics (small and imbalance dataset). Though the idea of GAN has been fervently researched in the computer vision domain in recent years [1719], this architecture is largely uncharted territory for bioinformatics except a few [20, 21]. The original GAN framework [16], which consists of two modules known as generator and discriminator, was designed to learn a generative distribution of data through a two-player minimax game. While the generator's goal is to “fool” the discriminator by producing samples that are as close to the real data distribution as possible, the discriminator strives to be not fooled by correctly classifying between real and fake data. The adversarial framework has been shown to provide a superior loss to other traditional ones based on mean square error or mutual information for learning data distributions.

Despite the success of GANs in mimicking the underlying data distribution, GANs cannot be readily applied to learn the abstract representation of data due to the lack of an efficient inference module. Hence, in order to learn the low-dimensional feature representation of RNA sequences using a GAN-based framework, we employed the adversarial learned inference (ALI) model [22], which jointly learns an encoder network and a decoder network using an adversarial process similar to GAN. The decoder network takes random noise as input and maps it to the data space, whereas the encoder network maps RNA sequences to the latent representation. Finally, the encoder, decoder, and a discriminator network get combined into an adversarial game, where the discriminator network is trained to distinguish between the joint distribution of latent/data space samples from the decoder and encoder networks.

We adapted the inference learning approach of ALI and employed it to a large compendium of the pre-mRNA sequence of the human transcriptome. The objective was to learn an abstract representation of transcriptomic sequences that are typically 51 bp long and utilize the embedding as feature vectors for predicting eight different types of transcriptome modification sites, namely, m1A, m1G, m2G, m5C, 2′O, m5U, pseudouridine (Ψ), and dihydrouridine (D). Evaluating the effectiveness of features by predicting multiple post-transcriptional modifications serves in the manifold. First, the computational prediction of distinct epitranscriptomic marks is biologically significant and a long-sought goal for bioinformatics researchers. Accurate identification of these marks is essential for deciphering their biological functions and mechanism. However, discriminating between RNA modifications using only genomic sequence is a challenging task because there are more than 100 different types of RNA modifications characterized so far in diverse RNA molecules, including mRNAs, tRNAs, rRNAs, and lncRNAs and they may share similar nucleotide sequence preference. Many transcriptome-wide sequencing technologies have been developed recently to determine the global landscape of RNA modifications (e.g., Pseudo-seq, Ψ-seq, CeU-seq, Aza-IP, MeRIP-seq, m6A-seq, miCLIP, m6A-CLIP, RiboMeth-seq, Nm-seq, and m1A-seq) that identifies distinct epitranscriptomic marks [23]. However, individual experimental identification of these modification sites is very costly, labor-intensive, and time-consuming. So we created a benchmark dataset by combining eight different types of post-transcriptional modification data (i.e., m1A, m1G, m2G, m5C, m5U, 2′-O, Ψ, and D). We compared the effectiveness of the embeddings with the one-hot encoding of RNA sequences and demonstrated that the learned embeddings outperform one-hot encoding by a significant margin (Figure 2). We have also applied our method on the m6A dataset provided by SRAMP and improved the performance by 4–12% for different case scenarios in predicting the m6A site (Figure 3). Finally, we carried out exploratory analysis via t-SNE to rationalize the superiority of MR-GAN features and investigated motifs learned by our model to determine the biological relevance of the embedded representation.

Materials and Methods

Transcriptome Dataset for Training MR-GAN

Because our goal is to learn a low-dimensional representation of transcriptome sequences using GAN, we employed pre-mRNA sequences from the entire human transcriptome to train the proposed MR-GAN model. Engaging the large corpus of pre-mRNA sequences is important as it aids our unsupervised training to learn patterns from coding and non-coding regions to ensure that sufficient contexts are observed. The training dataset was compiled from all the chromosomes of the human genome hg38 assembly [24]. Separate fasta format files (*.fa) were downloaded from the UCSC genome browser for each of the chromosomes. Each of the files comprises intronic and exonic sequences for all the genes belonging to the specific chromosome. We then chopped the sequences at the non-overlapping interval of every 51 bp and got rid of the sequences that contain character N (represents ambiguous nucleotide). We selected 51 bp as the input RNA sample length because previous studies involving the prediction of RNA modifications have discovered this length as preferable for capturing contextual information [5, 6]. This preprocessing step results in a total of 40.7 million samples of length 51 bp. Finally, in order to feed into the MR-GAN for unsupervised learning, each of these sequences was represented by a 4 × 51 binary one-hot encoded matrix with rows corresponding to A, C, G, and U.

Datasets for Epitranscriptome Modifications

To train and evaluate models for human transcriptome modification site prediction using transcript sequence features extracted by the MR-GAN encoder, we created two benchmark datasets. The first dataset includes sites from eight different types of transcriptome modification and negative random sequences (see Table 1). The positive sites were collected primarily from RMBase [25], which provides the location information of the modified single base. We utilized the modified single base as the center and extended it to 51 bp by including 25-bp nucleotide sequences from both the upstream and downstream directions to form positive sequences. We also extracted a set of background sequence samples, randomly sampled from the genomic locations that do not contain any modification sites; we ensured that the proportions of the sequences centered at A, U, C, or G are roughly balanced (see Table 1). Then, the negative training samples were created for different types of modifications separately. For 2′-O or pseudouridine, which occurs on all four different nucleotides, the negative samples include the background samples and samples from all other eight types of modification. Otherwise, for any of the other seven modifications that are associated with a unique nucleotide, the negative samples include only sequences centered at that unique nucleotide from the background samples and those from other types. While some of the modifications have large positive sample sizes, the majority of others, as in m2G and m1G, have very few positive samples. Training reliable prediction models with small, highly imbalanced training datasets is a highly challenging learning task. In this work, we show the power of the MR-GAN encoder in helping extract and discriminate sequencing features for different modification sites.

TABLE 1
www.frontiersin.org

Table 1. Number of samples for different transcriptome modifications in the benchmark dataset.

The second dataset consists of positive and negative samples for transcriptome m6A methylation. We choose to train an MR-GAN on m6A separately because m6A is the most abundant and widely studied transcriptome modification and there are also existing sequence-based prediction algorithms. This dataset was derived mostly from [6], which used a similar dataset to train an m6A site predictor called SRAMP. Positive samples of this dataset are miCLIP sites from [26] that also contain the m6A motif (DRACH), and the negative samples are random sequences that also have the DRACH motif located at the center but have no miCLIP detected in m6A sites. Similar to for SRAMP, two sets of training data, namely, the full transcript mode, where the training sequences were collected from the full transcripts, and the mature mRNA mode, which extracted training sequences from cDNA sequences, were prepared. Table 2 provides detailed information about the m6A training datasets.

TABLE 2
www.frontiersin.org

Table 2. Numbers of samples in the training and testing datasets for m6A.

Generative Adversarial Network

GAN, proposed by Goodfellow et al. [16], is a generative network that learns the distribution of data and produces samples of synthesized data from the captured distribution. GAN includes two differentiable functions characterized by neural networks: the discriminator function C(x; θC) with parameters θC that outputs a single scalar representing the probability of x from the real rather than synthesized data distribution and the generator D(z; θD) parameterized by θD that maps samples from a prior of input noise variables Pz(z) to data distribution pdata(x). GAN is trained by implementing a two-player minimax game, where D is optimized to tell apart real from synthesized or fake data and D is optimized to generate data (from noise) that “fool” the discriminator. This joint optimization can be formulated as

minDmaxCV(D,C)=Ex~pdata(x)[logC(x)]                                    +Ez~pz[log(1C(D(z)))]    (1)

Despite the recent successes of GAN in computer vision, the idea was unexplored in the genomics domain until Frey et al. [20] applied the framework on DNA sequences with some modifications. Instead of training a generative model that produces realistic DNA sequences, they synthesized sequences with certain desired properties. For instance, the data distribution captured by GAN was utilized to design DNA sequences with higher protein binding affinity than those real sequences found in the protein binding microarray (PBM) data.

MR-GAN for Unsupervised Learning of Transcriptome Sequences

As discussed earlier, GANs lack an inference network that prohibits them from understanding abstract data representations. People have used the discriminator network to extract features, but as a learning entity to separate the real and synthetic samples, the discriminator primarily learns to discriminate features between real and synthetic samples. Hence, the discriminator features are not a true representation of the underlying data. Learning an inverse mapping from generated data E(x) back to the latent input z can be one viable solution to the problem under consideration. To this end, we propose MR-GAN, a model inspired by the ALI and BiGAN framework, which consists of three multilayer neural networks as depicted in Figure 1. Briefly, in addition to the generator D (or decoder in this case) from the standard GAN framework, MR-GAN includes an encoder E, which maps data x to latent representations z. The discriminator C of MR-GAN discriminates joint samples of the data and corresponding latent component [pairs (x; E(x)) vs. (z; D(z))] as real fake pairs, where the latent variable is either an encoder output E(x) or a generator input z. Overall, the training is performed to optimize the following minimax objective.

minD,EmaxCV(D,E,C)=Ex~pdata[logC(x,E(x))]                                            +Ez~pz[log(1C(D(z),z))]    (2)
FIGURE 1
www.frontiersin.org

Figure 1. Unsupervised feature learning scheme of MR-GAN.

In MR-GAN's setting, aside from training a generator, we train an encoder Q:Ω(x) → Ω(z), which maps data points x into the latent feature space. The discriminator takes input from both the data and latent representation, producing PC(Y|x, z), where Y = 1 if x is sampled from the real data distribution px and Y = 0 if x is generated through the output of D(z) and z ~ pz. Each of the C, D, and E modules, which are parametric functions, are optimized simultaneously using a stochastic gradient descent algorithm. The specific architecture of C, D, and E modules can be found in Supplementary Table 1. Since the MR-GAN encoder learns to capture the semantic attributes of whole human transcriptome sequences, it produces more powerful feature representations than a fully supervised model that is trained using only the labeled transcriptome modification site sequences, especially for those occasions where training data are limited.

Training MR-GAN

The GAN framework trains a generator, such that no discriminative model can distinguish samples of the data distribution from samples of the generative distribution. Both generator and discriminator are trained using the optimization function noted in equation (1). In the MR-GAN setting, the loss function (2) is considered, which is optimized over the generator, encoder, and discriminator. However, both (1) and (2) can be considered as computing the Jensen–Shannon divergence, which is potentially not continuous concerning the generator's parameters, thus leading to poor training convergence and stability [27]. In Arjovsky et al. [27], an alternative optimization function was proposed that is based on the Earth-Mover (also called Wasserstein-1) distance W(q, p) and is defined as the minimum cost of transporting mass in order to transform the distribution q into the distribution p. In contrast to Equations (1) and (2), the Wasserstein distance is a continuous and differentiable function in the parameter space and therefore has improved training stability. However, it requires weight clipping to be implemented, which still leads to the generation of poor samples and failure to converge. Therefore, we trained MR-GAN using the Wasserstein optimization function, but instead of enforcing the weight clipping, we imposed a soft version of the constraint with a penalty on the gradient norm as follows

L=Ez~pz[ C(D(z),z) ]Ex~pdata[ C(x,E(x)) ]Original Wasserstein Loss    +λ*Ex^~px^[ ( x^C(x^) 21)2 ]Gradient Penalty    (3)

This technique is known as WGAN-GP (gradient penalty) and was initially proposed in Gulrajani et al. [28]. The entire MR-GAN framework was optimized using the standard stochastic gradient descent method, where the learning rate was set to 1e−4. In our implementation, the batch size was equal to 100 samples. We trained the network for two epochs, which takes around 4 days in a computer with a single GPU. In our experience, the loss keeps oscillating during the training process as has been observed in previous studies [2931]. However, WGAN-GP has shown that the discriminator loss is close to zero when the GAN model approaches convergence. As a result, we took a snapshot of our model at every 2,000 iterations during the training process and selected the model with minimum discriminator loss (−5.2) for downstream processing. In our training, the minimum discriminator loss occurs at iteration 809,600, whereas the total number of iterations for two epochs was 814,600.

Training Predictors for Epitranscriptome Sites Based on Features Extracted From the MR-GAN Encoder

After the MR-GAN model is trained in an unsupervised manner using transcriptome-wide pre-mRNA sequences, we retained the MR-GAN encoder and used it as a feature extractor for modification site sequences and trained prediction models based on these features [17]. To train a model to predict modification sites, we first fed each 51-bp-long sequences in our training and testing datasets into the encoder and concatenated the feature maps from the last four convolutional layers into a vector of 2,112 features. Then, support vector machine (SVM) classifiers with the radial basis function kernel were trained as the site predictors for each modification type based on the feature vectors extracted from the corresponding training sets. This resulted in eight predictors for the first dataset and a separate predictor for m6A. Note that these eight predictors together can also be considered as a multiclass classifier trained based on the one-vs.-rest approach.

To obtain a baseline prediction performance, we also trained SVM classifiers using the one-hot encoded modification sequences. One hot encoding is a widely used encoding approach in deep learning to represent biological sequences with numeric formats [3236]. The one-hot encoding converts a 51-bp mRNA sequence into a 4 × 51 binary matrix, where each row corresponds to A, C, G, or U, and a single “1” in each column encodes the corresponding nucleotide at that location of the sequence.

Results

Prediction Performance of the Eight Epitranscriptome Modifications

In this section, we comprehensively evaluate the performances of the eight modification predictors trained on MR-GAN generated features and the one-hot encoding using the first dataset. Because the positive and negative samples for any modification are heavily imbalanced, where negative samples are about 3–200 times more than the positive samples, the appropriate metric to gauge the classification performance is the area under the precision–recall curve (auPRC). Since both precision and recall are dependent on the number of true positives rather than true negatives, the auPRC is less prone to inflation by the class imbalance than auROC [36]. Figure 2 shows the auPRC performances achieved by a 5-fold cross-validation scheme for each of the modifications. As described earlier, each modification predictor was trained by selecting the samples of the particular modification as the positive class, whereas the samples having the same center nucleotide as a positive class from the rest of the modifications including the random negative ones were attributed to the negative class. It is clear from Figure 2 that MR-GAN extracted features outperform one-hot encoding representation by significant margins for most of the RNA modifications. The auPRC increase attained by MR-GAN features ranges from 5.6 to 19.2% across different modifications. Notably, the performance improvements are more prominent for the modifications with a comparatively lower number of samples. For instance, the average auPRC for predicting D, m2G, m5C, and m5U using MR-GAN embedding is 12.3% higher than the one-hot encoding compared to the 3.8% average auPRC increment for the rest of the modifications with a larger sample size (2′-O-Me, pseudouridine, m1A). This indicates that the unsupervised training of MR-GAN has helped the model to learn additional information that otherwise would be difficult to learn with a limited number of training samples. Furthermore, the MR-GAN features consistently deliver prediction results at higher than 85% auPRC, which is a remarkable feat considering the minuscule ratio of positive to negative samples in the heavily imbalanced dataset. Using a highly skewed dataset to train predictors generally leads to the outcome that most of the positive samples are misclassified as negative ones. To solve this complexity, many prediction algorithms resample the negative data to balance the ratio of positive and negative subsets. Remarkably, we did not perform any data-balancing technique in training, as the features learned by the MR-GAN encoder were powerful enough to handle the class imbalance problem by itself. Finally, we noticed that the one-hot encoded feature representation performs comparably to the MR-GAN features for two modification types (m1A and m1G). Because there are only 29 positive samples for m1G, it is reasonable to expect that the one-hot encoded feature representation and the MR-GAN features reach similar performance but these performances might not be generalized due to the small sample size. For m1A, where there are 3,173 positive training samples, it is likely that the positional sequence pattern in training samples already contains significant discriminating information for the one-hot predictor to reach maximum accuracy. Taken together, these results demonstrate that the proposed unsupervised representation by MR-GAN successfully captures the important features required for discriminating the RNA modifications, which might not be captured with small training samples.

FIGURE 2
www.frontiersin.org

Figure 2. Prediction performance (auPRC) of eight epitranscriptome modifications using MR-GAN features and one-hot encoding. The number next to a modification name is the number of corresponding positive samples. The auPRCs are listed on top of the bars.

Prediction Performance of m6A Sites

While in the previous section, we demonstrated the superior prediction performance of MR-GAN embedding on modifications with smaller training samples, we investigate in this section whether this unsupervised feature learning can help improve the performance of m6A prediction, where the number of positive samples required for training is large. Because of the availability of m6A data from SRAMP, we were able to evaluate the performance on independent testing datasets as opposed to cross-validation. Similar to the SRAMP evaluations in Zhou et al. [6], we investigated the MR-GAN model performance in terms of auPRC on the full transcript mode and mature RNA mode using the independent testing data as described in Table 2.

Figure 3 summarizes the prediction results of MR-GAN and SRAMP on testing data for both the input modes. As evident, the MR-GAN extracted features outperform SRAMP proposed encodings in both modes (11.1% for the mature mRNA mode and 3.6% for the full transcript mode), suggesting that unsupervised learning from the entire transcriptome sequence can improve the learning of modification-related features over the encodings employed in the SRAMP algorithm. It is pointed out in Zhou et al. [6] that SRAMP suffers in the mature mRNA mode due to the discarding of all introns, which may disrupt the original sequence context of an m6A site and therefore reduce the discriminative capability of the extracted features. Also, the distance between an m6A site and a non-m6A site generally becomes closer in mature mRNA sequences compared with that in the corresponding pre-mRNA sequences, which further aggravates the prediction outcome of SRAMP. By contrast, the unsupervised training enables MR-GAN to learn the information about pre-mRNA sequences and thus produce considerable prediction improvement.

FIGURE 3
www.frontiersin.org

Figure 3. m6A prediction performance (auPRCs) of MR-GAN and SRAMP.

Visualizing MR-GAN Features for Different Epitranscriptome Modifications

To gain an intuitive understanding of the superior performance of MR-GAN features and explore the relationship among features of different modifications, we applied t-SNE [37] to project the high-dimensional features onto 2D and 3D spaces. t-SNE is a popular non-linear dimension reduction method that optimizes the similarities of the probabilities of distances between high-dimensional samples space and corresponding lower-dimensional projected samples. It has been widely used for data visualization in a variety of fields, such as video, image, and audio signals [37]. For this visualization, we first utilized principal component analysis (PCA) to map the MR-GAN and one-hot encoded representations of different RNA modification samples to lower dimensions, which were then fed to t-SNE to obtain representations in 2D and 3D spaces. We first compared the t-SNE plots in a 3D space between MR-GAN features and the one-hot encodings (Figure 4). For this comparison, we only included the five epitranscriptome modifications with a larger number of samples. As evident in Figure 4A, all one-hot encoded samples were squeezed together and barely separable. In contrast, MR-GAN extracted features show well-separated groups of samples (Figure 4B). As a result, it is much easier to achieve separable hyperplanes in the SVM classifier when MR-GAN features are used.

FIGURE 4
www.frontiersin.org

Figure 4. 3D scatterplots of t-SNE projected (A) one-hot encoded samples and (B) MR-GAN features for five epitranscriptome modifications that have a larger number of samples.

One way to validate the data embedding is to verify if the samples with similar functionalities are clustered together in the embedding space. Therefore, we investigated the relationship of MR-GAN features between different modifications, where we projected the data samples of eight types of modifications to the 2D space using t-SNE (Figure 5). To produce a clear and uncluttered 2D plot, we randomly picked a maximum of 70 samples from each of the modifications. Figure 5 reveals that the four most abundant mRNA modifications, m6A, m1A, m5C, and 2′-O, each form unique groupings with somewhat small overlapping among them, suggesting that each of them has its distinct sequence patterns that might be associated with modification-specific RBPs and hence unique functions. Such distinct grouping is supported by their different transcript distributions, where m6A, m1A, m5C, and 2′-O are mostly enriched in the stop codon, the start codon, the 3′UTR, and the CDS, respectively. However, there is a small overlapping between m6A and 2′-O. Indeed, single-base mapping technology has found that the second nucleotide from the 5′ cap of certain mRNA has N6, 2′-O-dimethyladenosine (m6Am) [26]. Moreover, 2′-O exists in all four types of nucleotides, and it is not surprising to see 2′-O to share some overlapping with almost all other modifications. We also observe that the m1A cluster contains virtually all m1G samples, implying that m1A and m1G have significantly share sequence patterns and potentially similar regulatory functions. Certainly, m1A is shown to disrupt the Watson–Crick base pairing and found to collaborate with m1G to induce local duplex melting in RNA [38]. The fact that they are isolated from other modifications may suggest that disrupting the base-pairing might be their unique function. Compared with the other three clusters, the m5C cluster shows the most overlapping with m6A and 2′-O. Although the evidence of collaborations between these three methylations is limited, m6A and m5C have been shown to enhance the translation of p21 by jointly methylating its 3′UTR. In addition to these four clusters, Ψ shows overlapping with m6A, m5C, and 2′-O. Ψ is found throughout different regions in mRNA. The samples of the remaining two modifications, D and m2G, are widely scattered without any clear patterns. They are also much less studied, and their actual distributions in mRNA and their functions are mostly unknown.

FIGURE 5
www.frontiersin.org

Figure 5. 2D scatter plot of all modification sites using MR-GAN feature map modifications with similar functions altogether.

Features Learned by MR-GAN Confirmed Known Modification Sequence Motifs

In this section, we delve into the encoder network to comprehend the sequence motifs for different mRNA modifications captured by the latent semantic representation of MR-GAN. Following a similar strategy described in DeepBind [32] for the sequence logo generation, we extracted the subsequences that give the maximum response in the convolutional operation for a kernel at the first layer of the encoder network. We repeated this subsequence extraction task on each of the sequences of the benchmark dataset using the 32 kernels at the first layer. Next, we performed motif enrichment analysis using MEME-ChIP [39] by feeding the unique subsequences from each modification type as the positive set while the subsequences extracted from the random samples of the benchmark dataset were utilized as the control set for each of the modifications to ensure the common background. The top-three enriched motifs discovered by MEME-ChIP and the associated RBPs for the relatively abundant mRNA modifications are shown in Figure 6. Two of the most enriched motifs detected for m6A belong to the well-known DRACH motif. The RBPs that share the motifs include FMR1, whose binding mRNAs in the mouse brain have been shown to be significantly methylated with m6A marks [40]. Other RBPs associated with these top motifs include splicing factors SRSF1 and zinc finger CCCH domain-containing protein ZC3H10. Currently, no evidence of their interaction with m6A exist. However, m6A reader protein YTHDC1 is shown to regulate splicing through interacting with splicing factors SRSF3 and SRSF10 [41, 42], and the zinc finger CCCH domain-containing protein ZC3H13 is reported to regulate m6A to control embryonic stem cell self-renewal. These existing results may point to unknown interactions of m6A with SRSF1 and ZC3H10. Next, we examined the known motifs for each of the remaining abundant modifications in the top enriched motifs and reported many other new motifs associated with known RBPs. To systematically investigate the associated RPBs predicted by MR-GAN, we accumulated the list of RBPs for each of the mRNA modifications that were identified as discriminative by MEME-ChIP in the above experiment, resulting in a total of 72 RBPs. We then performed a clustergram analysis (Figure 7) such that the RBPs are clustered according to the significance with which they are found to be unique to these mRNA modifications by MEME-ChIP. Interestingly, we discovered that most of the modifications have at least one unique RBPs (Supplementary Table 2). For instance, RBM3 and RALY are unique to m6A, HNRNPC is unique to m1A, and SRSF2 is unique to m5C. RBM3 is shown to be a functional partner of the splicing factor SRSF3 [43], which is recruited by m6A to regulate alternative splicing [42]. RALY is a member of the hnRNP family, which are considered as indirect m6A readers [44, 45]. Also, there were also RBPs, such as SNRNP70, that are determined as discriminative for most of the modifications. SNRNP70 is a key early regulator of 5′ splice site selection. This result could suggest that regulating splicing would be a common function which various modifications possess. We also discovered that 55 of the 72 RBPs identified by this analysis overlap with the proteins isolated by the RICK experiment [46], which systematically captures proteins bound to a wide range of RNAs (Supplementary Table 3). This indicates that our unsupervised learning captures RBPs that are biologically meaningful and repeatedly identified by other related studies.

FIGURE 6
www.frontiersin.org

Figure 6. The top motifs of different transcriptome modifications learned by MR-GAN.

FIGURE 7
www.frontiersin.org

Figure 7. Clustergram analysis (heatmap) of RNA-binding proteins identified to most likely interact with different transcriptome modifications. The color represents the significance with which they are found to be unique to these mRNA modifications by MEME-ChIP. The significance goes from lowest to highest as color varies from blue to red.

In order to further validate the credibility of the features captured by our method, we converted the 32 convolutional kernels for m5C samples into position frequency matrices or motifs following the similar procedure of DeepBind. Then, we aligned these motifs to known motifs using the TOMTOM algorithm [47]. Of the 32 motifs learned by the first layer of the encoder network, 25 significantly matched known RBP motifs (E < 0.05). Subsequently, we proceeded to verify whether the RBPs identified by this analysis concurs with the results of other studies investigating similar problems. In Amort et al. [48], the authors carried out an analysis to determine the relationship between m5C sites and RBPs using CLIP-seq data and reported the RBPs showing statistically significant enrichment of m5C in their binding sites compared to randomly sampled Cs. Expectedly, several of the RBPs identified by our motif analysis for m5C (6 of 25) were also discovered by that study, which further endorses the significance of our work (Supplementary Table 4).

Conclusion

We considered the prediction of different transcriptome modifications based on RNA sequences. To address the problem of small sample size for many of the modifications, we developed a GAN model called MR-GAN, which is trained to learn low-dimensional embeddings of transcriptomic-wide sequences in an unsupervised manner. The learned embedding, as demonstrated through the experimental results, contain the improved representation of sequences for different modifications as it maps the RNA modifications with similar functionalities together. We have also demonstrated that the motifs learned by MR-GAN in the process of discriminating between various transcriptome modifications are biologically meaningful and conforms to the findings of some of the previous studies. It is worth mentioning that we analyzed only nine out of almost 100 well-known modifications. We believe there would be more interesting patterns revealed if the MR-GAN sequence features are applied to additional RNA modifications. The main advantage of MR-GAN is that the model can perform in a satisfactory accuracy even with the heavily skewed dataset without the need for employing data-balancing techniques. This is a significant contribution to the bioinformatics research community because we often fail to develop a well-performing computation prediction model due to the lack of enough labeled data. We hope to extend this work by applying the embedding into more genomic sequence-related classification problems.

Data Availability Statement

MR-GAN is available in the GitHub repository (https://github.com/sirajulsalekin/MR-GAN). The training data are available for download at https://drive.google.com/open?id=1aASppi8f0jWk-iGNcMV1iP6yzz_tPj4g.

Author Contributions

SS, JZ, and YH conceived the study and designed the methods. SS and MM collected the data. SS trained the model. SS, MM, and Y-CC performed the result analysis. SS, MM, Y-CC, YC, JZ, and YH interpreted the results. SS, MM, YC, JZ, and YH wrote the manuscript. All authors contributed to the article and approved the submitted version.

Funding

This work was supported by the National Institutes of Health (R01GM113245 to YH, CTSA 1UL1RR025767-01 to YC, and K99CA248944 to Y-CC), Cancer Prevention and Research Institute of Texas (RP190346 to YC and YH and RP160732 to YC), San Antonio Life Sciences Institute (SALSI Innovation Challenge Award 2016 to YH and YC and SALSI Post-doctoral Research Fellowship 2018 to Y-CC), and the Fund for Innovation in Cancer Informatics (ICI Fund to Y-CC and YC).

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

We thank the computational support from UTSA's HPC cluster Shamu, operated by the Office of Information Technology.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fphy.2020.00196/full#supplementary-material

References

1. Meyer KD, Saletore Y, Zumbo P, Elemento O, Mason CE, Jaffrey SR. Comprehensive analysis of mRNA methylation reveals enrichment in 3′ UTRs and near stop codons. Cell. (2012) 149:1635–46. doi: 10.1016/j.cell.2012.05.003

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Dominissini D, Moshitch-Moshkovitz S, Schwartz S, Salmon-Divon M, Ungar L, Osenberg S, et al. Topology of the human and mouse m6A RNA methylomes revealed by m6A-seq. Nature. (2012) 485:201–6. doi: 10.1038/nature11112

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Shi H, Wei J, He C. Where, when, and how: context-dependent functions of rna methylation writers, readers, and erasers. Mol Cell. (2019) 74:640–50. doi: 10.1016/j.molcel.2019.04.025

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Liu Z, Xiao X, Yu D-J, Jia J, Qiu W-R, Chou K-C. pRNAm-PC: predicting N 6-methyladenosine sites in RNA sequences via physical–chemical properties. Anal Biochem. (2016) 497:60–7. doi: 10.1016/j.ab.2015.12.017

PubMed Abstract | CrossRef Full Text | Google Scholar

5. Chen W, Feng P, Ding H, Lin H, Chou K-C. iRNA-Methyl: identifying N 6-methyladenosine sites using pseudo nucleotide composition. Analy Biochem. (2015) 490:26–33. doi: 10.1016/j.ab.2015.08.021

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Zhou Y, Zeng P, Li Y-H, Zhang Z, Cui Q. SRAMP: prediction of mammalian N6-methyladenosine (m6A) sites based on sequence-derived features. Nucleic Acids Res. (2016) 44:e91. doi: 10.1093/nar/gkw104

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Xiang S, Liu K, Yan Z, Zhang Y, Sun Z. RNAMethPre: a web server for the prediction and query of mRNA m 6 a Sites. PLoS ONE. (2016) 11:e0162707. doi: 10.1371/journal.pone.0162707

CrossRef Full Text | Google Scholar

8. Xing P, Su R, Guo F, Wei L. Identifying N6-methyladenosine sites using multi-interval nucleotide pair position specificity and support vector machine. Sci Rep. (2017) 7:46757. doi: 10.1038/srep46757

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Chen W, Tang H, Lin H. MethyRNA: a web server for identification of N6-methyladenosine sites. J Biomol Struct Dyn. (2017) 35:683–7. doi: 10.1080/07391102.2016.1157761

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Chen K, Wei Z, Zhang Q, Wu X, Rong R, Lu Z, et al. WHISTLE: a high-accuracy map of the human N6-methyladenosine (m6A) epitranscriptome predicted using a machine learning approach. Nucleic Acids Res. (2019) 47:e41. doi: 10.1093/nar/gkz074

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Yoon K, Kwek S editors. An unsupervised learning approach to resolving the data imbalanced issue in supervised learning problems in functional genomics. In: HIS'05 Fifth International Conference on Hybrid Intelligent Systems. Washington, DC: IEEE (2005).

Google Scholar

12. Spencer M, Eickholt J, Cheng J. A deep learning network approach to ab initio protein secondary structure prediction. IEEE/ACM Trans Comput Biol Bioinform. (2015) 12:103–12. doi: 10.1109/TCBB.2014.2343960

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Eickholt J, Cheng J. DNdisorder: predicting protein disorder using boosting and deep networks. BMC Bioinformatics. (2013) 14:88. doi: 10.1186/1471-2105-14-88

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Eickholt J, Cheng J. Predicting protein residue–residue contacts using deep networks and boosting. Bioinformatics. (2012) 28:3066–72. doi: 10.1093/bioinformatics/bts598

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Asgari E, Mofrad MR. Continuous distributed representation of biological sequences for deep proteomics and genomics. PLoS ONE. (2015) 10:e0141287. doi: 10.1371/journal.pone.0141287

PubMed Abstract | CrossRef Full Text | Google Scholar

16. Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, et al. Generative adversarial nets. In: Advances in Neural Information Processing Systems. Montreal, QC (2014).

Google Scholar

17. Radford A, Metz L, Chintala S. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv. (2015) 151106434.

Google Scholar

18. Berthelot D, Schumm T, Metz L. Began: boundary equilibrium generative adversarial networks. arXiv. (2017) 170310717.

Google Scholar

19. Springenberg JT. Unsupervised and semi-supervised learning with categorical generative adversarial networks. arXiv. (2015) 151106390.

Google Scholar

20. Killoran N, Lee LJ, Delong A, Duvenaud D, Frey BJ. Generating and designing DNA with deep generative models. arXiv. (2017) 171206148.

Google Scholar

21. Gupta A, Zou J. Feedback GAN (FBGAN) for DNA: a novel feedback-loop architecture for optimizing protein functions. arXiv. (2018) 180401694.

Google Scholar

22. Dumoulin V, Belghazi I, Poole B, Mastropietro O, Lamb A, Arjovsky M, et al. Adversarially learned inference. arXiv. (2016) 160600704.

Google Scholar

23. Xuan J-J, Sun WJ, Lin P-H, Zhou K-R, Liu S, Zheng L-L, et al. RMBase v2. 0: deciphering the map of RNA modifications from epitranscriptome sequencing data. Nucleic Acids Res. (2017) 46:D327–4. doi: 10.1093/nar/gkx934

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Kent WJ, Sugnet CW, Furey TS, Roskin KM, Pringle TH, Zahler AM, et al. The human genome browser at UCSC. Genome Res. (2002) 12:996–1006. doi: 10.1101/gr.229102

PubMed Abstract | CrossRef Full Text | Google Scholar

25. Sun W-J, Li J-H, Liu S, Wu J, Zhou H, Qu L-H, et al. RMBase: a resource for decoding the landscape of RNA modifications from high-throughput sequencing data. Nucleic Acids Res. (2015) 44:D259–65. doi: 10.1093/nar/gkv1036

PubMed Abstract | CrossRef Full Text | Google Scholar

26. Linder B, Grozhik AV, Olarerin-George AO, Meydan C, Mason CE, Jaffrey SR. Single-nucleotide-resolution mapping of m6A and m6Am throughout the transcriptome. Nat Methods. (2015) 12:767–72. doi: 10.1038/nmeth.3453

PubMed Abstract | CrossRef Full Text | Google Scholar

27. Arjovsky M, Chintala S, Bottou L. Wasserstein gan. arXiv. (2017) 170107875.

Google Scholar

28. Gulrajani I, Ahmed F, Arjovsky M, Dumoulin V, Courville AC editors. Improved training of wasserstein gans. In: Advances in Neural Information Processing Systems. Long Beach, CA (2017).

Google Scholar

29. Arjovsky M, Bottou L. Towards principled methods for training generative adversarial networks. arXiv. (2017) 170104862.

Google Scholar

30. Liang T, Stokes J. Interaction matters: a note on non-asymptotic local convergence of generative adversarial networks. arXiv. (2018) 180206132.

Google Scholar

31. Gidel G, Berard H, Vincent P, Lacoste-Julien S. A variational inequality perspective on generative adversarial nets. arXiv. (2018) 180210551.

Google Scholar

32. Alipanahi B, Delong A, Weirauch MT, Frey BJ. Predicting the sequence specificities of DNA-and RNA-binding proteins by deep learning. Nat Biotechnol. (2015) 33:831–8. doi: 10.1038/nbt.3300

PubMed Abstract | CrossRef Full Text | Google Scholar

33. Zhou J, Troyanskaya OG. Predicting effects of noncoding variants with deep learning-based sequence model. Nat Methods. (2015) 12:931–4. doi: 10.1038/nmeth.3547

PubMed Abstract | CrossRef Full Text | Google Scholar

34. Salekin S, Zhang JM, Huang Y editors. A deep learning model for predicting transcription factor binding location at single nucleotide resolution. In: IEEE EMBS International Conference on Biomedical & Health Informatics (BHI). Orlando, FL: IEEE (2017). doi: 10.1109/BHI.2017.7897204

PubMed Abstract | CrossRef Full Text | Google Scholar

35. Salekin S, Zhang JM, Huang Y. Base-pair resolution detection of transcription factor binding site by deep deconvolutional network. Bioinformatics. (2018) 34:3446–53. doi: 10.1093/bioinformatics/bty383

PubMed Abstract | CrossRef Full Text | Google Scholar

36. Quang D, Xie X. DanQ: a hybrid convolutional and recurrent deep neural network for quantifying the function of DNA sequences. Nucleic Acids Res. (2016) 44:e107. doi: 10.1093/nar/gkw226

PubMed Abstract | CrossRef Full Text | Google Scholar

37. Liu S, Maljovec D, Wang B, Bremer PT, Pascucci V. Visualizing high-dimensional data: advances in the past decade. IEEE Trans Vis Comput Graph. (2017) 23:1249–68. doi: 10.1109/TVCG.2016.2640960

PubMed Abstract | CrossRef Full Text | Google Scholar

38. Zhou H, Kimsey IJ, Nikolova EN, Sathyamoorthy B, Grazioli G, McSally J, et al. m(1)A and m(1)G disrupt A-RNA structure through the intrinsic instability of Hoogsteen base pairs. Nat Struct Mol Biol. (2016) 23:803. doi: 10.1038/nsmb.3270

PubMed Abstract | CrossRef Full Text | Google Scholar

39. Machanick P, Bailey TL. MEME-ChIP: motif analysis of large DNA datasets. Bioinformatics. (2011) 27:1696–7. doi: 10.1093/bioinformatics/btr189

PubMed Abstract | CrossRef Full Text | Google Scholar

40. Chang M, Lv H, Zhang W, Ma C, He X, Zhao S, et al. Region-specific RNA m(6)A methylation represents a new layer of control in the gene regulatory network in the mouse brain. Open Biol. (2017) 7:170166. doi: 10.1098/rsob.170166

PubMed Abstract | CrossRef Full Text | Google Scholar

41. Roundtree IA, He C. Nuclear m(6)a reader YTHDC1 regulates mRNA splicing. Trends Genet. (2016) 32:320–1. doi: 10.1016/j.tig.2016.03.006

PubMed Abstract | CrossRef Full Text | Google Scholar

42. Xiao W, Adhikari S, Dahal U, Chen YS, Hao YJ, Sun BF, et al. Nuclear m(6)A reader YTHDC1 regulates mRNA splicing. Mol Cell. (2016) 61:507–19. doi: 10.1016/j.molcel.2016.01.012

PubMed Abstract | CrossRef Full Text | Google Scholar

43. Fuentes-Fayos A, Vázquez-Borrego M, Jiménez-Vacas J, Bejarano L, Blanco-Acevedo C, Sánchez-Sánchez R, et al. P11. 17 Splicing dysregulation drives glioblastoma malignancy: SRSF3 as a potential therapeutic target to impair glioblastoma progression. Neuro-Oncology. (2019) 21:iii46. doi: 10.1093/neuonc/noz126.163

CrossRef Full Text | Google Scholar

44. Wu B, Su S, Patil DP, Liu H, Gan J, Jaffrey SR, et al. Molecular basis for the specific and multivariant recognitions of RNA substrates by human hnRNP A2/B1. Nat Commun. (2018) 9:420. doi: 10.1038/s41467-017-02770-z

PubMed Abstract | CrossRef Full Text | Google Scholar

45. Liu N, Zhou KI, Parisien M, Dai Q, Diatchenko L, Pan T. N6-methyladenosine alters RNA structure to regulate binding of a low-complexity protein. Nucleic Acids Res. (2017) 45:6051–63. doi: 10.1093/nar/gkx141

PubMed Abstract | CrossRef Full Text | Google Scholar

46. Bao X, Guo X, Yin M, Tariq M, Lai Y, Kanwal S, et al. Capturing the interactome of newly transcribed RNA. Nat Methods. (2018) 15:213. doi: 10.1038/nmeth.4595

PubMed Abstract | CrossRef Full Text | Google Scholar

47. Gupta S, Stamatoyannopoulos JA, Bailey TL, Noble WS. Quantifying similarity between motifs. Genome Biol. (2007) 8:R24. doi: 10.1186/gb-2007-8-2-r24

PubMed Abstract | CrossRef Full Text | Google Scholar

48. Amort T, Rieder D, Wille A, Khokhlova-Cubberley D, Riml C, Trixl L, et al. Distinct 5-methylcytosine profiles in poly (A) RNA from mouse embryonic stem cells and brain. Genome Biol. (2017) 18:1. doi: 10.1186/s13059-016-1139-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: N6-methyladenosine (m6A), epitranscriptome, RNA modification site prediction, generative adversarial networks (GANs), unsupervised representation learning, methylated RNA immunoprecipitation sequencing (MeRIP-Seq)

Citation: Salekin S, Mostavi M, Chiu Y-C, Chen Y, Zhang J and Huang Y (2020) Predicting Sites of Epitranscriptome Modifications Using Unsupervised Representation Learning Based on Generative Adversarial Networks. Front. Phys. 8:196. doi: 10.3389/fphy.2020.00196

Received: 23 January 2020; Accepted: 01 May 2020;
Published: 19 June 2020.

Edited by:

Andrew D. McCulloch, University of California, San Diego, United States

Reviewed by:

Fuhai Li, Washington University in St. Louis, United States
Vivek Kohar, Jackson Laboratory, United States
Wenjin Jim Zheng, University of Texas Health Science Center at Houston, United States

Copyright © 2020 Salekin, Mostavi, Chiu, Chen, Zhang and Huang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Jianqiu Zhang, michelle.zhang@utsa.edu; Yufei Huang, yufei.huang@utsa.edu