Community Impact on a Cryptocurrency: Twitter Comparison Example Between Dogecoin and Litecoin

Context: The third generation of cryptocurrencies gathers cryptocurrencies that are as diverse as the market is big (e.g., Dogecoin or Litecoin). While Dogecoin is seen as a memecoin, the other gathers a very different category of investors. To our knowledge, no study has independently assessed the crypto community’s economical impact on these cryptocurrencies. Furthermore, various methodological possibilities exist to forecast cryptocurrency price—mainly coming from online communities. Method: Our study has retrospectively studied (from 01/01/2015 to 03/11/2021)—using open access data—the association strength (using normalized mutual information) and the linear correlation (using Pearson’s correlation) between Twitter activity and cryptocurrency economical attributes. In addition, we have computed different models (ADF, ARIMA, and Interpretable MultiVvariable Long Short-Term Memory recurrent neural network) that forecast past price values and assessed their precision. Findings and conclusions: While the average Dogecoin transaction value is impacted by tweets, tweets are impacted by Litecoin transactions number and average Litecoin transaction value. Tweet number is impacted by Dogecoin whale behavior, but no significant relationship was found between Litecoin whales and tweets. The forecasting error resulting from our ARIMA (0,0,0) models was 0.08% (with Litecoin) and 0.22% (with Dogecoin). Therefore, those are just the beginning of scientific findings that may lead to building a trading robot based on these results. However, in itself, this study is only for academic discussion, and conclusions need to be drawn by further research. The authors cannot be liable if any financial investment is made based on its conclusions.


INTRODUCTION
Since Satoshi Nakamoto's whitepaper in 2008, cryptocurrencies have grown to a huge market capitalization-currently over $2T (as of December 2021). This huge rise in cryptocurrency market capitalization seems, at first glance, deeply linked to the cryptocurrency community. Indeed, most coins have a strong community promoting them through social networks. One of the most relevant examples when talking about online advertising of a coin might be Elon Musk's tweets. He seemed to have a huge impact on the cryptocurrency market as value seems to increase or decrease as he tweets, which could constitute an insider delay. However, according to a big data study (Tandon et al., 2021), it can clearly be stated that Elon Musk cannot control the utter volatile world of cryptocurrencies and especially Bitcoin and Dogecoin.
But the price of cryptocurrency could be more driven by the Lindy effect than anything else. This theory states that the future life expectancy of certain non-perishable goods-such as a technology or an idea-is proportional to their current age. Indeed, the longer something has been around, the more chances there are that it will survive longer, and among memes, competition for survival is fierce. In this jungle, the average lifespan is roughly four months. When compared to other memes, Doge is kind of a venerated elder. By surviving for eight years-The Doge meme first became popular in 2013-it has already proven to be one of the most resilient memes of the whole internet history. The Lindy effect suggests that for this reason alone, Dogecoin is more likely to persist into the future than any other meme. Just as the U.S. dollar is backed by America's hegemonic power, Dogecoin is backed by some of the most powerful memes in existence-and the communities behind them. Dogecoin has a real fan base promoting its use through social networks. Part of what has made Dogecoin a successful cryptocurrency is the non-tribalism of its community.
Moreover, while being technically very similar (i.e., almost the same PoW and use cases), Litecoin has a less loud community. Despite being an older and more stable cryptocurrency, Litecoin does not have the same online popularity. Litecoin's users are not that loud over social networks and do not mean (most of them) to organize coordinated buying in order to influence the currency's value.
Litecoin and Dogecoin have been selected for this study because of their similarities. Indeed, Litecoin was an early Bitcoin spinoff (or altcoin), starting in October 2011 (Ex-Googler Gives the Worl, 2018), and Dogecoin's protocol is based on the existing cryptocurrencies: Luckycoin and Litecoin (Gilbert, 2013). Dogecoin was launched on 6 December 2013 (Noyes, 2014). At first glance, these cryptocurrencies are different and might not be comparable.
As explained previously, the Lindy effect allowed Dogecoin-the "memecoin"-to free itself from being just a joke. Thus, after a period of time during which Dogecoin was ignored (Locke, 2021), it was pushed back by its community and became trendy again recently-both at the level of its adoption and at the level of its technical development (Dogecoin Foundation (2021) (reduction of transaction costs, bridge with ETH, first NFTs are being created on its network). Now, we can see a real decentralized approach coming from this cryptocurrency and its community. Litecoin, on the other hand, has a more centralized development-mainly because of its history-but does not prevent it from having one of the largest capitalizations. Therefore, Litecoin was one of the first altcoins, and Dogecoin was one of the first memecoins (if not the first). As with any cryptocurrency at launch, their future relies on their respective communities. We chose not to use other memecoins for comparison because of their speculative-driven mindset and have been so far removed from cypherpunk libertarian way of life (Hughes, 1993).
As it was broadly studied and well-documented, cryptoeconomy and the traditional financial economy behavior has about 20 well-known cognitive biases (Douziech, 2021). Here, we will be addressing the question: "how much of an impact does online activity-through Twitter in this study-have on the cryptocurrency market?" Questions like this one have already been raised and studied for Bitcoin.
-Temporal convolutional networks perform significantly better than both autoregressive and other deep learning-based models in the literature, and the tweet author meta-information, even detached from the tweet itself, is a better predictor than the semantic content and tweet volume statistics (Akbiyik et al., 2021); -Statistical tests show that the simplest GARCH(1,1) has the best reaction to the addition of an external signal in order to model the volatility process on out-of-sample data (Barjašić and Antulov-Fantulin, 2020).
In this way, we will study two main cryptocurrencies (Litecoin and Dogecoin) here.
• Are cryptocurrency prices and tradeoffs affected by the fluctuation in cryptocurrency community tweets? • Can the price of Litecoin and Dogecoin be forecasted (according to this assumption)? Which statistical forecasting model has the best performance? • Can Dogecoin or any Memecoin actually become a currency of the future?

METHODS
Our analysis is divided into two parts: -The first one covers the correlation/causality analysis -The second addresses the possible price forecasting for those cryptocurrencies based upon causality/correlation analysis.

Causality and Correlation Analysis
We used two methods to assess the association presence or absence between X and Y. These two methods were the classical Pearson's correlation and the normalized Shannon mutual information.

Data Sources
The data frame comes from three websites ( 1, 2, 3 ), but we use two dataframe versions here because of lacking figures for specific days. The first file is the original one which contains some "null" values. But, in order to work with our algorithm, they have been filled (in the second file) with the average value of the last existing value and the next one. This allows us to work with our files without introducing a new bias in our correlations.

Statistical Method
Obviously, correlation is not causation; but the lack of correlation implies absence of causality. Correlation (which might be negative or positive) is, therefore, a key component of the scientific process for it evinces collections of variables that may interact with each other, thereby warranting further study. Conversely, this methodology also accounts for the early dismissal of unwarranted hypotheses regarding such interplay between variables.
The first method we used is based on the standard Pearson's correlation matrix , and the computation was performed using the Python Numpy library. Then, we controlled the results with two Pearson's formulas for discrete series and continuous series. Specifically, we used the following function: numpy.corrcoef (df [cols]. values.T). where -df is the dataframe of the data -cols is the list of columns used for the matrix First, let us talk about Pearson's correlation: it is a commonly formulated criticism that one may not establish a linear correlation between a series of quantitative variables and another one of qualitative variables. However, it will help us identify those correlations as we are looking at them (Lev, 1949;Tate, 1954;Kornbrot, 2005). Pearson's correlation evaluates the linear relationship between two continuous variables. A relationship is said to be linear when a modification of one of the variables is associated with a proportional modification of the other variable.
Spearman's correlation evaluates the monotonic relationship between two continuous or ordinal variables. In a monotonic relationship, the variables tend to change together but not necessarily at a constant rate. This correlation coefficient is based on the ranked values of each variable rather than the raw data. Therefore, in view of the continuous variables that we will study, Pearson's correlation seems more suitable to study the immediate impact of tweets on the economic parameters of Dogecoin and Litecoin.
The second method is based on mutual information entropy (Pébaÿ, 2021;Pébaÿ et al., 2021). This allows us to be free from the limitation assumption of monotony required by linear correlation. It measures the quantity of information (in the sense defined by Claude Shannon in 1948(Shannon, 1948) that two distributions share. In other words, it measures the association ("clustering") between two variables: it is important to notice that his approach is not linear correlation but classical information entropy. Indeed, we computed a dimensionless quantity, generally expressed in units of bits (Thompson and Pebay, 2009), which may be thought of as the reduction in uncertainty about one random variable given the knowledge of another. For instance, high mutual information means a large reduction in uncertainty on one variable, given the other, whereas low mutual information indicates a small reduction in this uncertainty; eventually, zero mutual information between two random variables entails no association between the two distributions (McDaid et al., 2011). Furthermore, Shannon's source coding theorem establishes strict bounds on what can be known about one data series and state might be compressed-which, in turn, explains how and to what extent one variable might be a proxy of another one without data loss. Shannon information entropy has been demonstrated to be especially efficient for algorithmic complexity evaluation when evaluated with the block decomposition method (Zenil et al., 2016;Zenil, 2020). Moreover, according to N. N. Taleb, entropy metrics solve practically all correlation paradoxes in the field of social sciences (or rather, pseudo-paradoxes) (Taleb, 2019). Another important example of the relevance of this technique is that of mother wavelet selection, where it demonstrated superior sensitivity to quantify the changes in signal structure than classical mean-squared error and correlation coefficient (Wijaya et al., 2017).
Therefore, in order to compute reproducible results, we use the "muinther" R package available on GitHub, which uses these two statistical methods .

Biases
The first important bias is community size. Indeed, that could impact Pearson's method, more prone to these issues. However, a larger community will be able to reduce the extreme variations of the variables studied (number of tweets). Therefore, for the two methods, we will not be able to compare the raw data from the samples but only the coefficients (from Pearson's or the normalized information theory) between these two cryptocurrencies.
The second bias is Pearson's method in itself. Indeed, by its definition, Pearson's correlation evaluates the linear relationship between two continuous variables. A relationship is said to be linear when a modification of one of the variables is associated with a proportional modification of the other variable. However, if one moves in a monotonic relationship, the variables tend to change together but not necessarily at a constant rate (de Winter et al., 2016). In this case, Spearman's correlation would be better.

PRICE PREDICTION MODEL Settings
We used historical data, spanning from 01/01/2015 to 03/11/ 2021, by extracting various economic trackers as detailed below.

Variables
Three variables were primarily used: -"date" -"tweets" (a continuous variable describing the number of tweets per day with a mention of the interest in cryptocurrency), -"price" (a continuous variable describing adjusted closure price of the interested cryptocurrency).
If one of these values is missing, data were censored; in this way, we obtained 2,482 data for each variable.

Data Sources
Tweets were collected and extracted from the Twitter API, and prices were extracted on Yahoo Finance (3).

Price Correlation/Causation With Tweet Number
The same method was used to explore the correlation and causation relationship between Dogecoin/Litecoin price and tweet number. In other words, we first used (all from "muinther" R package) Pearson's correlation method and then Shannon mutual information entropy to assess that.

Application
Following our correlation/causation analysis, we will be able to determine two functions in order to establish a relationship between currency price and tweet number at some point "t".

Augmented Dickey-Fuller Test
Since our Litecoin and Dogecoin data are time series datasets, it is important to check whether the data might be somehow stationary. In order to check this, we considered an ADF (an acronym for Augmented Dickey-Fuller) test. These unit roots are the reason for causing unpredictable results in time series data analysis. Thus, the ADF is a significance test, so a null and alternative hypothesis comes into play by that, test statistics are calculated, and the p-value is reported. Based on p-values, the stationarity of data is determined. Basically, ADF determines the trend of data and determines how strongly or weakly the time series is defined by a trend. However, we use three linear regression models to evaluate this.
1) The first type (type1) is a linear model with no drift and linear trend with respect to time: where d is an operator of first order difference, that is, dx(t) = x(t) -x (t-1), and e(t) is an error term.
2) The second type (type2) is a linear model with drift but no linear trend: 3) The third type (type3) is a linear model with both drift and linear trend: Stationary data means that data statistical properties do not depend on time. If the given data are non-stationary, we need to change it to being stationary by applying a natural log.

ARIMA
The model we choose to use is ARIMA (stands for AutoRegressive Integrated Moving Average). It is a part of linear regression models mainly used to predict future values based on past behavior of the target. It is said that history does not repeat itself, but it surely has and had its own rhythm. The beauty of ARIMA models is that these do not use any exogenous values imposed on them but rather are completely dependent on the past target values for prediction. ARIMA can be broken as AR, I, and MA. As mentioned previously, AR stands for Auto-Regressive, and it works on the idea of regressing the target on its past variable, which is nothing but lagging on itself. Eq. 1 indicates a value Y is a linear function of its past n values. These n values can be chosen and are beta values used during fitting of the model. This equation helps forecast future values by simply making the following changes as in Eq. 2.
The Integrated part of ARIMA deals with stationalizing data. Here, the differencing is applied on data as shown in Eq. 3. It indicates that the future values of Y are some linear function of its past changes. The reason for differencing is that the time series data are not stable and the values of Y should have mean variance stationary.
Moving average is all summarized in Eq. 4, somewhat similar to AN equation with lags. E indicates that the error in the data is nothing but the residual derivations between the model and target value.
This is the standard notation to represent ARIMA models. These parameters can be replaced with integer values to specify the type of the model used. Parameter "p" is referred to as the lag order of AR, that is, the number of lags in Y to be included in the model, "d" is the order of differencing required for making data stationary, and "q" is referred to as the order of MA, that is, the number of lagged forecast errors.

Criterions
There are several different models (i.e., based on the orders such as AR and ARMA models of certain order or different order) for building time series. The lower the value obtained using these criteria, the better suitable will be the model for our time series data. Parameters used in these criteria include-log-likelihood (L), signifying how strong the model is in fitting the data. Generally, it is considered that the more complicated the model is, the better it fits the data. It is true though, in fitting, but also leverages the concept of overfitting (i.e., the model fits the training data better but loses its ability to generalize on test data). To prevent that, the number of predictors k (i.e., number of lags (fixed amount of passing time)) plus a constant is incorporated. Another parameter to consider here is T, the number of samples or observations used for estimation.
Below are listed the criteria used in this experiment for selecting the best model.  be used for selecting the predictors of the regression model. AIC can be calculated using the formula given below.
If T values are low, AIC may tend to predict too many predictors, so in order to prevent that bias, the corrected version of AIC, that is, AICc, will be considered.

AICc
AIC where L is the likelihood value, p is the order of the AR model, q is the order of the MA model, k is the number of predictors, and T is the number of observations which is used for estimation as mentioned above. In order to obtain the best model, we need to consider the model with a low AIC value. That means, the value of k should be low, and the value of L should be at its maximum, illustrating that the model will be simple as k is low and fitting the data well with max l.
• Bayesian Information Criterion (BIC) -BIC, also known as Schwarz information criterion, is used for model selection based on the score obtained.
Here also, the minimum value should be taken into consideration. BIC with a small value illustrates that the model is simple, with a relatively low k number, which fits best the model and is trained on a few observations. In addition to that, other studies have revealed that the ARIMA model has strong potential for short-term prediction and can compete with existing techniques in stock price prediction (Mahan et al., 2015).

IMV-LSTM
Eventually, we will use and assess the precision of our SoA Deep Learning-especially the Interpretable Multivariable-Long Short-Term Memory neural networks . We use the implementation available in Pytorch 4 . Indeed, according to a previous study (Barić et al., 2021), it seems to be the only model with both a satisfying performance score and correct interpretability, capturing both autocorrelations and crosscorrelations between multiple time series. Interestingly, while evaluating IMV-LSTM on simulated data from statistical and mechanistic models, the correctness of interpretability increases with more complex datasets.
Other models are often used in this domain, including stateof-the-art model extensions to probabilistic forecasting such as CatBoostLSS (or Quantile Regression Forests) (Daniel, 2019) and Gaussian processes in a dynamic linear regression, as a replacement for the Kalman Filter 5 .

Data Availability
All data generated or analyzed during this study are included in this published article (and its repository 6 ).

Correlation and Causation Analysis
Pearson's Correlation Analysis LITECOIN All Pearson's correlations studied with Litecoin (Supplementary File S1 and Figure 1A.) were significant with a p-value under 0.001, except the correlation between the Litecoin market cap and average Litecoin transaction value (p-value = 5.938*10^-3).
Tweets Mutual information theory analysis.

LITECOIN
"Association" Analysis. Community tweets are strongly (with a normalized mutual information coefficient of 0.9 at least) associated ( Figure 2A and Supplementary "Causality" Analysis. We will explore and emphasize only significant association causality (previously described). About tweet association with the economical trackers, tweet number was impacted by Litecoin transactions [conditional information entropy of tweets given Litecoin transactions (0.637) is higher than the conditional information entropy of Litecoin transactions given tweets (0.070) as Supplementary  Table S1 shows] and tweet number was impacted by the Litecoin average transaction value too [conditional information entropy of tweets given Litecoin average transaction value (0.637) is higher than the conditional information entropy of Litecoin average transaction value given tweets (0.071)].
When looking at other associations, we can see that Litecoin active addresses are impacted by Litecoin transactions [conditional information entropy of Litecoin active addresses given Litecoin transactions (0.071) is higher than the conditional information entropy of Litecoin transactions given Litecoin active addresses (0.037)] and Litecoin transactions are impacted by Litecoin average transaction value [conditional information entropy of Litecoin transactions given Litecoin average "Causality" Analysis. We will explore and emphasize only significant association causality (previously described).
When looking at associations between tweets and Dogecoin economical trackers, we notice that the average Dogecoin transaction value is impacted by tweets [conditional information entropy of average Dogecoin transaction value given tweets (0.861) is higher than the conditional information entropy of tweets given average Dogecoin transaction value (0.048)] and tweets are impacted by Dogecoin whales [conditional information entropy of tweets given Dogecoin whales (0.861) is higher than the conditional information entropy of Dogecoin whales given tweets (0.124)].
About other associations, mainly looking at whales; Dogecoin active addresses are impacted by Dogecoin whales [conditional information entropy of Dogecoin active addresses given Dogecoin whales (0.120) is higher than the conditional information entropy of Dogecoin whales given Dogecoin active addresses (0.029)]; Dogecoin whales are impacted by Dogecoin transactions [conditional information entropy of Dogecoin whales given Dogecoin transactions (0.124) is higher than the conditional information entropy of Dogecoin transactions given Dogecoin whales (0.049)].

Price Prediction Model
Considering historical price data of Litecoin/Dogecoin, we first analyzed the price and volume evolution over the years. Since prices data are a time series dataset, it is necessary to check this stationarity for model building.

Stationary Testing
We used the ADF test to determine the property of our time series variable. To do so, considering null and alternative hypotheses as the representation of test data can be performed using unit root or is non-stationary.  ADF statistics and p-value of time series variables were calculated. The results ( Table 1) were as follows: -Dogecoin non-stationary data show no trend in both of the three models (without a significant p-value) ( Figure 3A.). -Litecoin non-stationary data show no trend in both of the three models but with a significant p-value only in the third one ( Figure 4A.). -Dogecoin stationary data show a significant trend absence in the first model, a nonsignificant trend absence in the third model, and a nonsignificant trend presence in the second model ( Figure 3B). -Litecoin stationary data show a nonsignificant trend presence in the first model and a nonsignificant trend absence in the second and third model ( Figure 4B).
To make it stationary and remove the trend, we simply applied a natural logarithm on the price values of Litecoin and Dogecoin. The values of time series before and after removing trends can be found in the table. In this way, the third model (a linear model with both drift and linear trend) was the best one to study possible trends in those data. Figure 5 illustrates the non-stationary data (without and with a square scale transformation) and stationary data price v/s time line plot studied with the third model.

Price Correlation/Causation With Number of Tweets
About Pearson's correlation between currency prices and tweets ( Figure 6A and Table 2), Dogecoin prices are weakly correlated to Dogecoin community tweets (r = 0.29), while Litecoin ones are strongly correlated to Litecoin community tweets (r = 0.86). Those results are significant because our R algorithm computed a 0 rounded p-value.
About the association explored with Shannon mutual information entropy ( Figure 6B and Table 3), both Dogecoin and Litecoin links between their community tweet number and their price are strong (0.950 and 0.998, respectively). However, the first one is not as significant (p-value = 0.18) as the second one (p-value = 0.15).
About causality explored with Shannon mutual information entropy ( Table 3): -Dogecoin tweet number was impacting Dogecoin price. In fact, conditional information entropy of Dogecoin price given Dogecoin tweet number (0.724) is higher than conditional information entropy of Dogecoin tweet number given Dogecoin price (0.031). -Litecoin tweet number is impacted by Litecoin price. In fact, conditional information entropy of Litecoin tweet number given Litecoin price (0.037) is higher than conditional information entropy of Litecoin price given Litecoin tweet number. (0.002).

Application
According to previous correlation/causation analysis, we could describe the relationship between currency price P and community tweet number C at a time t (Eq 8 and 9). P dogecoin 0.0445124p C dogecoin + 2.160866 × 10 −0.6 p t − 0.0009299092.
Then, in order to analyze this regression quality, for each cryptocurrency ( Figure 7A, Figure 8A, Table 4.), we have performed ANOVA tests, compared residual data vs. fitted data, computed residual-scale location, performed normality verification by comparing the quantiles of the population with those of the normal law, and compared residual data vs. leverage data.
As the Pr (>F) is always under the alpha threshold (0.05), we reject the null hypothesis and therefore, non-null trends exist for both Litecoin and Dogecoin price. Furthermore, concerning residual data vs. fitted data studies, points are distributed randomly around the horizontal axis y = 0 and show no trend. "Scale-Location" charts show slight trends which, however, are not obvious. According to "QQ-Norm" charts, residues are normally distributed. The last graph, "Residuals vs. Leverage", highlights the importance of each point in the regression; as we can see, there is only one point in each sample with a Cook distance greater than 1 (making the data suspect with that aberrant point). We have also identified the suspect points, which are the points whose studentized residual is greater than 2 in absolute value and/or the Cook's distance is greater than 1 ( Figure 7B, Figure 8B). In the latter case, the point contributes very/too strongly to the determination of coefficients of the model compared to others. However, there is no one-size-fits-all method for dealing with these types of stitches. Thus, a machine learning modelization, as performed then, is required.

ARIMA
As given in Methods, the ARIMA model predicts future values based on past behavior. We considered certain criteria to know the condition of our model after training and the loss of information during training and to select the best model. Minimum loss indicates better training. Choices were automatically made using the "auto.arima" function of the R forecast package (Guasti Lima and Assaf Neto, 2022).  We chose about 0.99% of total Litecoin/Dogecoin price data for training 24,482 samples that we considered to predict all past values of Litecoin/Dogecoin pricing ( Figure 9). Then, we calculated the error (Eq. 10) for each predicted past value based on forecasted and actual values.

IMV-LSTM
About the IMV-LTSM method application on the Dogecoin data ( Figure 10), we have found a very weak match between forecasting and reality (e.g., the price forecasted blue curve is very distant from the real price curve in green). The tweet volume value only participated at a 45% level in the neural network price forecasting. About its application to Litecoin data (Figure 11), this method seems very interesting. After only 166 iterations, we can capture the main trends (e.g., the price forecasted blue curve is very close to the real price curve in green). The tweet volume participated at a level of 29% in the neural network price forecasting.

DISCUSSION
While the average Dogecoin transaction value is impacted by tweets, Litecoin's transaction number and tweets are impacted by the average Litecoin transaction value. About whales, tweets are impacted by Dogecoin whales, but no significant relationship was found between Litecoin whales and tweets. Furthermore, the lack of association was clearly observed using one fundamental approach (mutual information theory), resting on wholly different assumptions and principles: one with classical Pearson's correlation, the other Shannon mutual information (often used in this study field (Piškorec et al., 2014;Keskin and Aste, 2020)). Furthermore, we saw that these two approaches contradicted slightly but mainly, rather nicely, complemented each other, in that Pearson's correlation made it possible to study the sign of the correlation (positive vs negative), while normalized mutual information made it possible to assess the association strength in a finer way, independent from the assumption of monotony required by linear correlation.
About our price study and prediction based on community tweet volume, the most accurate ARIMA models are non-seasonal and are white noise (Mahan et al., 2015). In this way, those nonstationary (e.g., with a log transformation) ARIMA (0,0,0) models are the most accurate to predict cryptocurrencies prices based on their community tweet volume. White noise is a random signal with equal intensities at every frequency and is often defined in statistics as a signal whose samples are a sequence of unrelated, random variables with no mean and limited variance. In this way, in our study, this means that prices are independent between them (confirmed by prices correlation studies). At last, we have performed further analysis based upon Interpretable Multivariable-Long Short-Term Memory neural network application. IMV-LSTM seems interesting for crypto with less volatility such as Litecoin, but ARIMA (0,0,0) seems more interesting for cryptos more volatile such as Dogecoin.
We surmised that the main limitations of our work would most likely be grounded in community size. Indeed, a larger and more active community such as that of a "memecoin" could have a greater impact than a weaker community. In addition, a qualitative impact of some specific tweets by very well-known people could be studied (in a further study) in order to better understand how Twitter may impact cryptocurrency values. In addition, we need to complete our results in the near future with a real comparison between the different causality assessment techniques and a comparison between the various financial prediction techniques (based on the same variables). Very few studies addressing the behavioral impact on cryptocurrencies were made (Akbiyik et al., 2021;Barjašić and Antulov-Fantulin, 2020;Lansiaux et al., 2021;6 ;de Winter et al., 2016). Most of those look exclusively at Bitcoin (Akbiyik   Barjašić and Antulov-Fantulin, 2020;7 ;Piškorec et al., 2014;Kraaijeveld and De Smedt, 2020;Xu and Livshits, 2019;Ante, 2021;Abraham et al., 2018;Li et al., 2019;Beck et al., 2019). The last one is a comparison between Bitcoin and Dogecoin (Tandon et al., 2021); despite their main limitation, Dogecoin and Bitcoin are from different cryptocurrency generations. This study results in the unpredictability of prices when looking at tweets from a community. Therefore, we carried out the first study about behavioral impact analysis between cryptocurrencies. Our study has, thus, proven the interest of behavioral economics applied to the cryptocurrency world like others before it (Tandon et al., 2021;Akbiyik et al., 2021;Barjašić and Antulov-Fantulin, 2020;Lansiaux et al., 2021;4 ;de Winter et al., 2016;Ariyo et al., 2014;Mahan et al., 2015;Piškorec et al., 2014;Kraaijeveld and De Smedt, 2020;Xu and Livshits, 2019;Ante, 2021;Abraham et al., 2018;Li et al., 2019;Beck et al., 2019). However, many things remain to be improved: the application of those models to cryptocurrencies operating under other consensus mechanisms [Proof of Stake (PoS), Delegated Proof of Stake (DPoS), Practical Byzantine Fault Tolerance (PBFT)]. The application to cryptocurrencies with a voting consensus (for such consensus, the formulation of the problem would be more meaningful, and from the point of view of applications to commodity-backed cryptocurrency, their study would be more meaningful) and the creation of an online bot to predict cryptocurrencies future prices. However, before that, we still have to carry out a sensitivity study on all causal inference models on time series data (including our methods) and on all statistical prediction models (both regressive and based on neural networks) and on various cryptocurrencies in order to choose the most accurate. This will be our next study on cryptocurrency price prediction topic.

DATA AVAILABILITY STATEMENT
The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found in the article/Supplementary Material.