Skip to main content

BRIEF RESEARCH REPORT article

Front. Big Data, 13 March 2024
Sec. Data Mining and Management
Volume 7 - 2024 | https://doi.org/10.3389/fdata.2024.1308236

Data pipeline for real-time energy consumption data management and prediction

Jeonghwan Im1 Jaekyu Lee1 Somin Lee2 Hyuk-Yoon Kwon1*
  • 1Graduate School of Data Science, Seoul National University of Science and Technology, Seoul, Republic of Korea
  • 2Department of Global Technology Management, Seoul National University of Science and Technology, Seoul, Republic of Korea

With the increasing utilization of data in various industries and applications, constructing an efficient data pipeline has become crucial. In this study, we propose a machine learning operations-centric data pipeline specifically designed for an energy consumption management system. This pipeline seamlessly integrates the machine learning model with real-time data management and prediction capabilities. The overall architecture of our proposed pipeline comprises several key components, including Kafka, InfluxDB, Telegraf, Zookeeper, and Grafana. To enable accurate energy consumption predictions, we adopt two time-series prediction models, long short-term memory (LSTM), and seasonal autoregressive integrated moving average (SARIMA). Our analysis reveals a clear trade-off between speed and accuracy, where SARIMA exhibits faster model learning time while LSTM outperforms SARIMA in prediction accuracy. To validate the effectiveness of our pipeline, we measure the overall processing time by optimizing the configuration of Telegraf, which directly impacts the load in the pipeline. The results are promising, as our pipeline achieves an average end-to-end processing time of only 0.39 s for handling 10,000 data records and an impressive 1.26 s when scaling up to 100,000 records. This indicates 30.69–90.88 times faster processing compared to the existing Python-based approach. Additionally, when the number of records increases by ten times, the increased overhead is reduced by 3.07 times. This verifies that the proposed pipeline exhibits an efficient and scalable structure suitable for real-time environments.

1 Introduction

With the growing utilization of data in various industries and applications, the construction of efficient data pipelines becomes paramount. These pipelines are responsible for designing the entire process from data collection to data services, ensuring smooth and effective data management. In particular, real-time pipelines have the capability to dynamically collect and process data from Internet of Things (IoT) sensors, enabling them to provide instantaneous and up-to-date services (Gogineni et al., 2015; Rathore et al., 2015; Kalsoom et al., 2020). As a result, real-time data pipelines have gained significant traction in various industries (Gogineni et al., 2015; Rathore et al., 2015; Kalsoom et al., 2020). However, conventional Python scripts commonly used for MLOps often exhibit relatively long processing times in the overall pipeline. Moreover, they demonstrate a significant increase in overhead as the number of records grows. Therefore, the establishment of a real-time data pipeline is necessary to address these challenges.

This study focuses on energy consumption management systems, aiming to design an efficient data pipeline that covers the entire process from data collection to time series prediction models. The critical importance of energy management has led many countries to adopt energy consumption prediction models, which strive to optimize their performance (Ekonomou, 2010; Shapi et al., 2021). Machine learning (ML) has emerged as a powerful approach to tackling complex challenges across various industries, delivering effective solutions. To ensure the seamless operation of ML models, all interconnected system components, including data collection and management, must work in harmony, not just for the ML model itself (Sculley et al., 2014). In this context, machine learning operations (MLOps) have arisen as a viable solution to design an efficient system framework.

In this study, we introduce a specialized MLOps-focused framework that seamlessly integrates the ML model with the data pipeline for real-time energy consumption data management and prediction. In the framework, energy consumption data is collected from IoT sensors, undergoes preprocessing, and is then stored in databases. To deal with real-time collected data without loss, Kafka is employed for data ingestion, and Telegraf consumes the data from Kafka and feeds them to the database. Finally, the time-series prediction models, long short-term memory (LSTM) and seasonal autoregressive integrated moving average (SARIMA), are adopted.

To evaluate the performance of the models, we conducted a comprehensive comparison using a substantial dataset of electric power consumption. The results highlighted the trade-off between speed and accuracy. SARIMA exhibited a remarkable advantage in model learning time, making it suitable for scenarios with strict real-time requirements, while LSTM outperformed SARIMA in prediction accuracy.

To validate the effects of the pipeline, we measured the processing time of the overall pipeline by optimizing the configuration of Telegraf. In particular, we showed that our pipeline has an average end-to-end processing time of only 0.39 s to deal with 10,000 data records at a time in the pipeline and only 1.26 s when expanding it to 100,000 records. This indicates 30.69–90.88 times faster processing compared to the existing Python-based approach. Additionally, when the number of records increases by ten times, the increased overhead is reduced by 3.07 times. This verifies that the proposed pipeline exhibits an efficient and scalable structure suitable for real-time environments.

2 Related work

2.1 Data pipeline

Fu and Soman (2021) introduced a data infrastructure tailored for handling real-time data generated by end-users in Uber. The system employed Apache Kafka for real-time data ingestion and Apache Flink for stream processing. By utilizing these tools, they were able to effectively process and analyze the continuous flow of real-time data, enabling prompt responses to user actions and improving overall system performance. Similarly, Syafrudin et al. (2017) proposed a real-time data processing framework specifically designed for sustainability in the manufacturing sector. Their framework integrated Apache Kafka for detecting data transmitted from sensors, Apache Storm for real-time data processing, and MongoDB for storing and managing the processed data. With this setup, they were able to ensure timely data processing and storage, facilitating efficient decision-making and resource management in the manufacturing context. Both studies showcase the significance of real-time data processing frameworks in various industries. By leveraging powerful tools such as Apache Kafka, Apache Flink, Apache Storm, and MongoDB, these frameworks can handle large volumes of data in real-time, leading to improved operational efficiency, better decision-making, and enhanced user experiences.

2.2 Time series prediction model

Initially, various machine learning methodologies, such as autoregressive integrated moving average(ARIMA)-based statistical methods (Amjady, 2001; Chujai et al., 2013; Fard and Akbari-Zadeh, 2014), support vector machine (SVM) (Mohandes, 2002; Fan et al., 2016), and decision tree (Huang et al., 2016; Mayrink and Hippert, 2016), were commonly employed for predicting energy consumption. These traditional techniques provided valuable insights and paved the way for energy consumption forecasting.

However, in recent years, the field of energy consumption prediction has witnessed a significant shift toward the application of deep learning models. Artificial neural networks (ANN) (Gajowniczek and Zabkowski, 2014; Massana et al., 2015), especially LSTM models (Kong et al., 2017; Kim et al., 2018), which are capable of learning temporary gating functions, have shown remarkable performance in this domain. LSTM's ability to handle sequential data and capture long-term dependencies makes it well-suited for time series prediction tasks. Therefore, there have been various research efforts that utilize LSTM for predicting individual household energy consumption (Yan et al., 2019), gas consumption (Laib et al., 2019), and long-term energy consumption (Wang et al., 2020) have been proposed.

Additionally, convolutional neural networks (CNN) (Amarasinghe et al., 2017; Koprinska et al., 2018) have also demonstrated outstanding results in energy consumption prediction. CNNs excel at extracting local correlations between power spectrums, making them highly effective in analyzing spatial and temporal patterns in energy consumption data.

These deep learning models have revolutionized energy consumption prediction, offering improved accuracy and more robust predictions compared to traditional machine learning methods. However, in practice, we need to consider their excessive costs in an environment equipped with embedded devices such as IoT sensors.

3 Background

3.1 InfluxDB

InfluxDB is a renowned open-source time-series database, developed by InfluxData. Its primary focus lies in efficiently storing time-series data, offering fast processing and high availability for both write and read operations. This makes it a popular choice for applications that heavily rely on time information, such as IoT sensors, smart factories, operation monitoring, and real-time analysis. InfluxDB's versatility, performance, and ease of use make it a prominent tool in the time-series data management landscape.

3.2 Kafka

Kafka is a powerful open-source distributed event streaming platform developed by Apache. Its primary purpose is to capture data in real-time from various sources, including databases and sensors, in the form of a continuous stream of events. The architecture of Kafka revolves around three key components: producers, consumers, and brokers. Producers are responsible for publishing or writing the events into Kafka. Consumers, on the other hand, subscribe to the events and read them from Kafka. Brokers play a crucial role as mediators between producers and consumers. They efficiently manage and group the events into topics, which are essentially named streams of data.

3.3 Zookeeper

Zookeeper is a valuable open-source coordination service developed by Apache. Its main purpose is to simplify the implementation of various tasks in distributed environments, including synchronization and master node selection. In the context of Kafka, Zookeeper plays a crucial role in monitoring the health of the servers and storing the status of topics for both producing and consuming data. When new topics are created, Zookeeper keeps track of their status and ensures that consumers and producers are informed about the existence of these new topics. Zookeeper works closely with Kafka brokers to maintain a reliable and up-to-date system. It notifies consumers and producers about the creation of new Kafka brokers and any potential failures that may occur.

3.4 Telegraf

Telegraf is an open-source plugin developed by InfluxData, designed to efficiently collect and deliver events from databases, systems, and IoT sensors. One of the significant advantages of Telegraf is its versatility, as it can be seamlessly integrated into various server applications like InfluxDB and Kafka. This flexibility allows Telegraf to be easily incorporated into existing data flows, enabling users to leverage its capabilities without major disruptions to their current setups. By acting as a powerful data collection and delivery tool, Telegraf contributes to enhancing data management, analysis, and real-time processing.

3.5 Grafana

Grafana is a powerful open-source toolkit developed by Grafana Labs, specifically designed to offer a comprehensive dashboard for visualizing time-series data. One of Grafana's key strengths lies in its compatibility with multiple databases, including ElasticSearch, InfluxDB, and PostgreSQL. This enables users to seamlessly connect their data sources and create insightful visualizations. Grafana serves as an invaluable tool for data visualization and monitoring, offering a user-friendly interface and seamless integration with diverse data sources and notification systems.

4 Proposed framework

4.1 Data ingestion pipeline

Figure 1 illustrates the overall architecture of the proposed framework. The data ingestion pipeline is highlighted in blue, while the time-series prediction model is highlighted in green. The data ingestion pipeline consists of two main parts: data collection and data ingestion. The data collection, which is the initial stage of the architecture, involves gathering energy consumption data from IoT sensors. The collected data is transmitted to the central server. Subsequently, the Kafka source connector monitors the specified path and sends the data to the Kafka brokers for further processing. Once the data is in Kafka, it undergoes several processing steps. The processing steps occur before the data is transmitted to the main database, InfluxDB, after being consumed by Telegraf from Kafka. These steps consist of (1) parsing, (2) transformation, and (3) reconstruction. The parsing step involves decoding the collected sensor data into a usable format. The raw sensor data is encoded as a nested JSON in a single string, containing a timestamp, identifier, and sensor measurement values. To insert the encoded string data into the database, it is decoded to the individual attributes. The transformation step converts the decoded attributes from the parsing step into attribute types defined in the database schema. For example, it performs the conversion of a timestamp, initially formatted as a string, into a Unix timestamp defined in the database. The reconstruction step involves restructuring the original data for efficient data storage. Storing a nested JSON value directly in the database can lead to a degradation in query performance. Therefore, the nested data is appropriately restructured into a more manageable format in the database. For instance, sensor data incoming at the district level is reconstructed at the household level. These processes enhance the overall stability of the system, enabling efficient data storage and queries.

Figure 1
www.frontiersin.org

Figure 1. The overall architecture of the proposed framework.

To avoid an excessive accumulation of data in InfluxDB that could result in a decline in query performance and resource wastage, data older than a certain time threshold is compressed in the Apache Parquet format. These compressed files can be accessed and visualized in Grafana.

4.2 Time-series prediction model

To begin the prediction process, data is retrieved from both the Parquet files and InfluxDB. The retrieved data then undergoes extraction and preprocessing steps to prepare it for model training.

In the extraction step, the sensor data utilized for model training is extracted from the entire dataset1 and is passed through the process of model training. We read the target data used for training from Parquet files and InfluxDB. The read data is loaded into a Pandas dataframe, facilitating the training process. In the preprocessing step, we perform preprocessing on the sensor data to prepare it for model training. Three main processes are executed. First, we fill out the missing values. Sensor data often contains missing values, which can lead to degraded performance during model training and prediction. We replace missing values with the most recent value. Second, we adjust the time unit for model prediction. Before conducting model training and prediction, it is essential to adjust the time unit of the collected data to that for model training and prediction. In this step, we aggregate the collected sensor data to the time unit for prediction. For instance, if the collected data is at a 1-min interval and the prediction unit is at a 1-h interval, the collected data is aggregated into 1-h intervals. Third, we structure the original data into sliding window units. For time series prediction models, a sliding window configuration is crucial for training and prediction. This process involves grouping the continuous time series data for each time point into an array format. For example, if there is time series data with a time sequence from 1 to 15, and the sliding window size is 10, the sliding windows would be structured as 1–10, 2–11, ..., and 6–15.

The chosen prediction models, SARIMA (Valipour, 2015) and LSTM (Hochreiter and Schmidhuber, 1997), are trained using preprocessed data. SARIMA incorporates seasonal components in addition to ARIMA, often used to identify or impose regular patterns in time series data, especially in data types with periodicity. Temperatures are representative features with periodicity. Because energy consumption is significantly correlated with temperature, SARIMA is suitable for considering such seasonality. LSTM has the ability to retain information for a period and is beneficial for processing a long sequence. Because time series data is a prominent example of sequence data, LSTM is a suitable candidate for predicting time series data.

To monitor the models' ongoing performance, the evaluation metrics and prediction results are stored in the database. This enables easy access to the historical performance of the models and facilitates analysis and decision-making. The prediction process is applied to predict energy consumption in real-time, allowing for timely and informed management decisions. To capture recent trends in the data and account for changes in the system, the entire prediction process is periodically automated. This periodic automation ensures that the models remain up-to-date and can adapt to dynamic changes in the energy consumption data. By integrating SARIMA and LSTM models within the MLOps-based framework, the system can take advantage of both statistical time-series modeling and deep learning-based techniques. This comprehensive approach aims to optimize prediction efficiency while targeting real-time service requirements in energy management.

5 Performance evaluation

5.1 Experimental environments

For the experiments, we used a server equipped with Intel Xeon Silver 4210R CPU 2.40 GHz, 32 GB of memory, and Nvidia RTX A5000, where Ubuntu OS is installed. We use the following versions of the software: Python 3.6.9, Kafka 5.5.0, Zookeeper 3.4.9, InfluxDB 1.8.2, Telegraf, 1.19, and Grafana 7.2.0.

5.2 Experimental datasets and methods

The dataset (Hebrail and Berard, 2012) used in this study was collected with a one-minute sampling rate over a period between December 2006 and November 2010, encompassing a total of 47 months. It is a multivariate time-series data consisting of six independent variables, such as electrical quantities and sub-metering values, along with a numerical dependent variable, i.e., global active power. The dataset contains a total of 2,075,259 observations. During preprocessing, the data was resampled to an hourly frequency from the original one-minute sampling rate. The first 3 years of data (36 months) were used for model training, while the remaining data (11 months) were reserved for validation purposes.

To measure the prediction accuracy of the models, two types of metrics were employed: Mean Absolute Error (MAE) as shown in Equation (1), and Root Mean Squared Error (RMSE) as shown in Equation (2). MAE measures the average magnitude of the errors, ei, in a set of predictions, without considering their direction. It is the average over the test sample of the absolute differences between prediction and actual observation where all individual differences have equal weight. It is a suitable indicator when the loss increases linearly. RMSE represents the standard deviation of the prediction errors, ei. Since the error is squared, the larger the error, the higher the weight accordingly is reflected.

MAE=1ni=1n|ei|    (1)
RMSE=1ni=1nei2    (2)

For the hyperparameters of LSTM, ADAM (Kingma and Ba, 2014) was chosen as the optimizer, with a learning rate of 0.001, and MAE was used as the loss function. ADAM is a learning optimization algorithm that dynamically adjusts the learning rates for individual parameters, striking a balance between the global and local optima. It offers the advantage of relatively swift learning rates and excellent optimization performance. As for the SARIMA model, the ARIMA order was set as (1,1,1), and the seasonal order was set as (1,0,1,24). These parameter settings were determined based on the lowest Akaike Information Criterion (AIC) (Burnham and Anderson, 2004), which is a measure of the relative quality of statistical models for a given set of data. AIC represents the similarity between actual data and predicted data, which tends to be higher when there are more independent variables used in the estimation. As for AIC, it increases when unnecessary parameters are added to improve fitness, so a lower AIC suggests a better model with fewer unnecessary parameters. It is important to note that the results may vary depending on the chosen hyperparameters for both SARIMA and LSTM models. The presented hyperparameter settings were specifically chosen for this study to provide insights into the prediction accuracy of the models on the resampled data.

5.3 Experimental results for a prediction model

Figure 2 shows the predicted results of the energy consumption using SARIMA and LSTM. Table 1 presents the predicted results of SARIMA and LSTM. Overall, in terms of accuracy, LSTM outperforms SARIMA, but it is essential to highlight that SARIMA still demonstrates acceptable accuracy. In regard to model learning time, SARIMA shows a significant advantage over LSTM, being ~30.58 times faster. This indicates that SARIMA could be a more desirable choice when the prediction model needs to run on resource-constrained IoT devices or when frequent updates of the prediction model are required. On the other hand, LSTM proves to be more desirable in scenarios where more accurate predictions are needed and computational resources are less constrained, such as server environments.

Figure 2
www.frontiersin.org

Figure 2. Visualization of predicted results of LSTM and SARIMA. (A) LSTM and (B) SARIMA.

Table 1
www.frontiersin.org

Table 1. Comparing model learning time and accuracy.

5.4 Experimental results for a data pipeline

To optimize the performance of the entire pipeline, Telegraf plays a crucial role in consuming data from Kafka. To assess the scalability of the data pipeline, we conducted tests by generating artificial data at a rate of 500 records per second, randomly selected from the real-world dataset. To observe the overall processing time, we adjusted the flush interval of Telegraf from 10 to 1 s, loading 10,000 data records. Figure 3 shows the results of the overall processing time for the entire data pipeline of the proposed framework. The results showed that the smallest flush interval (1 s) yielded the fastest performance. Specifically, with a flush interval of 10 s, the average end-to-end processing time was 35.33 s. This time decreased significantly to 12.84 and 0.39 s when the flush interval was reduced to 5 and 1 s, respectively. It is worth noting that these improvements reached a saturation point. In this setup, even with an increased data load of 100,000 records, the processing time remained impressively low at only 1.26 s. This demonstrates the framework's ability to efficiently handle large volumes of data. For comparative experiments to show the effectiveness of the proposed method, we implement a Python-based script that performs the same function. The Python script performs data loading and processing in the same manner as the proposed pipeline. Specifically, it loads and parses json files, performs transformation and reconstruction steps, and then stores them in InfluxDB. The experiments using the Python script with the same data show an average processing time of 11.57 s for 10,000 records and 114.51 s for 100,000 records, respectively. This indicates that the proposed pipeline processes data 30.69 times faster for 10,000 records and 90.88 times faster for 100,000 records compared to the Python script, respectively. That is, with a tenfold increase in the number of records, the average processing time for the Python script increases by 9.90 times, while the proposed pipeline increases by only 3.23 times. This shows that the proposed pipeline is a highly efficient pipeline suitable for real-time environments, demonstrating superior scalability.

Figure 3
www.frontiersin.org

Figure 3. Average processing time from data collection to database.

6 Conclusions

In this study, we presented a comprehensive data pipeline that facilitates real-time data collection from IoT sensors, storage in databases, and time-series prediction models. Through the experiments, we showed the trade-off between the accuracy and efficiency of the ML-based prediction models under the proposed pipeline. Our pipeline showed 30.69–90.88 times faster processing than the existing Python-based approach. Additionally, when the number of records increased by ten times, the increased overhead was reduced by 3.07 times. This verifies that the proposed pipeline exhibits an efficient and scalable structure suitable for real-time environments. Given the limited research efforts in this domain, our proposed framework showcases its effectiveness in energy consumption management.

Based on the findings and insights gained from this study, we have identified two future research directions for improving the current framework. First, we need to enhance the coordination and optimization of the components within the pipeline to handle larger workloads more efficiently. As the system scales and deals with increased data volumes, it becomes crucial to ensure seamless interactions among different components and maximize resource utilization, particularly for real-time applications. Second, in this study, SARIMA and LSTM, which are simple yet effective predictive models considering the real-time nature of the proposed pipeline, were utilized. The proposed pipeline can also be extended with state-of-the-art predictive models such as MTAD-GAT (Zhao et al., 2020) and TFT (Lim et al., 2021). However, not only prediction performance but also computational overheads such as model training/inference time and resource usage should be considered. By addressing the trade-off, we aim to choose the most appropriate predictive model for a given environment.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

JI: Data curation, Formal analysis, Investigation, Methodology, Resources, Visualization, Writing – original draft. JL: Data curation, Investigation, Methodology, Resources, Writing – original draft. SL: Data curation, Investigation, Methodology, Resources, Writing – original draft. H-YK: Conceptualization, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Supervision, Validation, Writing – review & editing.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This work was supported in part by the Korea Institute of Energy Technology Evaluation and Planning (KETEP) and the Ministry of Trade, Industry & Energy (MOTIE) of the Republic of Korea (No. 2021202090028D), and in part by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (No. 2019R1A6A1A03032119).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Footnotes

References

Amarasinghe, K., Marino, D. L., and Manic, M. (2017). “Deep neural networks for energy load forecasting,” in 2017 IEEE 26th International Symposium on Industrial Electronics (ISIE) (Edinburgh: IEEE), 1483–1488.

Google Scholar

Amjady, N. (2001). Short-term hourly load forecasting using time-series modeling with peak load estimation capability. IEEE Transact. Power Syst. 16, 498–505. doi: 10.1109/59.932287

Crossref Full Text | Google Scholar

Burnham, K. P., and Anderson, D. R. (2004). Model Selection and Multimodel Inference. A Practical information-Theoretic Approach (New York, NY: Springer), 2.

Google Scholar

Chujai, P., Kerdprasop, N., and Kerdprasop, K. (2013). “Time series analysis of household electric consumption with arima and arma models,” in Proceedings of the International Multiconference of Engineers and Computer Scientists, Vol. 1 (Hong Kong: IAENG Hong Kong), 295–300.

Google Scholar

Ekonomou, L. (2010). Greek long-term energy consumption prediction using artificial neural networks. Energy 35, 512–517. doi: 10.1016/j.energy.2009.10.018

Crossref Full Text | Google Scholar

Fan, G.-F., Peng, L.-L., Hong, W.-C., and Sun, F. (2016). Electric load forecasting by the svr model with differential empirical mode decomposition and auto regression. Neurocomputing 173, 958–970. doi: 10.1016/j.neucom.2015.08.051

Crossref Full Text | Google Scholar

Fard, A. K., and Akbari-Zadeh, M.-R. (2014). A hybrid method based on wavelet, ann and arima model for short-term load forecasting. J. Exp. Theoret. Artif. Intell. 26, 167–182. doi: 10.1080/0952813X.2013.813976

Crossref Full Text | Google Scholar

Fu, Y., and Soman, C. (2021). “Real-time data infrastructure at uber,” in Proceedings of the 2021 International Conference on Management of Data (New York, NY), 2503–2516.

Google Scholar

Gajowniczek, K., and Zabkowski, T. (2014). Short term electricity forecasting using individual smart meter data. Proc. Comput. Sci. 35, 589–597. doi: 10.1016/j.procs.2014.08.140

Crossref Full Text | Google Scholar

Gogineni, V. R., Matcha, K., and Rao K, R. (2015). Real time domestic power consumption monitoring using wireless sensor networks. Int. J. Elect. Comp. Eng. 5:685. doi: 10.11591/ijece.v5i4.pp685-694

Crossref Full Text | Google Scholar

Hebrail, G., and Berard, A. (2012). Individual Household Electric Power Consumption. UCI Machine Learning Repository.

Google Scholar

Hochreiter, S., and Schmidhuber, J. (1997). Long short-term memory. Neural Comput. 9, 1735–1780. doi: 10.1162/neco.1997.9.8.1735

Crossref Full Text | Google Scholar

Huang, N., Lu, G., and Xu, D. (2016). A permutation importance-based feature selection method for short-term electricity load forecasting using random forest. Energies 9:767. doi: 10.3390/en9100767

Crossref Full Text | Google Scholar

Kalsoom, T., Ramzan, N., Ahmed, S., and Ur-Rehman, M. (2020). Advances in sensor technologies in the era of smart factory and industry 4.0. Sensors 20:6783. doi: 10.3390/s20236783

PubMed Abstract | Crossref Full Text | Google Scholar

Kim, N., Kim, M., and Choi, J. K. (2018). “Lstm based short-term electricity consumption forecast with daily load profile sequences,” in 2018 IEEE 7th Global Conference on Consumer Electronics (GCCE) (Nara: IEEE), 136–137.

Google Scholar

Kingma, D. P., and Ba, J. (2014). Adam: a method for stochastic optimization. arXiv [preprint]. doi: 10.48550/arXiv.1412.6980

Crossref Full Text | Google Scholar

Kong, W., Dong, Z. Y., Jia, Y., Hill, D. J., Xu, Y., and Zhang, Y. (2017). Short-term residential load forecasting based on lstm recurrent neural network. IEEE Trans. Smart Grid 10, 841–851. doi: 10.1109/TSG.2017.2753802

Crossref Full Text | Google Scholar

Koprinska, I., Wu, D., and Wang, Z. (2018). “Convolutional neural networks for energy time series forecasting,” in 2018 International Joint Conference on Neural Networks (IJCNN) (Rio de Janeiro: IEEE), 1–8.

Google Scholar

Laib, O., Khadir, M. T., and Mihaylova, L. (2019). Toward efficient energy systems based on natural gas consumption prediction with lstm recurrent neural networks. Energy 177, 530–542. doi: 10.1016/j.energy.2019.04.075

Crossref Full Text | Google Scholar

Lim, B., Arık, S. Ö., Loeff, N., and Pfister, T. (2021). Temporal fusion transformers for interpretable multi-horizon time series forecasting. Int. J. Forecast. 37, 1748–1764. doi: 10.1016/j.ijforecast.2021.03.012

Crossref Full Text | Google Scholar

Massana, J., Pous, C., Burgas, L., Melendez, J., and Colomer, J. (2015). Short-term load forecasting in a non-residential building contrasting models and attributes. Energy Build. 92, 322–330. doi: 10.1016/j.enbuild.2015.02.007

Crossref Full Text | Google Scholar

Mayrink, V., and Hippert, H. S. (2016). “A hybrid method using exponential smoothing and gradient boosting for electrical short-term load forecasting,” in 2016 IEEE Latin American Conference on Computational Intelligence (LA-CCI) (Cartagena: IEEE), 1–6.

Google Scholar

Mohandes, M. (2002). Support vector machines for short-term electrical load forecasting. Int. J. Energy Res. 26, 335–345. doi: 10.1002/er.787

Crossref Full Text | Google Scholar

Rathore, M. M., Ahmad, A., Paul, A., and Daniel, A. (2015). “Hadoop based real-time big data architecture for remote sensing earth observatory system,” in 2015 6th International Conference on Computing, Communication and Networking Technologies (ICCCNT) (Dallas-Fortworth, TX: IEEE), 1–7.

Google Scholar

Sculley, D., Holt, G., Golovin, D., Davydov, E., Phillips, T., Ebner, D., et al. (2014). Machine Learning: The High Interest Credit Card of Technical Debt. La Jolla, CA: NeurIPS.

Google Scholar

Shapi, M. K. M., Ramli, N. A., and Awalin, L. J. (2021). Energy consumption prediction by using machine learning for smart building: case study in malaysia. Dev. Built Environ. 5:100037. doi: 10.1016/j.dibe.2020.100037

Crossref Full Text | Google Scholar

Syafrudin, M., Fitriyani, N. L., Li, D., Alfian, G., Rhee, J., and Kang, Y.-S. (2017). An open source-based real-time data processing architecture framework for manufacturing sustainability. Sustainability 9:2139. doi: 10.3390/su9112139

Crossref Full Text | Google Scholar

Valipour, M. (2015). Long-term runoff study using sarima and arima models in the united states. Meteorol. Appl. 22, 592–598. doi: 10.1002/met.1491

Crossref Full Text | Google Scholar

Wang, J. Q., Du, Y., and Wang, J. (2020). Lstm based long-term energy consumption prediction with periodicity. Energy 197:117197. doi: 10.1016/j.energy.2020.117197

Crossref Full Text | Google Scholar

Yan, K., Li, W., Ji, Z., Qi, M., and Du, Y. (2019). A hybrid lstm neural network for energy consumption forecasting of individual households. IEEE Access 7, 157633–157642. doi: 10.1109/ACCESS.2019.2949065

Crossref Full Text | Google Scholar

Zhao, H., Wang, Y., Duan, J., Huang, C., Cao, D., Tong, Y., et al. (2020). “Multivariate time-series anomaly detection via graph attention network,” in 2020 IEEE International Conference on Data Mining (ICDM) (Sorrento: IEEE), 841–850.

Google Scholar

Keywords: energy consumption, MLOps-centric data pipeline, time-series forecasting, real-time data pipeline, scalable pipeline

Citation: Im J, Lee J, Lee S and Kwon H-Y (2024) Data pipeline for real-time energy consumption data management and prediction. Front. Big Data 7:1308236. doi: 10.3389/fdata.2024.1308236

Received: 06 October 2023; Accepted: 01 March 2024;
Published: 13 March 2024.

Edited by:

Jianwu Wang, University of Maryland, Baltimore County, United States

Reviewed by:

Weilong Ding, North China University of Technology, China
Junqi Yin, Oak Ridge National Laboratory (DOE), United States

Copyright © 2024 Im, Lee, Lee and Kwon. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Hyuk-Yoon Kwon, hyukyoon.kwon@seoultech.ac.kr

Download