Skip to main content

ORIGINAL RESEARCH article

Front. Psychol., 04 May 2022
Sec. Human-Media Interaction

Design of Aging Smart Home Products Based on Radial Basis Function Speech Emotion Recognition

  • 1School of Art and Design, Tianjin University of Technology, Tianjin, China
  • 2School of Control and Mechanical Engineering, Tianjin Chengjian University, Tianjin, China

The rapid development of computer technology and artificial intelligence is affecting people’s daily lives, where language is the most common way of communication in people’s daily life. To apply the emotion information contained in voice signals to artificial intelligence products after analysis, this article proposes a design based on voice emotion recognition for aging intelligent home products with RBF. The authors first aimed at a smart home design, and based on the problem of weak adaptability and learning ability of the aging population, a speech emotion recognition method based on a hybrid model of Hidden Markov/Radial Basis Function Neural Network (HMM/RBF) is proposed. This method combines the strong dynamic timing modeling capabilities of the HMM model and the strong classification decision-making ability of the RBF model, and by combining the two models, the speech emotion recognition rate is greatly improved. Furthermore, by introducing the concept of the dynamic optimal learning rate, the convergence speed of the network is reduced to 40.25s and the operation efficiency is optimized. Matlab’s simulation tests show that the recognition speed of the HMM/RBF hybrid model is 9.82–12.28% higher than that of the HMM model and the RBF model alone, confirming the accuracy and superiority of the algorithm and model.

Introduction

The rapid rise of artificial intelligence brings great convenience to people’s daily life. To make smart products more intelligent and humanized and the user experience more comfortable, the new human–computer interaction technology has gradually attracted the attention of the majority of researchers (Tao et al., 2021). Also, research by Reevess and Nass at Stanford University found that the key problem that “emotional intelligence” needs to solve is the ability of emotional intelligence; this is actually consistent with solving the problems of human–computer interaction and human-to-human communication (Peng et al., 2021). Therefore, to enable the computer actively and accurately perceive the user’s emotional state and natural human–computer interaction, first, the computer must be able to analyze and recognize the user’s emotional state, and then the way the computer interacts with the user can be determined and adjusted to have a more intelligent human–computer interaction (Xu and Wan, 2021). At present, there are various researches on emotional information processing of speech signals. It mainly includes the preprocessing of speech signals, the extraction of emotional feature parameters, the establishment and training of emotional models, etc. The classic AI interaction process is shown in Figure 1.

FIGURE 1
www.frontiersin.org

Figure 1. Schematic diagram of the artificial intelligence interaction process.

Language is an important means of human communication, and it is the most effective, natural, and direct way for human beings to exchange ideas, thoughts, and emotions. Emotional information can be transmitted not only through speech signals but also through facial expressions and physiological responses. Acoustic representation of language is speech, and speech signals contain a wealth of emotional information. In the case of effective calculations, the impact information contained in the speech is an important source of information for impact calculations. For example, if the same sentence is spoken with different emotions, the emotions expressed are completely different. For example, for the sentence “I will go even if it rains,” under two different emotions of anger and sadness the meanings expressed are very different. This is the so-called “obedience to listen to the sound” truth. Therefore, the voice file of the speaker is collected, analyzed for emotional features contained in speech signals, and this allows the computer to accurately assess the speaker’s emotional state, which has become an important area of research in the field of speech emotion recognition (El-Henawy and Abo-Elazm, 2020; Chatterjee et al., 2021). Speech emotion recognition is to preprocess the speech signal to be recognized, extract feature parameters with emotional tendencies, and through the established emotional model the characteristic parameters are analyzed and identified, a technique used to determine the emotional state of a speaker. By investigating the literature related to speech emotion recognition, the authors analyzed and studied the application of HMM and RBF models in speech emotion recognition, and an improved HMM/RBF hybrid model is proposed starting from the learning process of RBF neural network, introducing dynamic optimal learning rate, improving the stability, and converging the speed of the RBF network.

Literature Review

In recent years, scholars have researched on speech emotion recognition that originated in the United States. Williams and Stevens were the first to study the emotional information rich in speech, and it was found that the change of the pitch frequency in the speech signal is related to emotion. Deschamps-Berger et al. (2021) recognized it is possible to decode the speech signal frequency contour parameters and recognize speech emotions on four emotions using three different classification algorithms: the nearest neighbor method, the nuclear regression method, and the most probable Bayesian classification method, and obtain the highest recognition rate of about 68%. Sisavath and Yu (2021) analyzed 700 short sentences composed of five emotions (happiness, sadness, anger, fear, and naturalness), and different algorithms were used to differentiate, classify, and recognize the four emotional characteristics parameters (speech speed, shape, base frequency, and energy) and reach an acceptable level of about 70%. Qi et al. (2021) divided speech emotions into positive and negative categories, and two algorithms of linear discrimination and nuclear regression were used compared with Petrushin’s experiment. Li (2021) of the Technical University of Munich developed an open-source toolkit open SMILE, with which the function of batch automatic extraction of speech emotion feature parameters is realized, including common emotional characteristic parameters such as amplitude, fundamental frequency. Zhu et al. (2021) has gradually been Mel cepstral coefficient, and has gradually been widely recognized. Yang et al. (2021) used non-linear dynamic features (NLDs) to classify highly confusing emotions, such as anger and joy. The definitions of basic emotions by different scholars are shown in Table 1 below.

TABLE 1
www.frontiersin.org

Table 1. Definitions of basic emotions.

The International Speech Communication Association, abbreviated as ISCA, plays an important role in the study of speech emotions, and scholars from various countries discussed technical problems related to affective computing. ISCA now holds the Eurospeech or Interspeech International Conference every 2 years, and this has greatly contributed to the progress of research in the field of speech emotion recognition. In addition, many universities and scientific research institutions have also done a lot of constructive research work in speech emotion recognition, for example, Cameron University, University of Cambridge, University of Birmingham, and the Japan Research Institute ATR (Ren et al., 2019). In addition, many companies have successively established teams to study emotional computing and artificial intelligence, such as Microsoft, IBM, Google, Sony, etc. The research on speech emotion recognition in China started relatively late, and it is still in its infancy. In recent years, many universities, research institutes, and laboratories have conducted research on speech emotion recognition. The current study of emotion in mature language is based on the study of local culture and language, and Chinese scholars obviously cannot copy foreign research methods, so it is necessary to develop a recognition method that conforms to Chinese speaking habits (Agarwal and Om, 2021). This is because Chinese are not only semantically different from Western countries’ emphasis on light and stress but there are also differences in tone changes. Southeast University was the first to carry out research on speech emotion recognition, among them, Chen (2021) proposed a method of speech emotion recognition using three methods based on principal element analysis, and good recognition results have been achieved. The research team of Zhejiang University established a corpus MASC@CCNT, recorded in Mandarin by 68 young men and women, which included five emotional states, it has laid a solid foundation for the follow-up speech emotion recognition research. Liu and Zhang (2021) proposed a method to measure the effectiveness of emotional feature parameters by using fuzzy entropy. Abbas and Abdullah (2021), taking the prosodic feature as the extracted sentiment feature parameter, performed emotion recognition research on speech signals using two different algorithms, GMM and PNN algorithms, respectively. Professor Bao Canglong of Datong University in Taiwan led an experimental group to establish a Mandarin emotional speech database, the database includes five emotions (Mustaqeem Kwon, 2021). A full-attenuated neural network and a classification and regression tree (CART) algorithm are used to build an emotional model, analyze, and compare the acoustic and prosody features; experiments have shown that acoustic properties have a positive effect on the recognition of speech emotions.

With the development of speech recognition technology, more and more young scholars are also involved in this issue studied and published in the development history (Puri et al., 2022), application direction, technical research content, and key technologies required for emotional computing (Tarnowski et al., 2020). E-learning Lech et al. (2020) is also known as online learning through the Internet, and this also includes reading e-books, distance education is arranged through multimedia playback software, through the Internet, construct virtual online electronic classrooms and online digital libraries and other interactive methods. Liu et al. (2021) conducted research on the semantic aspects of emotional information extracted from images, humans comprehend and perceive images in a subjective way. Shemeis et al. (2021) proposed the Multi-Agent System architecture, the application of this system can realize the recognition of human emotional information, at the same time, it can also express the individual emotions of people in all aspects through the virtual way. Gu Xuejing and Shi Zhiguo proposed an emotional robot based on BDI Agent technology, where they mainly introduced the application of BDI Agent technology in emotional robots by using Agent-based technology to build a speech emotion recognition model, realize that the machine can correctly recognize the input of the external voice. Zhang Yanduo and Wu Hua et al. realized that robot dance can be better combined with music and the robot quickly choreographs its corresponding dance according to the rhythm of the music. This is by extracting the emotional feature parameters in music as the recognition basis, and using the dance system choreographed by robots as the application research object to extract the emotional feature parameters for automatic identification of music, a robot dance choreography system was implemented. Wang S. et al. (2021) proposed an emotion model based on emotion classification. Wang Z. et al. (2021) aimed at how computers can perform emotional calculations A probabilistic model of emotion space is proposed and computer simulation experiments are carried out on it. Wang Y. et al. (2021) adopted the basic theory based on Hidden Markov Model modeling, design an emotion recognition model. Song et al. (2021) summarize and study the current development of the use of speech emotion recognition, the architecture of the recognition model architecture of control robot based on speech emotion is realized, and verified by simulation experiments.

Research Methods

Preprocessing of Speech Signals

The original voice signal is processed to obtain computer-recognizable voice sample data, remove, or reduce the effects of noise in speech signals to obtain ideal speech data for training and recognition of emotion model, this process is the preprocessing of the speech signal. The preprocessing of speech signals is the first step in speech emotion recognition; this generally includes pre-highlighting, windowing for each frame, and endpoint detection, as shown in Figure 2 below.

FIGURE 2
www.frontiersin.org

Figure 2. The process of speech signal preprocessing.

Pre-emphasis

It is difficult to obtain the spectrum of the high-frequency part from the collected original voice signal as this is influenced by the speaker’s nasal radiation and glottic excitation. The high-frequency part generally drops by 6 dB/octave above 800 Hz, therefore, it is necessary to strengthen the high-frequency part of the speech signal, which is of preliminary importance. The high-frequency portion of the pre-stressed speech signal is highlighted and has certain characteristics, and the spectrum is flat, which makes it easy to analyze the spectrum.

In general, first-class FIR high-pass digital filters are used, and the transfer function is as follows:

H ( z ) = 1 - α z - 1 (1)

Here, α is a predictive coefficient, and the value of α ranges from 0.9 to 1.0.

Assuming that the value of the n-sample speech sample is x(n), then:

y ( n ) = x ( n ) - α x ( n - 1 ) (2)

Among them, y(n) is the result after pre-emphasis processing, and is a discrete speech signal.

Framing Windowing

The windowing of each speech signal frame is done after the pre-highlighting process. Speech signals have no short-term motion, i.e., in short-term motion the time (10–30 ms) appears to be constant and is determined by the inertial motion of the organ. In addition, although the speech signal is time-varying, it can be regarded as steady-state because the emotional characteristic parameters of the speech signal may appear unchanged in a short period of time. The framing operation of the voice signal is to divide the voice signal into some short segments during processing, an implementation method of weighting using a movable window of finite length.

However, once this short-term limit is exceeded, the characteristic parameters may change to take some between two adjacent frames’ overlapping part. With overlapping segments, the second half of the previous frame, the first half of the next frame overlapping (also called frame-shift), the continuity of the frame is maintained, which reflects the correlation between the frame and the frame data. The frame-to-frame ratio is generally 0–0.5 and the audible signal range is shown in Figure 3.

FIGURE 3
www.frontiersin.org

Figure 3. Windowing of speech signals.

For framing processing of speech signals, some kind of operation or transformation is performed on each frame of the speech signal. The expression of the output xw(n) after the speech signal x(n) is framed and windowed is as follows:

x w ( n ) = x ( n ) * w ( n ) (3)

In the formula, w (n) is the window function.

There are three commonly used window functions: Hamming windows, rectangular windows, and Hanning windows. This article selected the Hamming window function, which can reflect better the spectral characteristics of the speech signal. It is defined as follows:

w ( n ) = { 0 , n < 0 | n > N 0.54 - 0.46 cos ( 2 π n N - 1 ) , 0 n N (4)

Feature Extraction of Speech Signal

Since the original speech signal contains a large amount of emotional information such as intonation, emotion, and rhythm, the extractable emotional feature parameters are also diverse. So if you want to study speech emotion recognition, the first key question is how to get from the many emotional feature parameters, select valid feature parameters, build vectors that reflect personal emotional characteristics, and get better speech emotion recognition rate. An important selection strategy is: to extract the emotional feature parameters that are easier to improve the speech emotion recognition rate as much as possible and reduce those useless redundant information in the speech signal.

Pitch Frequency

The authors use the method of short-term autocorrelation function to obtain the fundamental frequency, and the pitch

frequency can be obtained by reciprocating the gene period. The short-term autocorrelation function nR is expressed as follows:

R n ( k ) = m = 0 N k - 1 x n ( m ) - x n ( m - k ) (5)

where x(m) is the voice signal, xn(m) is the nth windowed voice signal intercepted, N is the window length, and k is the amount of time delay.

The authors use the sentence “even if it rains” in the CASIA corpus, analyze the pitch frequency under the emotions of anger and happiness, as shown in Figures 4, 5 below.

FIGURE 4
www.frontiersin.org

Figure 4. The pitch frequency of anger emotion.

FIGURE 5
www.frontiersin.org

Figure 5. Pitch frequency of happy emotion.

Formant

A formant is a set of resonant frequencies produced by quasi-periodic impulses acting on our vocal tract and it is important to recognize the emotional state and quality of speech in Chinese. Formant is an important parameter of emotional traits in the study of speech emotion recognition. Formant is also an important emotion feature parameter in speech emotion recognition research. The spectral envelope of the speech signal contains formant information, the maximum value of which is the formant; therefore, there is a key point in extracting formant characteristic parameters, i.e., to estimate the spectral envelope of the speech signal. There are two extraction methods for formant characteristic parameters, one is the linear prediction method, and the other is the cepstral method. The authors use the standard complex root method to find the formant parameters. The method is to calculate the root of the prediction error filter A(z) and then the resonance peak can be determined according to the obtained complex root. At the same time, the bandwidth iB and the center frequency iF of the formant can be obtained by decomposing the A(z) polynomial coefficients. The complex roots of the polynomial A(z) can be found using the Newton-Raphson algorithm, where H(z) is related to A(z) as follows:

H ( z ) = G A ( z ) - G 1 - i = 1 p a i z - i (6)

α_i is the linear prediction coefficient, ip = 1,2,…., p is the order, then the relationship is as follows:

2 π θ F i = e - B i T π - r i (7)

where Fi is the formant frequency, the 3 dB bandwidth is Bi, T is the sampling period, and the equations are as follows:

F i = θ i / ( 2 π T ) (8)
B i = - ln r i / π T (9)

Since the number of complex conjugate pairs is at most half of the order p in the prediction error filter A(z), and the order p is preset, so the problem of determining the formant to which the pole belongs is not complicated. The speech signal uses the emotionally angry sentence in CASIA “even if it rains,” the vertical point on the power spectrum curve is the position of the formant frequency, and a total of 4 formants were detected. Their formant frequencies Fi and bandwidths Bi are shown in Table 2 below.

TABLE 2
www.frontiersin.org

Table 2. Formant frequency and bandwidth values.

The formant frequency is obtained using the LPC root-finding method, as shown in Figure 6 below.

FIGURE 6
www.frontiersin.org

Figure 6. Power spectrum curve of channel transfer function.

Linear Prediction Cepstral Coefficient

In the field of speech emotion recognition, the linear hypothesis coefficient (LPCC) is used as a characteristic of speech emotion recognition emotion, and LPCC can be derived from the linear prediction coefficient (LPC). The main advantage of LPCC is that it can simulate the human vocal tract model well and the excitation information in the process of speech generation can be completely eliminated. At the same time, the calculation of LPCC coefficients is simple and easy to implement. More than a dozen LPCC coefficients can well describe the formant characteristics of speech signals, therefore, a good recognition effect can be obtained in speech emotion recognition. By taking the Fourier transform of the speech signal, the logarithm of the module is taken and then the inverse Fourier transform and the cepstrum of the speech signal are determined.

H ^ ( z ) = ln H ( z ) n = 1 h ( n ) z - n (10)

On substituting Equation (6) into (10), and at the same time taking the derivative of 1z on both sides of the equation, we can get:

( 1 - i = 1 p α i z - i ) = n = 1 n h ( n ) z - n + 1 - i = 1 p i α z - i (11)

If the coefficients of z–1 in the above formula are equal, respectively, the cepstral coefficient can be obtained from the LPC coefficient. The recursive formula is as follows:

h ^ ( m ) = { α n , m = 1 α n + k = 1 n - 1 k h ^ ( k ) α n - k / m , 1 < m p + 1 k = 1 m - 1 k h ^ ( k ) α n - k / m , m > p + 1 (12)

Hidden Markov Model/Radial Basis Function Hybrid Model Based on Speech Emotion Recognition

Mel Frequency Cepstral Coefficient

Mel frequency Cepstrum coefficient (MFCC) extraction is based on the mechanism of human hearing, that is to say, the frame spectrum of speech is analyzed according to the characteristics of human hearing, and in the presence of spectral distortion and channel noise have a higher recognition accuracy. The process of MFCC coefficient extraction and calculation is shown in Figure 7.

FIGURE 7
www.frontiersin.org

Figure 7. Mel frequency cepstral coefficient extraction process.

After the original speech signal is preprocessed, each frame time domain signal is obtained.

x ( k ) = n = 0 N - 1 x ( n ) e - j 2 n k N ( 0 < n , k < N - 1 ) (13)

where x(k) is the linear spectrum and N is the DFT window width.

To find the logarithmic energy of each filter output, square the spectrum, the energy spectrum, is obtained and then multiplied by a triangular bandpass filter bank. For better robustness, especially to noise and spectral estimation errors, the output of each filter can be taken logarithmically to obtain the logarithmic power spectrum of the corresponding frequency band. Then the inverse discrete cosine transform is performed, thus, L MFCC coefficients are obtained. In general, L takes 12 to 16, and the MFCC coefficients are:

C n = log x ( k ) cos ( k - 0.5 ) n / M , n = 1 , 2 L (14)

Hidden Markov Model/Radial Basis Function Hybrid Model

The composition diagram of the Hidden Markov Model (HMM), called by the same name in Chinese, is shown in Figure 8.

FIGURE 8
www.frontiersin.org

Figure 8. Schematic diagram of the composition of the HMM model.

For an effective ensemble RBF neural network for the HMM model, applied to speech emotion recognition, a targeted algorithm must be selected, for which the authors use the Baum-Welch algorithm:

The expression is as follows:

ξ t ( i , j ) = P ( q t = i , q t + 1 = j , O | λ ) (15)

It can be deduced from this:

ξ t ( i , j ) = [ α t ( i ) α i j b j ( o t + 1 ) β t ( j ) ] / P ( O | λ ) (16)

The probability of state Si in the Markov chain is:

γ t ( i ) = P ( q t = i , O | λ ) = j = 1 N ξ t ( i , j ) = α t ( i ) β t ( i ) / P ( O | λ ) (17)

This iterative process is repeated to constantly get more optimized parameters, until the parameters are almost unchanged. The model at this time is the final model that meets the actual requirements.

Analysis of Results

Experiment Preparation

In this experiment, under the Windows operating system environment, the programming simulation is carried out with MATLAB as the working platform. The experimental data uses the CASIA Chinese corpus introduced in Chapter 2, choosing from 5 emotions (happy, sad, angry, surprised, and neutral) with a total of 200 sentences. The experiment adopts the method of ten-fold cross-validation, and the speech sample data is divided into 10 parts, 9 for training and 1 for testing, rotating verification. Then the average of the 10 recognition results is taken as the recognition rate. Sometimes ten-fold cross-validation is performed multiple times; the training process is shown in Figure 9.

FIGURE 9
www.frontiersin.org

Figure 9. Mixed model neural network training process.

In simulation experiments, the sampling frequency of the voice signal is set to 8 KHz and the quantization precision is 16 bit. Using a Hamming window, the frame length is 256 and the frame shift is 128. The simulation results of the preprocessing and feature parameter extraction of the speech signal have been shown in detail in the previous section. The authors select the first and second derivatives of the fundamental frequency 0F and 0F, the first three formants F1, F2, F3, the 10th order LPCC coefficient, and the 10th order MFCC coefficient, a total of 29 parameters for speech emotion recognition.

Due to the different physical meanings and value ranges of the extracted feature parameters, the quantitative relationship between the original data is quite different, it is unreasonable to use these feature parameters directly for identification, and the solution is to normalize these parameters. The code is shown in Table 3 below.

TABLE 3
www.frontiersin.org

Table 3. HMM-RBF hybrid model algorithm.

Comparison of Recognition Rates

In this context, the HMM/RBF model is compared to the HMM model and the RBF model. As shown in Figure 10 below, from 10 to 81.10% of the HMM model, the average RBF model is 82.84% and the HMM/RBF model is approved 90.98 percent higher. Therefore, the experiments show that the HMM/RBF hybrid model proposed by the authors have better recognition effect.

FIGURE 10
www.frontiersin.org

Figure 10. Comparison of recognition rates of three algorithms.

In addition, the authors use the CASIA Chinese sentiment corpus, where the corpus was recorded in a quiet experimental environment to obtain a speech signal without the influence of noise; however, in the process of actually collecting the voice signal, there will inevitably be interference such as background noise. Therefore, white noise with different signal-to-noise ratios were added to conduct the experiment again and verify the anti-noise performance of the established emotion model. The average recognition rate of three emotion models are shown in Figure 11 below, under different SNR environments. It can be clearly seen that, in the case of the HMM/RBF hybrid model with a signal-to-noise ratio of 30 db, the recognition rate can still reach about 89%, a better anti-noise performance than separate HMM and RBF models.

FIGURE 11
www.frontiersin.org

Figure 11. Average recognition rate of three algorithms under different signal-to-noise ratios.

Comparison of Convergence Speed

Aiming at the learning rate problem of the RBF network in the learning stage, the traditional fixed learning rate is abandoned, the speed of the learning is high, and the speed of the learning is re-calculated every time, and it improves the speed of the RBF network and reduces the time to work. From Table 4, it can be seen that the acceptance rate of the three events has not changed, this is because the initial parameters of the RBF model are the same, and the RBF network after reaching the learning goal is the same, then the recognition rate of the HMM/RBF hybrid model will not change. However, it can be seen that the HMM/RBF hybrid model with the optimal learning rate is better than the HMM/RBF hybrid model with a fixed learning rate, the running time and the number of iterations have been significantly improved, which not only ensures the stability of the network but also improves the operational efficiency.

TABLE 4
www.frontiersin.org

Table 4. Convergence speed comparison.

A dynamic optimal learning rate is introduced to the RBF network, although this method makes the RBF network in each iteration step recalculate the value of the learning rate and consumes a certain amount of computing time; however, the RBF network is simple and convenient and does not increase the running time of the HMM/RBF hybrid model. In addition, the number of iterations and running time of the improved RBF network are significantly reduced, therefore, the HMM/RBF hybrid model with dynamic optimal learning rate can greatly improve the operation efficiency.

Conclusion

The authors first introduce the basic theory and knowledge of HMM model and RBF model, the definition of HMM model, and the three basic algorithms are further elaborated. Then they outline the basic knowledge points of the two models, the RBF definition and the learning process. In addition, the concept of “dynamic optimal learning rate” is introduced. The HMM/RBF hybrid model with dynamic optimal learning rate makes the network more stable. Then, the theoretical basis and combination method of the HMM/ANN hybrid model are introduced, and using the strong dynamic time series modeling capabilities of HMM, respectively, and the strong classification decision-making ability of RBF, the two algorithms are combined to improve the recognition rate. Finally, the implementation of the HMM/RBF hybrid model is analyzed in detail, including the training and recognition process. Experiments show that the HMM/RBF hybrid model has better recognition effect than the separate HMM and RBF models and the HMM/RBF hybrid model with the optimal learning rate has a significantly faster convergence rate.

After extensive literature research, it was found that the authors still have many shortcomings that need to be further improved, there are two main aspects:

(1) There are many kinds of speech emotion feature parameters, the authors only select five kinds of feature parameters, including short-term energy, fundamental frequency, formant, LPCC, and MFCC. Further study of other emotional feature parameters is needed, such as selection of the parameters that can best express emotional characteristics, this is one of the issues that need improvement.

(2) The authors use the marked sentiment CASIA Chinese corpus where 20 sentences of text are read aloud by 4 professional actors simulating different emotions; however, this is different from the natural emotional expression of people in daily life. How to collect natural emotional speech signals and build a natural emotion corpus is a problem that needs to be solved.

Data Availability Statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author Contributions

Both authors listed have made a substantial, direct, and intellectual contribution to the work, and approved it for publication.

Funding

This work was supported by research project of the Humanities and Social Sciences Youth Fund of the Ministry of Education “Research on China’s Elderly Care under the Background of Aging – Intelligent Apartment Design based on the User Experience of the Elderly in the Era of Internet of Things” (Project No. 18YJC760097).

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Abbas, A. F., and Abdullah, M. Z. (2021). Design and implementation of a smart home gas detection based on mobile network system. J. Eng. Sustain. Dev. 25, 9–16. doi: 10.31272/jeasd.25.2.2

CrossRef Full Text | Google Scholar

Agarwal, G., and Om, H. (2021). Performance of deer hunting optimization based deep learning algorithm for speech emotion recognition. Multimed. Tools Appl. 80, 9961–9992. doi: 10.1007/s11042-020-10118-x

CrossRef Full Text | Google Scholar

Chatterjee, R., Mazumdar, S., Sherratt, R. S., Halder, R., and Giri, D. (2021). Real time speech emotion analysis for smart home assistants. IEEE Trans. Consum. Electron. 67, 68–76. doi: 10.1109/tce.2021.3056421

CrossRef Full Text | Google Scholar

Chen, X. (2021). Simulation of English speech emotion recognition based on transfer learning and cnn neural network. J. Intell. Fuzzy Syst. 40, 2349–2360. doi: 10.3233/jifs-189231

CrossRef Full Text | Google Scholar

Deschamps-Berger, T., Lamel, L., and Devillers, L. (2021). End-to-end speech emotion recognition: challenges of real-life emergency call centers data recordings. arXiv [Preprint]. doi: 10.48550/arXiv.2110.14957

CrossRef Full Text | Google Scholar

El-Henawy, I., and Abo-Elazm, M. (2020). Handling within-word and cross-word pronunciation variation for Arabic speech recognition (knowledge-based approach). J. Intell. Syst. Internet Things 1, 72–79. doi: 10.54216/jisiot.010202

CrossRef Full Text | Google Scholar

Lech, M., Stolar, M., Best, C., and Bolia, R. (2020). Real-Time speech emotion recognition using a pre-trained image classification network: effects of bandwidth reduction and companding. Front. Comput. Sci. 2:14. doi: 10.3389/fcomp.2020.00014

CrossRef Full Text | Google Scholar

Li, D. (2021). Design of english text-to-speech conversion algorithm based on machine learning. J. Intell. Fuzzy Syst. 40, 2433–2444. doi: 10.3233/jifs-189238

CrossRef Full Text | Google Scholar

Liu, M., and Zhang, C. (2021). Design of formaldehyde concentration detection system for smart home based on stm32 controller. J. Phys. Conf. Ser. 1780:012021. doi: 10.1088/1742-6596/1780/1/012021

CrossRef Full Text | Google Scholar

Liu, S., Zhang, M., Fang, M., Zhao, J., and Hung, C. C. (2021). Speech emotion recognition based on transfer learning from the facenet framework. J. Acoust. Soc. Am. 149, 1338–1345. doi: 10.1121/10.0003530

PubMed Abstract | CrossRef Full Text | Google Scholar

Mustaqeem Kwon, S. (2021). Speech Emotion Recognition Based on Deep Networks: A Review. Seoul: Korean Information Processing Society (KIPS-2021).

Google Scholar

Peng, J., Ye, H., He, Q., Qin, Y., and Lu, J. (2021). Design of smart home service robot based on ros. Mob. Inf. Syst. 2021, 1–14. doi: 10.1155/2021/5511546

CrossRef Full Text | Google Scholar

Puri, T., Soni, M., Dhiman, G., Ibrahim Khalaf, O., Alazzam, M., and Raza Khan, I. (2022). Detection of emotion of speech for RAVDESS audio using hybrid convolution neural network. J. Healthc. Eng. 2022:8472947. doi: 10.1155/2022/8472947

PubMed Abstract | CrossRef Full Text | Google Scholar

Qi, Q., Xie, X., Peng, C., and Hu, J. (2021). “Design of wireless smart home safety system based on visual identity,” in Proceedings of the 2021 International Conference on Communications, Information System and Computer Engineering (CISCE), Beijing.

Google Scholar

Ren, F., Dong, Y., and Wang, W. (2019). Emotion recognition based on physiological signals using brain asymmetry index and echo state network. Neural Comput. Applic. 31, 4491–4501. doi: 10.1007/s00521-018-3664-1

CrossRef Full Text | Google Scholar

Shemeis, M., Asad, T., and Attia, S. (2021). The effect of big five factors of personality on compulsive buying: the mediating role of consumer negative emotions. Am. J. Bus. Oper. Res. 2, 5–23. doi: 10.54216/ajbor.020102

CrossRef Full Text | Google Scholar

Sisavath, C., and Yu, L. (2021). Design and implementation of security system for smart home based on iot technology. Procedia Comput. Sci. 183, 4–13. doi: 10.3390/s19183973

PubMed Abstract | CrossRef Full Text | Google Scholar

Song, Y., Wang, X., Sun, F., and Wang, H. (2021). Camouflaged Object Segmentation With Transformer. Beijing: International Conference on Cognitive Systems and Signal Processing.

Google Scholar

Tao, M. Y., Ren, Y. H., Zhao, J. C., and Lu, Y. (2021). “Design of elevator auxiliary control system based on speech recognition,” in Proceedings of the 2021 IEEE 6th International Conference on Computer and Communication Systems (ICCCS), Chengdu.

Google Scholar

Tarnowski, P., Kołodziej, M., Majkowski, A., and Jan Rak, R. (2020). Eye-Tracking analysis for emotion recognition. Comput. Intell. Neurosci. 2020:2909267.

Google Scholar

Wang, S., and Xu, Z. (2021). Emotional attention detection and correlation exploration for image emotion distribution learning. IEEE Trans. Affect. Comput. 1:99. doi: 10.1109/TAFFC.2021.3071131

CrossRef Full Text | Google Scholar

Wang, Y., Wu, Y., Fu, L., Wang, R., and Feng, S. (2021). Distribution characteristics based on data analysis of edible vegetable oil supermarkets in sy city. J. Phys. Confer. Ser. 1852:042051. doi: 10.1088/1742-6596/1852/4/042051

CrossRef Full Text | Google Scholar

Wang, Z., Pan, H., Xie, L., Liu, B., Yang, M., and Tao, J. (2021). Review of micro-expression spotting and recognition in video sequences – sciencedirect. Virt. Real. Intell. Hard. 3, 1–17. doi: 10.1016/j.vrih.2020.10.003

CrossRef Full Text | Google Scholar

Xu, J., and Wan, Y. (2021). Design of smart home decoration energy saving system based on context awareness. J. Ambient Intell. Humaniz. Comput. 1–11.

Google Scholar

Yang, L., Xie, K., Wen, C., and He, J. B. (2021). Speech emotion analysis of netizens based on bidirectional lstm and pgcdbn. IEEE Access 9, 59860–59872. doi: 10.1109/access.2021.3073234

CrossRef Full Text | Google Scholar

Zhu, Z., Dai, W., Hu, Y., Wang, J., and Li, J. (2021). “Speech emotion recognition model based on CRNN-CTC,” in 2020 International Conference on Applications and Techniques in Cyber Intelligence. ATCI 2020. Advances in Intelligent Systems and Computing, eds J. Abawajy, K. K. Choo, Z. Xu, and M. Atiquzzaman (Cham: Springer). doi: 10.1007/978-3-030-53980-1_113

CrossRef Full Text | Google Scholar

Keywords: artificial intelligence, speech emotion recognition, HMM, RBF, dynamic optimal learning rate, aging users

Citation: Wu X and Zhang Q (2022) Design of Aging Smart Home Products Based on Radial Basis Function Speech Emotion Recognition. Front. Psychol. 13:882709. doi: 10.3389/fpsyg.2022.882709

Received: 24 February 2022; Accepted: 14 March 2022;
Published: 04 May 2022.

Edited by:

Deepak Kumar Jain, Chongqing University of Posts and Telecommunications, China

Reviewed by:

Ping Wang, Jiangxi University of Technology, China
Zhen Bai, Henan University, China
Xuetao Li, Hubei University of Automotive Technology, China

Copyright © 2022 Wu and Zhang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Xu Wu, wuxu_19871214@email.tjut.edu.cn

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.