Human-centric Computing and Information Sciences volume 13, Article number: 12 (2023)
Cite this article 1 Accesses
https://doi.org/10.22967/HCIS.2023.13.012
Amid the rapid development of science and technology and the increasingly fierce market competition, low cost coupled with high performance is the key to a company's competitiveness. This requires optimizing all the links in the actual output. Hybrid intelligent algorithms can combine the advantages of different algorithms to solve a large number of optimization problems in engineering practice. Therefore, this study is focused on a hybrid swarm intelligence algorithm and its application. As a classic method of machine learning, the kernel function, which is based on the support vector machine (SVM) and the selection of the parameters in the kernel function, has an important influence on the performance of the classifier. The use of kernel function technology cannot only greatly reduce the amount of calculation in the input space, but can also effectively improve machine learning classification performance. In the field of machine learning, choosing and building the core functions is a notable difficulty. However, little research has been conducted in this area so far. In view of the above problems, this study discusses and analyzes the structure of the support frame machine core in detail, and improves the traditional parameter optimization algorithm. It also proposes a new method of fuzzy clustering algorithm automatic parameter learning combined with the basic ideas of a genetic algorithm in order to improve the parameter optimization strategy of support vector regression, so as to obtain better prediction results. Through simulation experiments, the improved hybrid core SVM and parameter optimization algorithm were applied to the ORL face database, greatly improving the recognition rate, and experiments were carried out after adding noise to the images in the face database to verify the practicability and practicality of the algorithm. The robustness and reliability of the algorithm were improved by at least 30%, thus confirming the feasibility of the proposed algorithm.
Hybrid Intelligence, Optimization Algorithm, Feature Selection, Parameter Synchronization Optimization
With the explosive development of the internet and information technology, network information has brought us many conveniences, but negative effects in such as areas as information security have also gradually emerged. Some criminals have used information networks to illegally steal state secrets and commercial materials. Due to these threats to social stability, the importance of information security has become even more prominent. Traditional methods of identity verification mainly include ID cards, driving licenses, passwords, passwords, and so forth. Although these methods can encrypt information to a certain extent, magnetic cards, IC cards, and the like are easy to lose, and passwords are easy to forget, among many other problems, so they are gradually being abandoned by the great majority of people.
The purpose of optimization is to optimize target measurement. Efficient optimization technology is the key to saving energy, reducing costs, and improving reliability. The basic principle of traditional optimization methods is to use the idea of calculus to determine the extreme value of the objective function. However, with the increasing number of industries covered by optimization techniques, the emerging problems are becoming increasingly complicated, and traditional optimization methods seem unable to solve these problems, and are accordingly judged to be incompetent. At the present time, compared with traditional methods, some intelligent algorithms have begun to enter people's field of vision due to their own unique advantages. Intelligent algorithms are produced by imitating biological behaviors, and their principles are simple and easy to implement. Search it is highly efficient and can solve a wide range of practical engineering problems.
A study by Wen Haibiao et al. reported that when the supported carrier machine (SVM) processes large sample datasets with a large number of feature dimensions, the algorithm is very time-consuming and difficult to use in local optimal solutions. Choosing inappropriate SVM algorithm parameters will affect the classification performance of the SVM model. In order to improve the performance of the SVM, he proposed an algorithm for optimizing the synchronization of the SVM parameters, which combines particle swarm optimization (PSO), genetic algorithm (GA) perform feature selection, and parameter synchronization. Experiments using standard UCI datasets show that the polygenic scores (PGS) algorithm can effectively find the appropriate function set and parameters of the SVM algorithm, improve the convergence speed, and achieve higher classification accuracy in a smaller feature subset. However, the range of experimental error is relatively large, making it necessary to perform multiple experiments to obtain accurate results. Wang Lianhong believes that in order to eliminate unnecessary data in feature extraction and to optimize the classification performance of the SVM (support machine), some researchers have proposed to use particle algorithms to select the features and optimize the classification parameters. However, the particle algorithm is too fast and simple to calculate, so the classifier parameters often cannot be synchronized. It is applied to a discrete space, and the attribute selection and classifier parameters are not synchronized or optimized. To resolve the aforementioned problems, this study proposes a new SVM algorithm to select the features in a discrete space and to SVMs. Regarding modern parameter optimization, experimental results show that the SVM algorithm can effectively filter the feature sets, optimize the SVM parameters, reduce space complexity, and improve the face detection rate and robustness. But as the algorithm has not been practically applied, it needs to be promoted and optimized. A study by Fan reports that vector-assisted theory has been widely used in pattern classification in recent years, but the two main factors affecting classification accuracy, namely, feature selection and parameter optimization, interact with and restrict each other. He proposed a BA+SVM algorithm that uses the Bat algorithm (BA) to optimize the SVM parameters and select input data at the same time, thereby improving the classification ability of the SVM, and developed three experimental methods in 10 test datasets. The results show that, compared with the algorithm used for parameter optimization or single feature selection, the modern BA+SVM optimization algorithm has the advantages of having fewer input functions and greater accuracy. However, the algorithm runs slowly and has low practicability, and thus needs to be strengthened.
Adding the gravity of the intermediate fitness particle to the current particle to the speed update formula of the standard PSO algorithm to avoid the blind search of the particle only under the guidance of the global optimal particle and the individual optimal particle. The introduction of the intermediate fitness particle can provide the particle multiple experiences. In summary, recommendation systems are used in various fields of the Internet, which can solve the problem of "information overload" and help users to locate the range of resources they need very quickly. Therefore, studying the problem of recommender systems has become an important topic in Internet applications. The parameter synchronization optimization problem of the SVM has become the focus of research by countless experts and scholars [1]. A good optimization algorithm can produce huge application value. Collaborative filtering algorithm and particle swarm algorithm can play a role in solving the recommendation algorithm problem. When analyzing the collaborative filtering algorithm and the particle swarm algorithm, this paper mainly considers the influence of the time factor on the correctness of the collaborative filtering algorithm and the particle swarm algorithm, so as to improve the accuracy of optimization [2].
At present, China's Doppler weather radar is usually used for quantitative precipitation estimation (QPE) based on the Z-R relationship. However, the estimation error of mixed precipitation is very large. In order to improve the accuracy of radar QPE, Changjiang proposed a 6-minute dynamic radar QPE algorithm based on the reflectivity data of the Doppler radar Z9002 in the Qingpu district of Shanghai and the precipitation data of automatic weather stations (AWSs) in eastern China.
Considering the time dependence and mutation of precipitation, the first 30 minutes of data were selected as training data. In order to reduce the complexity of radar QPE, the stationary wavelet transform (SWT) is used to convert the weather data into the wavelet domain, and the high and low frequency reflectivity and precipitation information are extracted. Using wavelet coefficients, a SVM was constructed on all scales to estimate the wavelet coefficients of precipitation. Finally, the estimated rainfall was obtained by inverse wavelet transform (IWT). However, the variables controlled in the experiment were not very clear [3, 4]. Cloud computing is widely accepted by individuals and enterprises in storing multimedia content. This is due to the introduction of a new architecture, which has lower service costs for computing, storage, and maintaining multimedia storage. However, cloud users must take some measures to avoid privacy problems. In order to provide multimedia security, Sukumar et al. [5] transformed the multimedia content using discrete Rajan transform (DRT) and embedded it into the selected cover image, which was generated by integer wavelet using the diamond coding scheme. The generated hidden images were stored in the cloud. When multimedia content was needed, the hidden image was downloaded from the cloud and inversely transformed by IWT. SVM provides good learning ability for the extraction process, which makes the algorithm more robust to various attacks (i.e., salt and pepper noise, Gaussian noise, clipping, compression, etc.). The experimental value of the peak signal-to-noise ratio (PSNR) was 53 and 50 for two secret images, which is better than the existing scheme. Similarly, the scheme provided better results for robustness and security evaluation. But it is not very suitable for practical applications [5]. The SVM is a powerful technology in terms of pattern classification, but its performance depends on its parameters to a great extent. Li et al. [6] proposed a new SVM, which was optimized by a new differential evolution (DE), adopted the mixed parameter setting strategy and population size adaptive method, and was simplified to FDE-PS-SVM. In the mixed parameter setting strategy, the parameter offspring of the SVM was generated by the operator with fixed evolutionary parameters, or by fuzzy logic reasoning (FLR) according to a given probability. In the population size adaptive method, the population size was gradually reduced in the search process, by trying to balance the diversity and concentration ability of the algorithm and find better SVM parameters. Some benchmark datasets were used to evaluate the proposed algorithm. The experimental results show that these two strategies can effectively search for better SVM parameters, and the performance of the FDE-PS-SVM algorithm is better than other algorithms proposed in other literatures. However, the simplification of the experimental methods needs to be improved [6].
Air pollution prediction plays an important role in helping to reduce air pollutant emissions, guiding people's daily activities and warning the public in advance. However, previous studies still have many shortcomings, such as ignoring the importance of outlier detection and correction of the original time series, and the randomness of the initial parameters of the model. To solve such problems, Wang et al. [7] proposed a hybrid model based on an outlier detection and correction algorithm and a heuristic intelligent optimization algorithm. First, the data preprocessing algorithm was used to detect and correct outliers and mine the main features of the original time series; second, a widely used heuristic intelligent optimization algorithm was used to optimize the parameters of the limit learning machine, and the prediction results of each subclass were obtained, improving prediction accuracy; and, finally, the experimental results and analysis showed that the proposed hybrid model provides accurate prediction superior to that of other comparative models, but it is not practical [7]. Over the past few decades, through research on natural organisms, a large number of intelligent algorithms based on social intelligent behavior have been widely studied and applied to various optimization fields. The learning-based intelligent optimization algorithm (LIOA) is an intelligent optimization algorithm with certain learning abilities. This is how the traditional intelligent optimization algorithm combines the learning operator or specific learning mechanism to give itself a certain learning ability, so as to achieve better optimization behavior. Li et al. [8] conducted a comprehensive survey of the LIOA. The research contents included a statistical analysis of the LIOA, classification of LIOA learning methods, application of the LIOA in complex optimization scenarios, and application of the LIOA in engineering applications. The future views and development direction of the LIOA were also discussed, but it was not applied in detail [8]. In recent years, with the widespread application of the SVM in machine learning applications, it has become very important to obtain a sparse model that is sufficiently robust to withstand the noise in the dataset. Singla et al. [9] sought to enhance the sparsity of RSVM-RHHQ (robust support vector machine-rescaled hinge loss function) using a non-smooth regularizer with non-convex and non-smooth loss functions, and used the primal dual approximation method to solve the non-smooth non-convex problem. It was soon found that this combination not only increased the sparsity of the model, but also outperformed the existing robust SVM methods in terms of robustness to label noise. In addition, the time complexity of the optimization technology was also considered. The experimental results showed that this method was superior to existing methods in terms of sparsity, accuracy, and robustness. In addition, a sensitivity analysis of the label noise regularization parameters in the dataset was conducted. However, the experimental operation was too cumbersome and too many aspects were considered, resulting in too many constraints [9]. Based on artificial intelligence technology, modern irrigation systems need to be improved all the time. In his research work, Ali et al. [10] proposed a new population-based meta-heuristic algorithm called the “control shower optimization” (CSO) algorithm for the global optimization of unconstrained problems. Modern irrigation systems are equipped with intelligent tools made and controlled by human intelligence. The proposed CSO algorithm was inspired by the function of the water allocation tool, which is used to model the search agent that performs the optimization process. CSO simulates the mechanism of the sprinkler projecting the water unit and its platform moving to the required position to plan the best search program. The proposed method has been tested with many low-dimensional and high-dimensional benchmark functions with different properties. Statistical analysis of the empirical data showed that CSO provides a solution of higher quality than several other effective algorithms, including the GA, the PSO algorithm, the DE algorithm, the artificial bee colony (ABC) algorithm, and the covariance matrix adaptive evolution strategy (CMA-ES), However, in the specific application, it is still necessary to continue investigating according to the actual situation [10]. In order to improve the marketing effects of e-commerce products, Cui et al. [11] constructed an e-commerce product marketing model based on machine learning and the SVM based on a machine learning algorithm. In addition, he also studied the classical reinforcement learning algorithm known as “Q-learning” and proposed an improved Q-learning algorithm. In addition, the mean normalization method was used to reduce the noise impact of the reward signal caused by the non-fixed time interval between decision points. Aiming at the deviation caused by the asynchronous update of time intervals in the iterative process of the Q-value function, the standardization factor was further constructed. However, the experimental data of this algorithm were not very convincing [11]. Research on intelligent algorithms has always been a hot topic in the field of human-centered computing, and continues to expand with the development of artificial intelligence. Usually, coupled data fusion algorithms usually use the information of one dataset to improve estimation accuracy and explain the relevant latent variables of other coupled datasets. Lu et al. [12] proposed several coupled image decomposition algorithms based on a coupling matrix with a tensor decomposition optimization (CMTF-OPT) algorithm and a flexible coupling algorithm, which are called the “coupled image decomposition optimization” (CIF-OPT) algorithm and the “improved flexible coupling” algorithm, respectively. Theory and experiments have shown that the effect of the CIF-OPT algorithm is robust under the influence of different noise. In particular, the CIF-OPT algorithm can accurately recover images that have lost some data elements. However, the experiment is not very representative [12]. With the continuous progress of computer and information technology, a large number of research papers are now being published online and offline all the time. With the continuous emergence of new research fields, it is difficult for users to find interesting research papers and classify them. In order to overcome these limitations, Kim and Gil [13] proposed a research paper classification system capable of clustering research papers into meaningful categories, in which papers are likely to have similar topics. The proposed system extracts representative keywords from the topics of each paper and topic through the potential Dirichlet assignment (LDA) scheme. Then, based on the word frequency inverse document frequency (TF-IDF) value of each paper, all papers are classified into research papers with similar topics by the K-means clustering algorithm. However, this experimental method did not prove to be sufficiently innovative [13].
Basic Principles of the Particle Swarm Algorithm
At present, the most commonly used task scheduling algorithm in the cloud computing workflow management system is the particle swarm algorithm proposed by Eberhart and Shi. The origin of the particle swarm algorithm has been described in detail in the research background provided in the previous chapter. As each and every particle in the particle swarm algorithm is like a bird, the optimization process of the algorithm is like a flock of birds looking for food. The particle gradually approaches the optimal solution by adjusting its own speed and direction, just as birds flying together in a flock adjust their speed and direction of flight by cooperating with each other and sharing positional information, and then find food.
The particle swarm algorithm is the most widely used intelligent optimization algorithm because it has several characteristics that are very convenient for researchers to learn and improve, i.e., the algorithm has fewer parameters, is easy to program, and, most importantly, has a faster convergence speed. However, the particle swarm algorithm also has an obvious shortcoming in that it is very easy to fall into a local optimal solution, which makes the algorithm unable to find an accurate optimal solution after iteration [14].
Features of the Particle Swarm Algorithm
The good performance of the particle swarm algorithm mainly stems from some of its advantages, which are mainly as follows:
(1) The principles and implementation process of the particle swarm algorithm are simple and easy to understand, and relatively easy to expand. Most researchers choose the particle swarm algorithm and other intelligent optimization algorithms for fusion, and its effects are remarkable [15].
(2) The particle swarm algorithm entails the adjustment of fewer parameters and is easier to understand than other intelligent algorithms. Researchers can make appropriate improvements and optimizations to the particle swarm algorithm according to the specific optimization problems that need to be solved. Generally, it is more extensive to choose to improve the inertia weight, and the optimization effect is more obvious. The principle is shown in Fig. 1.
Linear Support Vector Machine
The kernel function is the most important part of the support vector machine. The different choices of the kernel function lead to different types of SVMs. The introduction of the kernel function can transform complex nonlinear problems into easy-to-solve linear problems [18]. It has a similar function to that of the converter. This algorithm can map the original data space to a high-dimensional data space, so that the data have better linear separability in the feature space. The concept of the linear SVM is to search for an optimal classification hyperplane, which cannot only correctly separate the two types of samples, but also satisfy the interval between the classification samples to the largest extent possible and the error to the smallest extent possible. In the low-dimensional space, L is the classification line [19, 20]. If this idea is extended to the high-dimensional space, the optimal classification line becomes the corresponding optimal classification surface.
Assuming that the given sample training set T={(x1,y1),(x2,y2),...,(xn,yn)} is the category label, the expression of the classification line L in the m-space is as follows:
(1)
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
Stock Dataset Experiment and Analysis
This section gives a brief introduction to the singular spectrum transformation before discussing the experimental data. Singular spectrum analysis is mainly used to study nonlinear time series data. This method consists in constructing a trajectory matrix that can be decomposed and reconstructed from the time series it observes. With this operation, some signals representing different components of the original time series are obtained, so as to structure the time series and analyze and perform further forecasting operations
[27, 28]. In this dataset, the main purpose of preprocessing using singular spectrum analysis is to reduce noise and achieve better prediction purposes.
Table 1. Grid algorithm of the stock dataset
First | Second | Third | Average value | |
c | 2 | 2 | 2 | 2 |
σ (represented as g in MATLAB) | 1 | 1 | 1 | 1 |
Time (s) | 1755.9 | 1975.1 | 1851.2 | 1860.7 |
Mean training model (MSE) | 0.00881757 |
Model type | Type of data | MAPE (%) | RMSE | R |
PCA-LSSVM NH4 | Training data | 47.62 | 12.38 | 0.9926 |
N removal model | Forecast data | 9.93 | 15.07 | - |
PCA-LSSVM TN | Training data | 3.47 | 12.48 | 0.9849 |
Remove model | Forecast data | 3.41 | 13.66 | - |
Serial number | Influent pH value | NH4-N (mg/L) | NO2 (mg/L) | COD (mg/L) |
A1 | 7.51 | 184.61 | 301.11 | 74.58 |
A2 | 7.52 | 186.49 | 305.77 | 96.37 |
B1 | 7.56 | 188.94 | 313.15 | 132.3 |
B2 | 7.54 | 188.19 | 319.88 | 157.67 |
Parameter name | Symbol | Value |
Population size | s | 100 |
Number of resources (number of virtual machines) | r | 5 |
Number of tasks | tn | 50–300 |
Maximum inertia weight | ωmax | 1 |
Minimum inertia weight | ωmin | 0 |
When attempting to research a single-species intelligent algorithm, there are bound to be optimization and application defects due to its own shortcomings, so the fusion of two or more single-species intelligent algorithms to form a mixed-group intelligent algorithm can effectively maximize their respective strengths and avoid their weaknesses, and further enhance the specialties of the single-species intelligent algorithm. The PSO has good global search ability, and has advantages in terms of high- and low-dimensional optimization problems, but it can easily fall into the local optimum, while the local refined search ability of the collaborative algorithm is more prominent. Therefore, this paper attempted to improve the particle swarm algorithm and perform a parallel game search. The better result of the search stall and delay was selected as the initial threshold of the improved collaborative algorithm, and then the improved collaborative algorithm was used for a local refined search to form a hybrid swarm intelligence algorithm of two algorithms. The core strategy of this paper consisted in conducting detailed and comprehensive experiments comprising five basic time series data categories, which basically represent all types of time series data.
Through a large number of experiments and illustrations, it has been confirmed that the core strategy of this paper can indeed obtain the optimal parameter set within a reasonable time cost, and overcome the limitations of other classic algorithms where the time cost is too large or which cannot be universally applied to every dataset type. The shortcomings of better solution results are obtained, and then better prediction results are obtained. At the same time, time efficiency can also be controlled within a more appropriate range. Although the fusion algorithm, ITAIWPSACO, proposed in this paper has obtained relatively good experimental results, there are still shortcomings regarding task scheduling in the cloud computing environment, so it needs to be further improved and perfected. In the ITAIWPSACO algorithm proposed in this paper, only the optimization condition of the task completion time is considered. Although the task scheduling cost is added to the ITAIWPSO algorithm at the initial stage of scheduling, the final optimization result is only the task completion time. Therefore, in the real cloud, many factors should be considered in the computing environment, and future studies will be required to improve the algorithm.
Conceptualization, ZM, QW. Funding acquisition ZM. Investigation and methodology, ZM. Project administration ZM, QW. Resources ZM, QW. Supervision QW. Writing of the original draft, QW. Writing of the review and editing, QW. Software, QW. Validation, ZM. Formal analysis, ZM. Data curation, ZM, QW. Visualization, ZM, QW.
This work was supported by the Jiangxi Education and Teaching Reform Project (No. JXJG-17-24-12), Science and Technology Project of the Jiangxi Provincial Department of Education (No. GJJ191000, GJJ180979).
The authors declare that they have no competing interests.
Name: Qingjun Wang
Affiliation: College of Economics and Management, Shenyang Aerospace University, Shenyang, 110136, Liaoning, China. Nanjing University of Aeronautics and Astronautics, Nanjing 210016, China.
Biography: Qingjun Wang, received his M.S. degree from the Northeast University, in 2009. He is currently a graduate student studying for Ph.D. degree in the College of Automation Engineering, Nanjing University of Aeronautics and Astronautics. His research interests include Pattern Recognition, Artificial Intelligent.
Name: Zhendong Mu
Affiliation: The Center of Collaboration and Innovation, Jiangxi University of Technology, Nanchang 330098, Jiangxi, China.
Biography: Zhendong Mu (1975.11), male, graduated from Nanchang University with a master's degree. Working in Jiangxi University of Technology, professor. Mainly engaged in intelligent computing and brain-computer interface research.
Qingjun WangM1,2 and Zhendong Mu3,*, Feature Selection and SVM Parameter Synchronous Optimization Based on a Hybrid Intelligent Optimization Algorithm, Article number: 13:12 (2023) Cite this article 1 Accesses
Download citationAnyone you share the following link with will be able to read this content:
Provided by the Springer Nature SharedIt content-sharing initiative