Acoustic emission(AE)source localization is a fundamental element of rock fracture damage imaging.To improve the efficiency and accuracy of AE source localization,this paper proposes a joint method comprising a three-...Acoustic emission(AE)source localization is a fundamental element of rock fracture damage imaging.To improve the efficiency and accuracy of AE source localization,this paper proposes a joint method comprising a three-dimensional(3D)AE source localization simplex method and grid search scanning.Using the concept of the geometry of simplexes,tetrahedral iterations were first conducted to narrow down the suspected source region.This is followed by a process of meshing the region and node searching to scan for optimal solutions,until the source location is determined.The resulting algorithm was tested using the artificial excitation source localization and uniaxial compression tests,after which the localization results were compared with the simplex and exhaustive methods.The results revealed that the localization obtained using the proposed method is more stable and can be effectively avoided compared with the simplex localization method.Furthermore,compared with the global scanning method,the proposed method is more efficient,with an average time of 10%–20%of the global scanning localization algorithm.Thus,the proposed algorithm is of great significance for laboratory research focused on locating rupture damages sustained by large-sized rock masses or test blocks.展开更多
Coronary artery disease(CAD)is one of themost authentic cardiovascular afflictions because it is an uncommonly overwhelming heart issue.The breakdown of coronary cardiovascular disease is one of the principal sources ...Coronary artery disease(CAD)is one of themost authentic cardiovascular afflictions because it is an uncommonly overwhelming heart issue.The breakdown of coronary cardiovascular disease is one of the principal sources of death all over theworld.Cardiovascular deterioration is a challenge,especially in youthful and rural countries where there is an absence of humantrained professionals.Since heart diseases happen without apparent signs,high-level detection is desirable.This paper proposed a robust and tuned random forest model using the randomized grid search technique to predictCAD.The proposed framework increases the ability of CADpredictions by tracking down risk pointers and learning the confusing joint efforts between them.Nowadays,the healthcare industry has a lot of data but needs to gain more knowledge.Our proposed framework is used for extracting knowledge from data stores and using that knowledge to help doctors accurately and effectively diagnose heart disease(HD).We evaluated the proposed framework over two public databases,Cleveland and Framingham datasets.The datasets were preprocessed by using a cleaning technique,a normalization technique,and an outlier detection technique.Secondly,the principal component analysis(PCA)algorithm was utilized to lessen the feature dimensionality of the two datasets.Finally,we used a hyperparameter tuning technique,randomized grid search,to tune a random forest(RF)machine learning(ML)model.The randomized grid search selected the best parameters and got the ideal CAD analysis.The proposed framework was evaluated and compared with traditional classifiers.Our proposed framework’s accuracy,sensitivity,precision,specificity,and f1-score were 100%.The evaluation of the proposed framework showed that it is an unrivaled perceptive outcome with tuning as opposed to other ongoing existing frameworks.展开更多
Rockburst is a common geological disaster in underground engineering,which seriously threatens the safety of personnel,equipment and property.Utilizing machine learning models to evaluate risk of rockburst is graduall...Rockburst is a common geological disaster in underground engineering,which seriously threatens the safety of personnel,equipment and property.Utilizing machine learning models to evaluate risk of rockburst is gradually becoming a trend.In this study,the integrated algorithms under Gradient Boosting Decision Tree(GBDT)framework were used to evaluate and classify rockburst intensity.First,a total of 301 rock burst data samples were obtained from a case database,and the data were preprocessed using synthetic minority over-sampling technique(SMOTE).Then,the rockburst evaluation models including GBDT,eXtreme Gradient Boosting(XGBoost),Light Gradient Boosting Machine(LightGBM),and Categorical Features Gradient Boosting(CatBoost)were established,and the optimal hyperparameters of the models were obtained through random search grid and five-fold cross-validation.Afterwards,use the optimal hyperparameter configuration to fit the evaluation models,and analyze these models using test set.In order to evaluate the performance,metrics including accuracy,precision,recall,and F1-score were selected to analyze and compare with other machine learning models.Finally,the trained models were used to conduct rock burst risk assessment on rock samples from a mine in Shanxi Province,China,and providing theoretical guidance for the mine's safe production work.The models under the GBDT framework perform well in the evaluation of rockburst levels,and the proposed methods can provide a reliable reference for rockburst risk level analysis and safety management.展开更多
Diagnosing the cardiovascular disease is one of the biggest medical difficulties in recent years.Coronary cardiovascular(CHD)is a kind of heart and blood vascular disease.Predicting this sort of cardiac illness leads ...Diagnosing the cardiovascular disease is one of the biggest medical difficulties in recent years.Coronary cardiovascular(CHD)is a kind of heart and blood vascular disease.Predicting this sort of cardiac illness leads to more precise decisions for cardiac disorders.Implementing Grid Search Optimization(GSO)machine training models is therefore a useful way to forecast the sickness as soon as possible.The state-of-the-art work is the tuning of the hyperparameter together with the selection of the feature by utilizing the model search to minimize the false-negative rate.Three models with a cross-validation approach do the required task.Feature Selection based on the use of statistical and correlation matrices for multivariate analysis.For Random Search and Grid Search models,extensive comparison findings are produced utilizing retrieval,F1 score,and precision measurements.The models are evaluated using the metrics and kappa statistics that illustrate the three models’comparability.The study effort focuses on optimizing function selection,tweaking hyperparameters to improve model accuracy and the prediction of heart disease by examining Framingham datasets using random forestry classification.Tuning the hyperparameter in the model of grid search thus decreases the erroneous rate achieves global optimization.展开更多
In computer vision,emotion recognition using facial expression images is considered an important research issue.Deep learning advances in recent years have aided in attaining improved results in this issue.According t...In computer vision,emotion recognition using facial expression images is considered an important research issue.Deep learning advances in recent years have aided in attaining improved results in this issue.According to recent studies,multiple facial expressions may be included in facial photographs representing a particular type of emotion.It is feasible and useful to convert face photos into collections of visual words and carry out global expression recognition.The main contribution of this paper is to propose a facial expression recognitionmodel(FERM)depending on an optimized Support Vector Machine(SVM).To test the performance of the proposed model(FERM),AffectNet is used.AffectNet uses 1250 emotion-related keywords in six different languages to search three major search engines and get over 1,000,000 facial photos online.The FERM is composed of three main phases:(i)the Data preparation phase,(ii)Applying grid search for optimization,and(iii)the categorization phase.Linear discriminant analysis(LDA)is used to categorize the data into eight labels(neutral,happy,sad,surprised,fear,disgust,angry,and contempt).Due to using LDA,the performance of categorization via SVM has been obviously enhanced.Grid search is used to find the optimal values for hyperparameters of SVM(C and gamma).The proposed optimized SVM algorithm has achieved an accuracy of 99%and a 98%F1 score.展开更多
Waste production rises in tandem with population growth and increased utilization.The indecorous disposal of waste paves the way for huge disaster named as climate change.The National Environment Agency(NEA)of Singapo...Waste production rises in tandem with population growth and increased utilization.The indecorous disposal of waste paves the way for huge disaster named as climate change.The National Environment Agency(NEA)of Singapore oversees the sustainable management of waste across the country.The three main contributors to the solid waste of Singapore are paper and cardboard(P&C),plastic,and food scraps.Besides,they have a negligible rate of recycling.In this study,Machine Learning techniques were utilized to forecast the amount of garbage also known as waste audits.The waste audit would aid the authorities to plan their waste infrastructure.The applied models were k-nearest neighbors,Support Vector Regressor,ExtraTrees,CatBoost,and XGBoost.The XGBoost model with its default parameters performed better with a lower Mean Absolute Percentage Error(MAPE)of 8.3093(P&C waste),8.3217(plastic waste),and 6.9495(food waste).However,Grid Search Optimization(GSO)was used to enhance the parameters of the XGBoost model,increasing its effectiveness.Therefore,the optimized XGBoost algorithm performs the best for P&C,plastics,and food waste with MAPE of 4.9349,6.7967,and 5.9626,respectively.The proposed GSO-XGBoost model yields better results than the other employed models in predicting municipal solid waste.展开更多
The use of machine learning to predict student employability is important in order to analyse a student’s capability to get a job.Based on the results of this type of analysis,university managers can improve the empl...The use of machine learning to predict student employability is important in order to analyse a student’s capability to get a job.Based on the results of this type of analysis,university managers can improve the employability of their students,which can help in attracting students in the future.In addition,learners can focus on the essential skills identified through this analysis during their studies,to increase their employability.An effectivemethod calledOPT-BAG(OPTimisation of BAGging classifiers)was therefore developed to model the problem of predicting the employability of students.This model can help predict the employability of students based on their competencies and can reveal weaknesses that need to be improved.First,we analyse the relationships between several variables and the outcome variable using a correlation heatmap for a student employability dataset.Next,a standard scaler function is applied in the preprocessing module to normalise the variables in the student employability dataset.The training set is then input to our model to identify the optimal parameters for the bagging classifier using a grid search cross-validation technique.Finally,the OPT-BAG model,based on a bagging classifier with optimal parameters found in the previous step,is trained on the training dataset to predict student employability.The empirical outcomes in terms of accuracy,precision,recall,and F1 indicate that the OPT-BAG approach outperforms other cutting-edge machine learning models in terms of predicting student employability.In this study,we also analyse the factors affecting the recruitment process of employers,and find that general appearance,mental alertness,and communication skills are the most important.This indicates that educational institutions should focus on these factors during the learning process to improve student employability.展开更多
The load types in low-voltage distribution systems are diverse.Some loads have current signals that are similar to series fault arcs,making it difficult to effectively detect fault arcs during their occurrence and sus...The load types in low-voltage distribution systems are diverse.Some loads have current signals that are similar to series fault arcs,making it difficult to effectively detect fault arcs during their occurrence and sustained combustion,which can easily lead to serious electrical fire accidents.To address this issue,this paper establishes a fault arc prototype experimental platform,selects multiple commonly used loads for fault arc experiments,and collects data in both normal and fault states.By analyzing waveform characteristics and selecting fault discrimination feature indicators,corresponding feature values are extracted for qualitative analysis to explore changes in timefrequency characteristics of current before and after faults.Multiple features are then selected to form a multidimensional feature vector space to effectively reduce arc misjudgments and construct a fault discrimination feature database.Based on this,a fault arc hazard prediction model is built using random forests.The model’s multiple hyperparameters are simultaneously optimized through grid search,aiming tominimize node information entropy and complete model training,thereby enhancing model robustness and generalization ability.Through experimental verification,the proposed method accurately predicts and classifies fault arcs of different load types,with an average accuracy at least 1%higher than that of the commonly used fault predictionmethods compared in the paper.展开更多
Currently, the second most devastating form of cancer in people, particularly in women, is Breast Cancer (BC). In the healthcare industry, Machine Learning (ML) is commonly employed in fatal disease prediction. Due to...Currently, the second most devastating form of cancer in people, particularly in women, is Breast Cancer (BC). In the healthcare industry, Machine Learning (ML) is commonly employed in fatal disease prediction. Due to breast cancer’s favourable prognosis at an early stage, a model is created to utilize the Dataset on Wisconsin Diagnostic Breast Cancer (WDBC). Conversely, this model’s overarching axiom is to compare the effectiveness of five well-known ML classifiers, including Logistic Regression (LR), Decision Tree (DT), Random Forest (RF), K-Nearest Neighbor (KNN), and Naive Bayes (NB) with the conventional method. To counterbalance the effect with conventional methods, the overarching tactic we utilized was hyperparameter tuning utilizing the grid search method, which improved accuracy, secondary precision, third recall, F1 score and finally the AUC & ROC curve. In this study of hyperparameter tuning model, the rate of accuracy increased from 94.15% to 98.83% whereas the accuracy of the conventional method increased from 93.56% to 97.08%. According to this investigation, KNN outperformed all other classifiers in terms of accuracy, achieving a score of 98.83%. In conclusion, our study shows that KNN works well with the hyper-tuning method. These analyses show that this study prediction approach is useful in prognosticating women with breast cancer with a viable performance and more accurate findings when compared to the conventional approach.展开更多
An improved hybrid Time of Arrival (ToA)/ Angle of Arrival (AoA) location algorithm by adopting Gauss-Newton iterative algorithm is proposed. It is with the advantage of fast convergence and combining with the grid-se...An improved hybrid Time of Arrival (ToA)/ Angle of Arrival (AoA) location algorithm by adopting Gauss-Newton iterative algorithm is proposed. It is with the advantage of fast convergence and combining with the grid-search-based method to optimize the initial object coordinates of the iteration, meanwhile, under the condition of small measurement errors caused by noises of ToA and AoA, the algorithm performance can be improved effectively. In the Non-Line-of-Sight (NLoS) environments of the Wireless Sensor Network (WSN), simulation results show that improved accuracy is gained with moderate flexibility and fast steady convergence compared with the existing algorithms.展开更多
In this paper, a template matching and location method, which has been rapidly adopted in microseismic research in recent years, is applied to laboratory acoustic emission(AE) monitoring. First, we used traditional me...In this paper, a template matching and location method, which has been rapidly adopted in microseismic research in recent years, is applied to laboratory acoustic emission(AE) monitoring. First, we used traditional methods to detect P-wave first motions and locate AE hypocenters in three dimensions. In addition, we selected events located with sufficient accuracy(normally corresponding AE events of relatively larger energy, showing clear P-wave first motion and a higher signal-to-noise ratio in most channels) as template events. Then, the template events were used to scan and match other poorly located events in triggered event records or weak events in continuous records. Through crosscorrelation of the multi-channel waveforms between the template and the event to be detected, the weak signal was detected and located using a grid-searching algorithm(with the grid centered at the template hypocenter). In order to examine the performance of the approach, we calibrated the proposed method using experimental data of different rocks and different types of experiments. The results show that the proposed method can significantly improve the detection capability and location accuracy, and can be applied to various laboratory and in situ experiments, which use multi-channel AE monitoring with waveforms recorded in either triggering or continuous mode.展开更多
Lumber moisture content(LMC) is the important parameter to judge the dryness of lumber and the quality of wooden products.Nevertheless the data acquired are mostly redundant and incomplete because of the complexity of...Lumber moisture content(LMC) is the important parameter to judge the dryness of lumber and the quality of wooden products.Nevertheless the data acquired are mostly redundant and incomplete because of the complexity of the course of drying,by interference factors that exist in the dryness environment and by the physical characteristics of the lumber itself.To improve the measuring accuracy and reliability of LMC,the optimal support vector machine(SVM) algorithm was put forward for regression analysis LMC.Environmental factors such as air temperature and relative humidity were considered,the data of which were extracted with the principle component analysis method.The regression and prediction of SVM was optimized based on the grid search(GS) technique.Groups of data were sampled and analyzed,and simulation comparison of forecasting performance shows that the main component data were extracted to speed up the convergence rate of the optimum algorithm.The GS-SVM shows a better performance in solving the LMC measuring and forecasting problem.展开更多
Novel Coronavirus Disease(COVID-19)is a communicable disease that originated during December 2019,when China officially informed the World Health Organization(WHO)regarding the constellation of cases of the disease in...Novel Coronavirus Disease(COVID-19)is a communicable disease that originated during December 2019,when China officially informed the World Health Organization(WHO)regarding the constellation of cases of the disease in the city of Wuhan.Subsequently,the disease started spreading to the rest of the world.Until this point in time,no specific vaccine or medicine is available for the prevention and cure of the disease.Several research works are being carried out in the fields of medicinal and pharmaceutical sciences aided by data analytics and machine learning in the direction of treatment and early detection of this viral disease.The present report describes the use of machine learning algorithms[Linear and Logistic Regression,Decision Tree(DT),K-Nearest Neighbor(KNN),Support Vector Machine(SVM),and SVM with Grid Search]for the prediction and classification in relation to COVID-19.The data used for experimentation was the COVID-19 dataset acquired from the Center for Systems Science and Engineering(CSSE),Johns Hopkins University(JHU).The assimilated results indicated that the risk period for the patients is 12–14 days,beyond which the probability of survival of the patient may increase.In addition,it was also indicated that the probability of death in COVID cases increases with age.The death probability was found to be higher in males as compared to females.SVM with Grid search methods demonstrated the highest accuracy of approximately 95%,followed by the decision tree algorithm with an accuracy of approximately 94%.The present study and analysis pave a way in the direction of attribute correlation,estimation of survival days,and the prediction of death probability.The findings of the present study clearly indicate that machine learning algorithms have strong capabilities of prediction and classification in relation to COVID-19 as well.展开更多
Using Support Vector Machine(SVM)requires the selection of several parameters such as multi-class strategy type(one-against-all or one-against-one),the regularization parameter C,kernel function and their parameters.T...Using Support Vector Machine(SVM)requires the selection of several parameters such as multi-class strategy type(one-against-all or one-against-one),the regularization parameter C,kernel function and their parameters.The choice of these parameters has a great influence on the performance of the final classifier.This paper considers the grid search method and the particle swarm optimization(PSO)technique that have allowed to quickly select and scan a large space of SVM parameters.A comparative study of the SVM models is also presented to examine the convergence speed and the results of each model.SVM is applied to handwritten Arabic characters learning,with a database containing 4840 Arabic characters in their different positions(isolated,beginning,middle and end).Some very promising results have been achieved.展开更多
As the main body of air traffic control safety,the air traffic controller is an important part of the whole air traffic control system. According to the relevant data of civil aviation over the years,a mapping model b...As the main body of air traffic control safety,the air traffic controller is an important part of the whole air traffic control system. According to the relevant data of civil aviation over the years,a mapping model between flight support sorties and air traffic controller demand is constructed by using the prediction algorithm of support vector regression(SVR) based on grid search and cross-validation. Then the model predicts the demand for air traffic controllers in seven regions. Additionally,according to the employment data of civil aviation universities,the future training scale of air traffic controller is predicted. The forecast results show that the average relative error of the number of controllers predicted by the algorithm is 1.73%,and the prediction accuracy is higher than traditional regression algorithms. Under the influence of the epidemic,the demand for air traffic controllers will decrease in the short term,but with the control of the epidemic,the demand of air traffic controllers will return to the pre-epidemic level and gradually increase. It is expected that the controller increment will be about 816 by 2028. The forecast results of the demand for air traffic controllers provide a theoretical basis for the introduction and training of medium and long-term air traffic controllers,and also provide method guidance and decision support for the establishment of professional reserve and dynamic control mechanism in the air traffic control system.展开更多
A dense seismic network was installed in the capital region of China in recent years,which makes it possible to resolve the focal mechanisms of small earthquakes. We gathered large earthquake focal mechanisms from the...A dense seismic network was installed in the capital region of China in recent years,which makes it possible to resolve the focal mechanisms of small earthquakes. We gathered large earthquake focal mechanisms from the last fifty years and moderate or small earthquake focal mechanisms from between 2002 and 2004,and calculated the present tectonic stress field of the capital region by the grid search method, which weighs different sized earthquakes and can improve the accuracy of the stress field inversion. The analysis of inversion results of different sub-regions shows that the azinuth of the maximum principal compressive stress axis is NE43°- 86° in the Beijing-Zhangjiakou-Datong area,NE38°-86° in the Tangshan area,and NE79°- 81° in the Xingtai area. Inversion results of this paper are similar to previous results,which proves the correctness of the approach. As revealed by the results,the stress field of the capital region is characterized by overall consistency and sub-regional differences. This study provides reference for earthquake mechanism explanation and geodynamics research.展开更多
Microseismic monitoring provides a valuable tool for evaluating the effectiveness of hydraulic fracturing operations.However,robust detection and accurate location of microseismic events are challenging due to the low...Microseismic monitoring provides a valuable tool for evaluating the effectiveness of hydraulic fracturing operations.However,robust detection and accurate location of microseismic events are challenging due to the low signal to noise ratio(SNR)of their signals on seismograms.In a downhole monitoring setting,P-wave polarization direction measured from 3-component records is usually considered as the backazimuth of the microseismic event,i.e.,the direction of the event.The direction and arrival time difference between the P and S waves is used to locate the seismic event.When SNR is low,an accurate estimate of event backazimuth becomes very challenging with the traditional covariance matrix method.Here we propose to employ a master event and use a grid search method to find the backazimuth of a target event that maximizes the dot product of the two backazimuthal vectors of the master and target events.We compared the backazimuths measured with the proposed grid-search and the conventional covariance-matrix methods using a large synthetic dataset.We found that the grid-search method yields more accurate backazimuth estimates from low SNR records when measurements are made at single geophone level.When array data are combined,the proposed method also has some advantage over the covariance-matrix method,especially when the number of geophones is low.We also applied the method to a microseismic dataset acquired by a hydraulic fracturing project at a shale play site in southwestern China and found that the relocated microseismic events tend to align along existing faults more tightly than those in the original catalog.展开更多
The optimization of network topologies to retain the generalization ability by deciding when to stop overtraining an artificial neural network(ANN)is an existing vital challenge in ANN prediction works.The larger the ...The optimization of network topologies to retain the generalization ability by deciding when to stop overtraining an artificial neural network(ANN)is an existing vital challenge in ANN prediction works.The larger the dataset the ANN is trained with,the better generalization the prediction can give.In this paper,a large dataset of atmospheric corrosion data of carbon steel compiled from several resources is used to train and test a multilayer backpropagation ANN model as well as two conventional corrosion prediction models(linear and Klinesmith models).Unlike previous related works,a grid search-based hyperparameter tuning is performed to develop multiple hyperparameter combinations(network topologies)to train multiple ANNs with mini-batch stochastic gradient descent optimization algorithm to facilitate the training of a large dataset.After that,one selection strategy for the optimal hyperparameter combination is applied by an early stopping method to guarantee the generalization ability of the optimal network model.The correlation coefficients(R)of the ANN model can explain about 80%(more than 75%)of the variance of atmospheric corrosion of carbon steel,and the root mean square errors(RMSE)of three models show that the ANN model gives a better performance than the other two models with acceptable generalization.The influence of input parameters on the output is highlighted by using the fuzzy curve analysis method.The result reveals that TOW,Cl-and SO2 are the most important atmospheric chemical variables,which have a well-known nonlinear relationship with atmospheric corrosion.展开更多
Generative Adversarial Networks(GANs)are neural networks that allow models to learn deep representations without requiring a large amount of training data.Semi-Supervised GAN Classifiers are a recent innovation in GAN...Generative Adversarial Networks(GANs)are neural networks that allow models to learn deep representations without requiring a large amount of training data.Semi-Supervised GAN Classifiers are a recent innovation in GANs,where GANs are used to classify generated images into real and fake and multiple classes,similar to a general multi-class classifier.However,GANs have a sophisticated design that can be challenging to train.This is because obtaining the proper set of parameters for all models-generator,discriminator,and classifier is complex.As a result,training a single GAN model for different datasets may not produce satisfactory results.Therefore,this study proposes an SGAN model(Semi-Supervised GAN Classifier).First,a baseline model was constructed.The model was then enhanced by leveraging the Sine-Cosine Algorithm and Synthetic Minority Oversampling Technique(SMOTE).SMOTE was used to address class imbalances in the dataset,while Sine Cosine Algorithm(SCA)was used to optimize the weights of the classifier models.The optimal set of hyperparameters(learning rate and batch size)were obtained using grid manual search.Four well-known benchmark datasets and a set of evaluation measures were used to validate the proposed model.The proposed method was then compared against existing models,and the results on each dataset were recorded and demonstrated the effectiveness of the proposed model.The proposed model successfully showed improved test accuracy scores of 1%,2%,15%,and 5%on benchmarking multimedia datasets;Modified National Institute of Standards and Technology(MNIST)digits,Fashion MNIST,Pneumonia Chest X-ray,and Facial Emotion Detection Dataset,respectively.展开更多
基金supported by the Natural Science Foundation of Henan Province(No.222300420596)China Railway Science and Technology Innovation Program Funded Project(CZ02-Special-03)Science and Technology Innovation Project funded by China Railway Tunnel Group(Tunnel Research 2021-03)。
文摘Acoustic emission(AE)source localization is a fundamental element of rock fracture damage imaging.To improve the efficiency and accuracy of AE source localization,this paper proposes a joint method comprising a three-dimensional(3D)AE source localization simplex method and grid search scanning.Using the concept of the geometry of simplexes,tetrahedral iterations were first conducted to narrow down the suspected source region.This is followed by a process of meshing the region and node searching to scan for optimal solutions,until the source location is determined.The resulting algorithm was tested using the artificial excitation source localization and uniaxial compression tests,after which the localization results were compared with the simplex and exhaustive methods.The results revealed that the localization obtained using the proposed method is more stable and can be effectively avoided compared with the simplex localization method.Furthermore,compared with the global scanning method,the proposed method is more efficient,with an average time of 10%–20%of the global scanning localization algorithm.Thus,the proposed algorithm is of great significance for laboratory research focused on locating rupture damages sustained by large-sized rock masses or test blocks.
文摘Coronary artery disease(CAD)is one of themost authentic cardiovascular afflictions because it is an uncommonly overwhelming heart issue.The breakdown of coronary cardiovascular disease is one of the principal sources of death all over theworld.Cardiovascular deterioration is a challenge,especially in youthful and rural countries where there is an absence of humantrained professionals.Since heart diseases happen without apparent signs,high-level detection is desirable.This paper proposed a robust and tuned random forest model using the randomized grid search technique to predictCAD.The proposed framework increases the ability of CADpredictions by tracking down risk pointers and learning the confusing joint efforts between them.Nowadays,the healthcare industry has a lot of data but needs to gain more knowledge.Our proposed framework is used for extracting knowledge from data stores and using that knowledge to help doctors accurately and effectively diagnose heart disease(HD).We evaluated the proposed framework over two public databases,Cleveland and Framingham datasets.The datasets were preprocessed by using a cleaning technique,a normalization technique,and an outlier detection technique.Secondly,the principal component analysis(PCA)algorithm was utilized to lessen the feature dimensionality of the two datasets.Finally,we used a hyperparameter tuning technique,randomized grid search,to tune a random forest(RF)machine learning(ML)model.The randomized grid search selected the best parameters and got the ideal CAD analysis.The proposed framework was evaluated and compared with traditional classifiers.Our proposed framework’s accuracy,sensitivity,precision,specificity,and f1-score were 100%.The evaluation of the proposed framework showed that it is an unrivaled perceptive outcome with tuning as opposed to other ongoing existing frameworks.
基金Project(52161135301)supported by the International Cooperation and Exchange of the National Natural Science Foundation of ChinaProject(202306370296)supported by China Scholarship Council。
文摘Rockburst is a common geological disaster in underground engineering,which seriously threatens the safety of personnel,equipment and property.Utilizing machine learning models to evaluate risk of rockburst is gradually becoming a trend.In this study,the integrated algorithms under Gradient Boosting Decision Tree(GBDT)framework were used to evaluate and classify rockburst intensity.First,a total of 301 rock burst data samples were obtained from a case database,and the data were preprocessed using synthetic minority over-sampling technique(SMOTE).Then,the rockburst evaluation models including GBDT,eXtreme Gradient Boosting(XGBoost),Light Gradient Boosting Machine(LightGBM),and Categorical Features Gradient Boosting(CatBoost)were established,and the optimal hyperparameters of the models were obtained through random search grid and five-fold cross-validation.Afterwards,use the optimal hyperparameter configuration to fit the evaluation models,and analyze these models using test set.In order to evaluate the performance,metrics including accuracy,precision,recall,and F1-score were selected to analyze and compare with other machine learning models.Finally,the trained models were used to conduct rock burst risk assessment on rock samples from a mine in Shanxi Province,China,and providing theoretical guidance for the mine's safe production work.The models under the GBDT framework perform well in the evaluation of rockburst levels,and the proposed methods can provide a reliable reference for rockburst risk level analysis and safety management.
文摘Diagnosing the cardiovascular disease is one of the biggest medical difficulties in recent years.Coronary cardiovascular(CHD)is a kind of heart and blood vascular disease.Predicting this sort of cardiac illness leads to more precise decisions for cardiac disorders.Implementing Grid Search Optimization(GSO)machine training models is therefore a useful way to forecast the sickness as soon as possible.The state-of-the-art work is the tuning of the hyperparameter together with the selection of the feature by utilizing the model search to minimize the false-negative rate.Three models with a cross-validation approach do the required task.Feature Selection based on the use of statistical and correlation matrices for multivariate analysis.For Random Search and Grid Search models,extensive comparison findings are produced utilizing retrieval,F1 score,and precision measurements.The models are evaluated using the metrics and kappa statistics that illustrate the three models’comparability.The study effort focuses on optimizing function selection,tweaking hyperparameters to improve model accuracy and the prediction of heart disease by examining Framingham datasets using random forestry classification.Tuning the hyperparameter in the model of grid search thus decreases the erroneous rate achieves global optimization.
文摘In computer vision,emotion recognition using facial expression images is considered an important research issue.Deep learning advances in recent years have aided in attaining improved results in this issue.According to recent studies,multiple facial expressions may be included in facial photographs representing a particular type of emotion.It is feasible and useful to convert face photos into collections of visual words and carry out global expression recognition.The main contribution of this paper is to propose a facial expression recognitionmodel(FERM)depending on an optimized Support Vector Machine(SVM).To test the performance of the proposed model(FERM),AffectNet is used.AffectNet uses 1250 emotion-related keywords in six different languages to search three major search engines and get over 1,000,000 facial photos online.The FERM is composed of three main phases:(i)the Data preparation phase,(ii)Applying grid search for optimization,and(iii)the categorization phase.Linear discriminant analysis(LDA)is used to categorize the data into eight labels(neutral,happy,sad,surprised,fear,disgust,angry,and contempt).Due to using LDA,the performance of categorization via SVM has been obviously enhanced.Grid search is used to find the optimal values for hyperparameters of SVM(C and gamma).The proposed optimized SVM algorithm has achieved an accuracy of 99%and a 98%F1 score.
文摘Waste production rises in tandem with population growth and increased utilization.The indecorous disposal of waste paves the way for huge disaster named as climate change.The National Environment Agency(NEA)of Singapore oversees the sustainable management of waste across the country.The three main contributors to the solid waste of Singapore are paper and cardboard(P&C),plastic,and food scraps.Besides,they have a negligible rate of recycling.In this study,Machine Learning techniques were utilized to forecast the amount of garbage also known as waste audits.The waste audit would aid the authorities to plan their waste infrastructure.The applied models were k-nearest neighbors,Support Vector Regressor,ExtraTrees,CatBoost,and XGBoost.The XGBoost model with its default parameters performed better with a lower Mean Absolute Percentage Error(MAPE)of 8.3093(P&C waste),8.3217(plastic waste),and 6.9495(food waste).However,Grid Search Optimization(GSO)was used to enhance the parameters of the XGBoost model,increasing its effectiveness.Therefore,the optimized XGBoost algorithm performs the best for P&C,plastics,and food waste with MAPE of 4.9349,6.7967,and 5.9626,respectively.The proposed GSO-XGBoost model yields better results than the other employed models in predicting municipal solid waste.
文摘The use of machine learning to predict student employability is important in order to analyse a student’s capability to get a job.Based on the results of this type of analysis,university managers can improve the employability of their students,which can help in attracting students in the future.In addition,learners can focus on the essential skills identified through this analysis during their studies,to increase their employability.An effectivemethod calledOPT-BAG(OPTimisation of BAGging classifiers)was therefore developed to model the problem of predicting the employability of students.This model can help predict the employability of students based on their competencies and can reveal weaknesses that need to be improved.First,we analyse the relationships between several variables and the outcome variable using a correlation heatmap for a student employability dataset.Next,a standard scaler function is applied in the preprocessing module to normalise the variables in the student employability dataset.The training set is then input to our model to identify the optimal parameters for the bagging classifier using a grid search cross-validation technique.Finally,the OPT-BAG model,based on a bagging classifier with optimal parameters found in the previous step,is trained on the training dataset to predict student employability.The empirical outcomes in terms of accuracy,precision,recall,and F1 indicate that the OPT-BAG approach outperforms other cutting-edge machine learning models in terms of predicting student employability.In this study,we also analyse the factors affecting the recruitment process of employers,and find that general appearance,mental alertness,and communication skills are the most important.This indicates that educational institutions should focus on these factors during the learning process to improve student employability.
基金This work was funded by Beijing Key Laboratory of Distribution Transformer Energy-Saving Technology(China Electric Power Research Institute).
文摘The load types in low-voltage distribution systems are diverse.Some loads have current signals that are similar to series fault arcs,making it difficult to effectively detect fault arcs during their occurrence and sustained combustion,which can easily lead to serious electrical fire accidents.To address this issue,this paper establishes a fault arc prototype experimental platform,selects multiple commonly used loads for fault arc experiments,and collects data in both normal and fault states.By analyzing waveform characteristics and selecting fault discrimination feature indicators,corresponding feature values are extracted for qualitative analysis to explore changes in timefrequency characteristics of current before and after faults.Multiple features are then selected to form a multidimensional feature vector space to effectively reduce arc misjudgments and construct a fault discrimination feature database.Based on this,a fault arc hazard prediction model is built using random forests.The model’s multiple hyperparameters are simultaneously optimized through grid search,aiming tominimize node information entropy and complete model training,thereby enhancing model robustness and generalization ability.Through experimental verification,the proposed method accurately predicts and classifies fault arcs of different load types,with an average accuracy at least 1%higher than that of the commonly used fault predictionmethods compared in the paper.
文摘Currently, the second most devastating form of cancer in people, particularly in women, is Breast Cancer (BC). In the healthcare industry, Machine Learning (ML) is commonly employed in fatal disease prediction. Due to breast cancer’s favourable prognosis at an early stage, a model is created to utilize the Dataset on Wisconsin Diagnostic Breast Cancer (WDBC). Conversely, this model’s overarching axiom is to compare the effectiveness of five well-known ML classifiers, including Logistic Regression (LR), Decision Tree (DT), Random Forest (RF), K-Nearest Neighbor (KNN), and Naive Bayes (NB) with the conventional method. To counterbalance the effect with conventional methods, the overarching tactic we utilized was hyperparameter tuning utilizing the grid search method, which improved accuracy, secondary precision, third recall, F1 score and finally the AUC & ROC curve. In this study of hyperparameter tuning model, the rate of accuracy increased from 94.15% to 98.83% whereas the accuracy of the conventional method increased from 93.56% to 97.08%. According to this investigation, KNN outperformed all other classifiers in terms of accuracy, achieving a score of 98.83%. In conclusion, our study shows that KNN works well with the hyper-tuning method. These analyses show that this study prediction approach is useful in prognosticating women with breast cancer with a viable performance and more accurate findings when compared to the conventional approach.
基金supported by National Natural Science Foundation of China under Grant No.61172073State Key Laboratory of Networking and Switching Technology (Beijing Universityof Posts and Telecommunications) under Grant No.SKLNST-2009-1-09+1 种基金Open Research Fund of National Mobile Communications Research Laboratory, Southeast University, P. R.ChinaChina Fundamental Research Funds for the Central Universities:Beijing Jiaotong University
文摘An improved hybrid Time of Arrival (ToA)/ Angle of Arrival (AoA) location algorithm by adopting Gauss-Newton iterative algorithm is proposed. It is with the advantage of fast convergence and combining with the grid-search-based method to optimize the initial object coordinates of the iteration, meanwhile, under the condition of small measurement errors caused by noises of ToA and AoA, the algorithm performance can be improved effectively. In the Non-Line-of-Sight (NLoS) environments of the Wireless Sensor Network (WSN), simulation results show that improved accuracy is gained with moderate flexibility and fast steady convergence compared with the existing algorithms.
基金funding support from Grant-in-Aid for Scientific Research(Grant No.19H00722)by Japan Society for the Promotion of Science(JSPS)。
文摘In this paper, a template matching and location method, which has been rapidly adopted in microseismic research in recent years, is applied to laboratory acoustic emission(AE) monitoring. First, we used traditional methods to detect P-wave first motions and locate AE hypocenters in three dimensions. In addition, we selected events located with sufficient accuracy(normally corresponding AE events of relatively larger energy, showing clear P-wave first motion and a higher signal-to-noise ratio in most channels) as template events. Then, the template events were used to scan and match other poorly located events in triggered event records or weak events in continuous records. Through crosscorrelation of the multi-channel waveforms between the template and the event to be detected, the weak signal was detected and located using a grid-searching algorithm(with the grid centered at the template hypocenter). In order to examine the performance of the approach, we calibrated the proposed method using experimental data of different rocks and different types of experiments. The results show that the proposed method can significantly improve the detection capability and location accuracy, and can be applied to various laboratory and in situ experiments, which use multi-channel AE monitoring with waveforms recorded in either triggering or continuous mode.
基金supported by the Natural Science Foundation of China(Grant No.31470715),(Grant No.31470714)the Fundamental Research Funds for the Central Universities(2572016EBT1)
文摘Lumber moisture content(LMC) is the important parameter to judge the dryness of lumber and the quality of wooden products.Nevertheless the data acquired are mostly redundant and incomplete because of the complexity of the course of drying,by interference factors that exist in the dryness environment and by the physical characteristics of the lumber itself.To improve the measuring accuracy and reliability of LMC,the optimal support vector machine(SVM) algorithm was put forward for regression analysis LMC.Environmental factors such as air temperature and relative humidity were considered,the data of which were extracted with the principle component analysis method.The regression and prediction of SVM was optimized based on the grid search(GS) technique.Groups of data were sampled and analyzed,and simulation comparison of forecasting performance shows that the main component data were extracted to speed up the convergence rate of the optimum algorithm.The GS-SVM shows a better performance in solving the LMC measuring and forecasting problem.
文摘Novel Coronavirus Disease(COVID-19)is a communicable disease that originated during December 2019,when China officially informed the World Health Organization(WHO)regarding the constellation of cases of the disease in the city of Wuhan.Subsequently,the disease started spreading to the rest of the world.Until this point in time,no specific vaccine or medicine is available for the prevention and cure of the disease.Several research works are being carried out in the fields of medicinal and pharmaceutical sciences aided by data analytics and machine learning in the direction of treatment and early detection of this viral disease.The present report describes the use of machine learning algorithms[Linear and Logistic Regression,Decision Tree(DT),K-Nearest Neighbor(KNN),Support Vector Machine(SVM),and SVM with Grid Search]for the prediction and classification in relation to COVID-19.The data used for experimentation was the COVID-19 dataset acquired from the Center for Systems Science and Engineering(CSSE),Johns Hopkins University(JHU).The assimilated results indicated that the risk period for the patients is 12–14 days,beyond which the probability of survival of the patient may increase.In addition,it was also indicated that the probability of death in COVID cases increases with age.The death probability was found to be higher in males as compared to females.SVM with Grid search methods demonstrated the highest accuracy of approximately 95%,followed by the decision tree algorithm with an accuracy of approximately 94%.The present study and analysis pave a way in the direction of attribute correlation,estimation of survival days,and the prediction of death probability.The findings of the present study clearly indicate that machine learning algorithms have strong capabilities of prediction and classification in relation to COVID-19 as well.
文摘Using Support Vector Machine(SVM)requires the selection of several parameters such as multi-class strategy type(one-against-all or one-against-one),the regularization parameter C,kernel function and their parameters.The choice of these parameters has a great influence on the performance of the final classifier.This paper considers the grid search method and the particle swarm optimization(PSO)technique that have allowed to quickly select and scan a large space of SVM parameters.A comparative study of the SVM models is also presented to examine the convergence speed and the results of each model.SVM is applied to handwritten Arabic characters learning,with a database containing 4840 Arabic characters in their different positions(isolated,beginning,middle and end).Some very promising results have been achieved.
基金supported by the National Natural Science Foundation of China(No.71971114)。
文摘As the main body of air traffic control safety,the air traffic controller is an important part of the whole air traffic control system. According to the relevant data of civil aviation over the years,a mapping model between flight support sorties and air traffic controller demand is constructed by using the prediction algorithm of support vector regression(SVR) based on grid search and cross-validation. Then the model predicts the demand for air traffic controllers in seven regions. Additionally,according to the employment data of civil aviation universities,the future training scale of air traffic controller is predicted. The forecast results show that the average relative error of the number of controllers predicted by the algorithm is 1.73%,and the prediction accuracy is higher than traditional regression algorithms. Under the influence of the epidemic,the demand for air traffic controllers will decrease in the short term,but with the control of the epidemic,the demand of air traffic controllers will return to the pre-epidemic level and gradually increase. It is expected that the controller increment will be about 816 by 2028. The forecast results of the demand for air traffic controllers provide a theoretical basis for the introduction and training of medium and long-term air traffic controllers,and also provide method guidance and decision support for the establishment of professional reserve and dynamic control mechanism in the air traffic control system.
基金sponsored by the Special Fund of Fundamental Scientific Research Operating Expenses for Higher School of Central Government(Projects for creation teams ZY20110101)the Special Fund for the Earthquake Scientific Research of China(201208009)National Natural Science Foundation of China(41074072)
文摘A dense seismic network was installed in the capital region of China in recent years,which makes it possible to resolve the focal mechanisms of small earthquakes. We gathered large earthquake focal mechanisms from the last fifty years and moderate or small earthquake focal mechanisms from between 2002 and 2004,and calculated the present tectonic stress field of the capital region by the grid search method, which weighs different sized earthquakes and can improve the accuracy of the stress field inversion. The analysis of inversion results of different sub-regions shows that the azinuth of the maximum principal compressive stress axis is NE43°- 86° in the Beijing-Zhangjiakou-Datong area,NE38°-86° in the Tangshan area,and NE79°- 81° in the Xingtai area. Inversion results of this paper are similar to previous results,which proves the correctness of the approach. As revealed by the results,the stress field of the capital region is characterized by overall consistency and sub-regional differences. This study provides reference for earthquake mechanism explanation and geodynamics research.
基金supported by the National Natural Science Foundation of China(Grants No.41630209 and 41974045)
文摘Microseismic monitoring provides a valuable tool for evaluating the effectiveness of hydraulic fracturing operations.However,robust detection and accurate location of microseismic events are challenging due to the low signal to noise ratio(SNR)of their signals on seismograms.In a downhole monitoring setting,P-wave polarization direction measured from 3-component records is usually considered as the backazimuth of the microseismic event,i.e.,the direction of the event.The direction and arrival time difference between the P and S waves is used to locate the seismic event.When SNR is low,an accurate estimate of event backazimuth becomes very challenging with the traditional covariance matrix method.Here we propose to employ a master event and use a grid search method to find the backazimuth of a target event that maximizes the dot product of the two backazimuthal vectors of the master and target events.We compared the backazimuths measured with the proposed grid-search and the conventional covariance-matrix methods using a large synthetic dataset.We found that the grid-search method yields more accurate backazimuth estimates from low SNR records when measurements are made at single geophone level.When array data are combined,the proposed method also has some advantage over the covariance-matrix method,especially when the number of geophones is low.We also applied the method to a microseismic dataset acquired by a hydraulic fracturing project at a shale play site in southwestern China and found that the relocated microseismic events tend to align along existing faults more tightly than those in the original catalog.
基金supported by National Key R&D Program of China[Grant Number 2017YFB0203703]111 Project[Grant Number B12012]Fundamental Research Funds for the Central Universities[Grant Number FRF-GF-19-029B].
文摘The optimization of network topologies to retain the generalization ability by deciding when to stop overtraining an artificial neural network(ANN)is an existing vital challenge in ANN prediction works.The larger the dataset the ANN is trained with,the better generalization the prediction can give.In this paper,a large dataset of atmospheric corrosion data of carbon steel compiled from several resources is used to train and test a multilayer backpropagation ANN model as well as two conventional corrosion prediction models(linear and Klinesmith models).Unlike previous related works,a grid search-based hyperparameter tuning is performed to develop multiple hyperparameter combinations(network topologies)to train multiple ANNs with mini-batch stochastic gradient descent optimization algorithm to facilitate the training of a large dataset.After that,one selection strategy for the optimal hyperparameter combination is applied by an early stopping method to guarantee the generalization ability of the optimal network model.The correlation coefficients(R)of the ANN model can explain about 80%(more than 75%)of the variance of atmospheric corrosion of carbon steel,and the root mean square errors(RMSE)of three models show that the ANN model gives a better performance than the other two models with acceptable generalization.The influence of input parameters on the output is highlighted by using the fuzzy curve analysis method.The result reveals that TOW,Cl-and SO2 are the most important atmospheric chemical variables,which have a well-known nonlinear relationship with atmospheric corrosion.
基金This research was supported by Universiti Teknologi PETRONAS,under the Yayasan Universiti Teknologi PETRONAS(YUTP)Fundamental Research Grant Scheme(YUTPFRG/015LC0-308).
文摘Generative Adversarial Networks(GANs)are neural networks that allow models to learn deep representations without requiring a large amount of training data.Semi-Supervised GAN Classifiers are a recent innovation in GANs,where GANs are used to classify generated images into real and fake and multiple classes,similar to a general multi-class classifier.However,GANs have a sophisticated design that can be challenging to train.This is because obtaining the proper set of parameters for all models-generator,discriminator,and classifier is complex.As a result,training a single GAN model for different datasets may not produce satisfactory results.Therefore,this study proposes an SGAN model(Semi-Supervised GAN Classifier).First,a baseline model was constructed.The model was then enhanced by leveraging the Sine-Cosine Algorithm and Synthetic Minority Oversampling Technique(SMOTE).SMOTE was used to address class imbalances in the dataset,while Sine Cosine Algorithm(SCA)was used to optimize the weights of the classifier models.The optimal set of hyperparameters(learning rate and batch size)were obtained using grid manual search.Four well-known benchmark datasets and a set of evaluation measures were used to validate the proposed model.The proposed method was then compared against existing models,and the results on each dataset were recorded and demonstrated the effectiveness of the proposed model.The proposed model successfully showed improved test accuracy scores of 1%,2%,15%,and 5%on benchmarking multimedia datasets;Modified National Institute of Standards and Technology(MNIST)digits,Fashion MNIST,Pneumonia Chest X-ray,and Facial Emotion Detection Dataset,respectively.