A rotating packed bed is a typical chemical process enhancement equipment that can strengthen micromixing and mass transfer.During the operation of the rotating packed bed,the nonreactants and products irregularly adh...A rotating packed bed is a typical chemical process enhancement equipment that can strengthen micromixing and mass transfer.During the operation of the rotating packed bed,the nonreactants and products irregularly adhere to the wire mesh packing in the rotor,thus resulting in an imbalance in the vibration of the rotor,which may cause serious damage to the bearing and material leakage.This study proposes a model prediction for estimating the bearing residual life of a rotating packed bed based on rotor imbalance response analysis.This method is used to determine the influence of the mass on the imbalance in the vibration of the rotor on bearing damage.The major influence on rotor vibration was found to be exerted by the imbalanced mass and its distribution radius,as revealed by the results of orthogonal experiments.Through implementing finite element analysis,the imbalance response curve for the rotating packed bed rotor was obtained,and a correlation among rotor imbalance mass,distribution radius of imbalance mass,and bearing residue life was established via data fitting.The predicted value of the bearing life can be used as the reference basis for an early safety warning of a rotating packed bed to effectively avoid accidents.展开更多
To improve the hit probability of tank at high speed,a prediction method of projectile-target intersection based on adaptive robust constraint-following control and interval uncertainty analysis is proposed.The method...To improve the hit probability of tank at high speed,a prediction method of projectile-target intersection based on adaptive robust constraint-following control and interval uncertainty analysis is proposed.The method proposed provides a novel way to predict the impact point of projectile for moving tank.First,bidirectional stability constraints and stability constraint-following error are constructed using the Udwadia-Kalaba theory,and an adaptive robust constraint-following controller is designed considering uncertainties.Second,the exterior ballistic ordinary differential equation with uncertainties is integrated into the controller,and the pointing control of stability system is extended to the impact-point control of projectile.Third,based on the interval uncertainty analysis method combining Chebyshev polynomial expansion and affine arithmetic,a prediction method of projectile-target intersection is proposed.Finally,the co-simulation experiment is performed by establishing the multi-body system dynamic model of tank and mathematical model of control system.The results demonstrate that the prediction method of projectile-target intersection based on uncertainty analysis can effectively decrease the uncertainties of system,improve the prediction accuracy,and increase the hit probability.The adaptive robust constraint-following control can effectively restrain the uncertainties caused by road excitation and model error.展开更多
Cross-Project Defect Prediction(CPDP)is a method that utilizes historical data from other source projects to train predictive models for defect prediction in the target project.However,existing CPDP methods only consi...Cross-Project Defect Prediction(CPDP)is a method that utilizes historical data from other source projects to train predictive models for defect prediction in the target project.However,existing CPDP methods only consider linear correlations between features(indicators)of the source and target projects.These models are not capable of evaluating non-linear correlations between features when they exist,for example,when there are differences in data distributions between the source and target projects.As a result,the performance of such CPDP models is compromised.In this paper,this paper proposes a novel CPDP method based on Synthetic Minority Oversampling Technique(SMOTE)and Deep Canonical Correlation Analysis(DCCA),referred to as S-DCCA.Canonical Correlation Analysis(CCA)is employed to address the issue of non-linear correlations between features of the source and target projects.S-DCCA extends CCA by incorporating the MlpNet model for feature extraction from the dataset.The redundant features are then eliminated by maximizing the correlated feature subset using the CCA loss function.Finally,cross-project defect prediction is achieved through the application of the SMOTE data sampling technique.Area Under Curve(AUC)and F1 scores(F1)are used as evaluation metrics.This paper conducted experiments on 27 projects from four public datasets to validate the proposed method.The results demonstrate that,on average,our method outperforms all baseline approaches by at least 1.2%in AUC and 5.5%in F1 score.This indicates that the proposed method exhibits favorable performance characteristics.展开更多
Predicting election outcomes is a crucial undertaking,and various methods are employed for this purpose,such as traditional opinion polling,and social media analysis.However,traditional polling approaches often strugg...Predicting election outcomes is a crucial undertaking,and various methods are employed for this purpose,such as traditional opinion polling,and social media analysis.However,traditional polling approaches often struggle to capture the intricate nuances of voter sentiment at local levels,resulting in a limited depth of analysis and understanding.In light of this challenge,this study focuses on predicting elections at the state/regional level along with the country level,intending to offer a comprehensive analysis and deeper insights into the electoral process.To achieve this,the study introduces the Location-Based Election Prediction Model(LEPM),which utilizes social media data,specifically Twitter,and integrates location-aware sentiment analysis techniques at both the state/region and country levels.LEPM predicts the support and opposing strength of each political party/candidate.To determine the location of users/voters who have not disclosed their location information in tweets,the model utilizes a Voter Location Detection(VotLocaDetect)approach,which leverages recent tweets/posts.The sentiment analysis techniques employed in this study include rule-based sentiment analysis,Valence Aware Dictionary and Sentiment Reasoner(VADER)as well as transformers-based sentiment analysis such as Bidirectional Encoder Representations from Transformers(BERT),BERTweet,and Election based BERT(ElecBERT).This study uses the 2020 United States(US)Presidential Election as a case study.By applying the LEPM model to the election,the study demonstrates its ability to accurately predict outcomes in forty-one states,achieving an 0.84 accuracy rate at the state level.Moreover,at the country level,the LEPM model outperforms traditional polling results.With a low Mean Absolute Error(MAE)of 0.87,the model exhibits more precise predictions and serves as a successful alternative to conventional polls and other methodologies.Leveraging the extensive social media data,the LEPM model provides nuanced insights into voter behavior,enabling policymakers to make informed decisions and facilitating in-depth analyses of elections.The study emphasizes the importance of using social media data for reliable election prediction and offers implications for enhancing prediction accuracy and understanding voter sentiment and behavior.展开更多
The price prediction task is a well-studied problem due to its impact on the business domain.There are several research studies that have been conducted to predict the future price of items by capturing the patterns o...The price prediction task is a well-studied problem due to its impact on the business domain.There are several research studies that have been conducted to predict the future price of items by capturing the patterns of price change,but there is very limited work to study the price prediction of seasonal goods(e.g.,Christmas gifts).Seasonal items’prices have different patterns than normal items;this can be linked to the offers and discounted prices of seasonal items.This lack of research studies motivates the current work to investigate the problem of seasonal items’prices as a time series task.We proposed utilizing two different approaches to address this problem,namely,1)machine learning(ML)-based models and 2)deep learning(DL)-based models.Thus,this research tuned a set of well-known predictive models on a real-life dataset.Those models are ensemble learning-based models,random forest,Ridge,Lasso,and Linear regression.Moreover,two new DL architectures based on gated recurrent unit(GRU)and long short-term memory(LSTM)models are proposed.Then,the performance of the utilized ensemble learning and classic ML models are compared against the proposed two DL architectures on different accuracy metrics,where the evaluation includes both numerical and visual comparisons of the examined models.The obtained results show that the ensemble learning models outperformed the classic machine learning-based models(e.g.,linear regression and random forest)and the DL-based models.展开更多
Machine learning algorithms (MLs) can potentially improve disease diagnostics, leading to early detection and treatment of these diseases. As a malignant tumor whose primary focus is located in the bronchial mucosal e...Machine learning algorithms (MLs) can potentially improve disease diagnostics, leading to early detection and treatment of these diseases. As a malignant tumor whose primary focus is located in the bronchial mucosal epithelium, lung cancer has the highest mortality and morbidity among cancer types, threatening health and life of patients suffering from the disease. Machine learning algorithms such as Random Forest (RF), Support Vector Machine (SVM), K-Nearest Neighbor (KNN) and Naïve Bayes (NB) have been used for lung cancer prediction. However they still face challenges such as high dimensionality of the feature space, over-fitting, high computational complexity, noise and missing data, low accuracies, low precision and high error rates. Ensemble learning, which combines classifiers, may be helpful to boost prediction on new data. However, current ensemble ML techniques rarely consider comprehensive evaluation metrics to evaluate the performance of individual classifiers. The main purpose of this study was to develop an ensemble classifier that improves lung cancer prediction. An ensemble machine learning algorithm is developed based on RF, SVM, NB, and KNN. Feature selection is done based on Principal Component Analysis (PCA) and Analysis of Variance (ANOVA). This algorithm is then executed on lung cancer data and evaluated using execution time, true positives (TP), true negatives (TN), false positives (FP), false negatives (FN), false positive rate (FPR), recall (R), precision (P) and F-measure (FM). Experimental results show that the proposed ensemble classifier has the best classification of 0.9825% with the lowest error rate of 0.0193. This is followed by SVM in which the probability of having the best classification is 0.9652% at an error rate of 0.0206. On the other hand, NB had the worst performance of 0.8475% classification at 0.0738 error rate.展开更多
In the existing landslide susceptibility prediction(LSP)models,the influences of random errors in landslide conditioning factors on LSP are not considered,instead the original conditioning factors are directly taken a...In the existing landslide susceptibility prediction(LSP)models,the influences of random errors in landslide conditioning factors on LSP are not considered,instead the original conditioning factors are directly taken as the model inputs,which brings uncertainties to LSP results.This study aims to reveal the influence rules of the different proportional random errors in conditioning factors on the LSP un-certainties,and further explore a method which can effectively reduce the random errors in conditioning factors.The original conditioning factors are firstly used to construct original factors-based LSP models,and then different random errors of 5%,10%,15% and 20%are added to these original factors for con-structing relevant errors-based LSP models.Secondly,low-pass filter-based LSP models are constructed by eliminating the random errors using low-pass filter method.Thirdly,the Ruijin County of China with 370 landslides and 16 conditioning factors are used as study case.Three typical machine learning models,i.e.multilayer perceptron(MLP),support vector machine(SVM)and random forest(RF),are selected as LSP models.Finally,the LSP uncertainties are discussed and results show that:(1)The low-pass filter can effectively reduce the random errors in conditioning factors to decrease the LSP uncertainties.(2)With the proportions of random errors increasing from 5%to 20%,the LSP uncertainty increases continuously.(3)The original factors-based models are feasible for LSP in the absence of more accurate conditioning factors.(4)The influence degrees of two uncertainty issues,machine learning models and different proportions of random errors,on the LSP modeling are large and basically the same.(5)The Shapley values effectively explain the internal mechanism of machine learning model predicting landslide sus-ceptibility.In conclusion,greater proportion of random errors in conditioning factors results in higher LSP uncertainty,and low-pass filter can effectively reduce these random errors.展开更多
Cardiovascular Diseases (CVDs) pose a significant global health challenge, necessitating accurate risk prediction for effective preventive measures. This comprehensive comparative study explores the performance of tra...Cardiovascular Diseases (CVDs) pose a significant global health challenge, necessitating accurate risk prediction for effective preventive measures. This comprehensive comparative study explores the performance of traditional Machine Learning (ML) and Deep Learning (DL) models in predicting CVD risk, utilizing a meticulously curated dataset derived from health records. Rigorous preprocessing, including normalization and outlier removal, enhances model robustness. Diverse ML models (Logistic Regression, Random Forest, Support Vector Machine, K-Nearest Neighbor, Decision Tree, and Gradient Boosting) are compared with a Long Short-Term Memory (LSTM) neural network for DL. Evaluation metrics include accuracy, ROC AUC, computation time, and memory usage. Results identify the Gradient Boosting Classifier and LSTM as top performers, demonstrating high accuracy and ROC AUC scores. Comparative analyses highlight model strengths and limitations, contributing valuable insights for optimizing predictive strategies. This study advances predictive analytics for cardiovascular health, with implications for personalized medicine. The findings underscore the versatility of intelligent systems in addressing health challenges, emphasizing the broader applications of ML and DL in disease identification beyond cardiovascular health.展开更多
In recent decades, Urban Heat Island Effects have become more pronounced and more widely examined. Despite great technological advances, our current societies still experience great spatial disparity in urban forest a...In recent decades, Urban Heat Island Effects have become more pronounced and more widely examined. Despite great technological advances, our current societies still experience great spatial disparity in urban forest access. Urban Heat Island Effects are measurable phenomenon that are being experienced by the world’s most urbanized areas, including increased summer high temperatures and lower evapotranspiration from having impervious surfaces instead of vegetation and trees. Tree canopy cover is our natural mitigation tool that absorbs sunlight for photosynthesis, protects humans from incoming radiation, and releases cooling moisture into the air. Unfortunately, urban areas typically have low levels of vegetation. Vulnerable urban communities are lower-income areas of inner cities with less access to heat protection like air conditioners. This study uses mean evapotranspiration levels to assess the variability of urban heat island effects across the state of Tennessee. Results show that increased developed land surface cover in Tennessee creates measurable changes in atmospheric evapotranspiration. As a result, the mean evapotranspiration levels in areas with less tree vegetation are significantly lower than the surrounding forested areas. Central areas of urban cities in Tennessee had lower mean evapotranspiration recordings than surrounding areas with less development. This work demonstrates the need for increased tree canopy coverage.展开更多
An efficient and accurate prediction of a precise tidal level in estuaries and coastal areas is indispensable for the management and decision-making of human activity in the field wok of marine engineering. The variat...An efficient and accurate prediction of a precise tidal level in estuaries and coastal areas is indispensable for the management and decision-making of human activity in the field wok of marine engineering. The variation of the tidal level is a time-varying process. The time-varying factors including interference from the external environment that cause the change of tides are fairly complicated. Furthermore, tidal variations are affected not only by periodic movement of celestial bodies but also by time-varying interference from the external environment. Consequently, for the efficient and precise tidal level prediction, a neuro-fuzzy hybrid technology based on the combination of harmonic analysis and adaptive network-based fuzzy inference system(ANFIS)model is utilized to construct a precise tidal level prediction system, which takes both advantages of the harmonic analysis method and the ANFIS network. The proposed prediction model is composed of two modules: the astronomical tide module caused by celestial bodies' movement and the non-astronomical tide module caused by various meteorological and other environmental factors. To generate a fuzzy inference system(FIS) structure,three approaches which include grid partition(GP), fuzzy c-means(FCM) and sub-clustering(SC) are used in the ANFIS network constructing process. Furthermore, to obtain the optimal ANFIS based prediction model, large numbers of simulation experiments are implemented for each FIS generating approach. In this tidal prediction study, the optimal ANFIS model is used to predict the non-astronomical tide module, while the conventional harmonic analysis model is used to predict the astronomical tide module. The final prediction result is performed by combining the estimation outputs of the harmonious analysis model and the optimal ANFIS model. To demonstrate the applicability and capability of the proposed novel prediction model, measured tidal level samples of Fort Pulaski tidal station are selected as the testing database. Simulation and experimental results confirm that the proposed prediction approach can achieve precise predictions for the tidal level with high accuracy, satisfactory convergence and stability.展开更多
Due to the advances of intelligent transportation system(ITSs),traffic forecasting has gained significant interest as robust traffic prediction acts as an important part in different ITSs namely traffic signal control...Due to the advances of intelligent transportation system(ITSs),traffic forecasting has gained significant interest as robust traffic prediction acts as an important part in different ITSs namely traffic signal control,navigation,route mapping,etc.The traffic prediction model aims to predict the traffic conditions based on the past traffic data.For more accurate traffic prediction,this study proposes an optimal deep learning-enabled statistical analysis model.This study offers the design of optimal convolutional neural network with attention long short term memory(OCNN-ALSTM)model for traffic prediction.The proposed OCNN-ALSTM technique primarily preprocesses the traffic data by the use of min-max normalization technique.Besides,OCNN-ALSTM technique was executed for classifying and predicting the traffic data in real time cases.For enhancing the predictive outcomes of the OCNN-ALSTM technique,the bird swarm algorithm(BSA)is employed to it and thereby overall efficacy of the network gets improved.The design of BSA for optimal hyperparameter tuning of the CNN-ALSTM model shows the novelty of the work.The experimental validation of the OCNNALSTM technique is performed using benchmark datasets and the results are examined under several aspects.The simulation results reported the enhanced outcomes of the OCNN-ALSTM model over the recent methods under several dimensions.展开更多
An accurate landslide displacement prediction is an important part of landslide warning system. Aiming at the dynamic characteristics of landslide evolution and the shortcomings of traditional static prediction models...An accurate landslide displacement prediction is an important part of landslide warning system. Aiming at the dynamic characteristics of landslide evolution and the shortcomings of traditional static prediction models, this paper proposes a dynamic prediction model of landslide displacement based on singular spectrum analysis(SSA) and stack long short-term memory(SLSTM) network. The SSA is used to decompose the landslide accumulated displacement time series data into trend term and periodic term displacement subsequences. A cubic polynomial function is used to predict the trend term displacement subsequence, and the SLSTM neural network is used to predict the periodic term displacement subsequence. At the same time, the Bayesian optimization algorithm is used to determine that the SLSTM network input sequence length is 12 and the number of hidden layer nodes is 18. The SLSTM network is updated by adding predicted values to the training set to achieve dynamic displacement prediction. Finally, the accumulated landslide displacement is obtained by superimposing the predicted value of each displacement subsequence. The proposed model was verified on the Xintan landslide in Hubei Province, China. The results show that when predicting the displacement of the periodic term, the SLSTM network has higher prediction accuracy than the support vector machine(SVM) and auto regressive integrated moving average(ARIMA). The mean relative error(MRE) is reduced by 4.099% and 3.548% respectively, while the root mean square error(RMSE) is reduced by 5.830 mm and 3.854 mm respectively. It is concluded that the SLSTM network model can better simulate the dynamic characteristics of landslides.展开更多
Blast-induced dominant vibration frequency(DVF)involves a complex,nonlinear and small sample system considering rock properties,blasting parameters and topography.In this study,a combination of grey relational analysi...Blast-induced dominant vibration frequency(DVF)involves a complex,nonlinear and small sample system considering rock properties,blasting parameters and topography.In this study,a combination of grey relational analysis and dimensional analysis procedures for prediction of dominant vibration frequency are presented.Six factors are selected from extensive effect factor sequences based on grey relational analysis,and then a novel blast-induced dominant vibration frequency prediction is obtained by dimensional analysis.In addition,the prediction is simplified by sensitivity analysis with 195 experimental blast records.Validation is carried out for the proposed formula based on the site test database of the firstperiod blasting excavation in the Guangdong Lufeng Nuclear Power Plant(GLNPP).The results show the proposed approach has a higher fitting degree and smaller mean error when compared with traditional predictions.展开更多
Despite of the small amount in the atmosphere,ozone is one of the most critical atmospheric component as it protects human beings and any other life on the earth from the sun's high frequency ultraviolet radiation...Despite of the small amount in the atmosphere,ozone is one of the most critical atmospheric component as it protects human beings and any other life on the earth from the sun's high frequency ultraviolet radiation. In recent decades,the global ozone depletion caused by human activities is w ell know n and produces an " ozone hole",the most direct consequence of w hich is the increase in ultraviolet radiation,w hich w ill affect human survival,climatic environment,ecological environment and other important adverse impacts. Due to the implementation of the M ontreal protocol and other agreement,the total amount of ozone depleting substance in the atmosphere has been prominent reduced,w hich w ill lead to a new round of regional climate change.Therefore,predicting the changes of the total ozone in the future w ill have an important guiding significance for predicting the future climate change and making reasonable measures to deal w ith the climate change. In this paper,based on the ozone data of 1979 to 2016 in the southern hemisphere and ARIM A model algorithm,using time series analysis,w e obtain prediction effect of ARIM A model is good by Ljung-Box Q-test and R^2,and the model can be used to predict the future ozone change. With the help of SPSS softw are,the future trend of the total ozone can be predicted in the future 50 years. Based on the above experiment results,the global ozone change in the future 50 years can be forecasted,namely the atmospheric ozone layer w ill return to its 1980's standard by the middle of this century at the global scale.展开更多
Electricity consumption forecasting is one of the most important tasks for power system workers,and plays an important role in regional power systems.Due to the difference in the trend of power load and the past in th...Electricity consumption forecasting is one of the most important tasks for power system workers,and plays an important role in regional power systems.Due to the difference in the trend of power load and the past in the new normal,the influencing factors are more diversified,which makes it more difficult to predict the current electricity consumption.In this paper,the grey system theory and BP neural network are combined to predict the annual electricity consumption in Jiangsu.According to the historical data of annual electricity consumption and the six factors affecting electricity consumption,the gray correlation analysis method is used to screen the important factors,and three factors with large correlation degree are selected as the input parameters of BP neural network.The power forecasting model uses nearly 18 years of data to train and validate the model.The results show that the gray correlation analysis and BP neural network method have higher accuracy in power consumption prediction,and the calculation is more convenient than traditional methods.展开更多
Based on the principle of Bayesian discriminant analysis, we established a model of Bayesian discriminant analysis for predicting coal and gas outbursts. We selected five major indices which affect outbursts, i.e., in...Based on the principle of Bayesian discriminant analysis, we established a model of Bayesian discriminant analysis for predicting coal and gas outbursts. We selected five major indices which affect outbursts, i.e., initial speed of methane diffusion, a consistent coal coefficient, gas pressure, destructive style of coal and mining depth, as discriminating factors of the model. In our model, we divided the type of coal and gas outbursts into four grades regarded as four normal populations. We then obtained the corresponding discriminant functions through training a set of data from engineering examples as learning samples and evaluated their criteria by a back substitution method to verify the optimal properties of the model. Finally, we applied the model to the prediction of coal and gas outbursts in the Yunnan Enhong Mine. Our results coincided completely with the actual situation. These results show that a model of Bayesian discriminant analysis has excellent recognition performance, high prediction accuracy and a low error rate and is an effective method to predict coal and gas outbursts.展开更多
The first automobile engine remanufacture company in China, Jinan Fuqiang Power Co, Ltd, was introduced. The engine remanufacturing technological process of this company was described. The benefit statistic of remanuf...The first automobile engine remanufacture company in China, Jinan Fuqiang Power Co, Ltd, was introduced. The engine remanufacturing technological process of this company was described. The benefit statistic of remanufacturing 10000 Styer engines were analyzed, and the contribution of engine remanufacturing to cycle economy was predicted. The results show that remanufacturing engineering could use the maximal additional values of obsolete engines, and make contributions to materials conservation, capital saving, energy conservation and environment protection. 10000 engines are supposed to be remanufactured per year, the following benefits would be obtained: reclaiming additional values of RMB3.23 hundred million, saving metallic materials about 7.65 thousand tons, saving energy of 16 million kilowatt-hours, reducing emission of CO2 about 11.3[CD*2]15.3 thousand tons, and providing employment for 500 persons. According to the survey and analysis, tremendous benefits will be gained by the year of 2020. For example, reclaiming additional values per year of 1424[CD*2]2236 hundred million RMB, saving energy per year of 60[CD*2]90 hundred million kilowatt-hours, reducing emission of CO2 about 6.67[CD*2]9.69 million tons. It can be deduced that developing remanufacturing will play an important role in enriching the cycle economy and accelerating the development of national economy.展开更多
By comparing with ENSO events that ever happened in the history, the basic features and probable causes of the anomalous sea surface temperature of the tropical Pacific Ocean during 1997 and 1998 have been analyzed di...By comparing with ENSO events that ever happened in the history, the basic features and probable causes of the anomalous sea surface temperature of the tropical Pacific Ocean during 1997 and 1998 have been analyzed diagnostically. It is found that the 1997/1998 El Nino had significant abnormalities and peculiarities. It differs from the previous El Ni駉 events falling into the simple eastern pattern or western pattern. The predictions of 1997/1998 El Ni駉 event have also been tested with an intermediate ocean-atmosphere coupled dynamic model. The results show that the skills of the 0~24 lead month forecasts for the warm event are all above 0.5. The predictions of the mature phase and the later stages of the warm event are better than those of the beginning phase.展开更多
基金the High-Performance Computing Platform of Beijing University of Chemical Technology(BUCT)for supporting this papersupported by the Fundamental Research Funds for the Central Universities(JD2319)+2 种基金the CNOOC Technical Cooperation Project(ZX2022ZCTYF7612)the National Natural Science Foundation of China(51775029,52004014)the Chinese Universities Scientific Fund(XK2020-04)。
文摘A rotating packed bed is a typical chemical process enhancement equipment that can strengthen micromixing and mass transfer.During the operation of the rotating packed bed,the nonreactants and products irregularly adhere to the wire mesh packing in the rotor,thus resulting in an imbalance in the vibration of the rotor,which may cause serious damage to the bearing and material leakage.This study proposes a model prediction for estimating the bearing residual life of a rotating packed bed based on rotor imbalance response analysis.This method is used to determine the influence of the mass on the imbalance in the vibration of the rotor on bearing damage.The major influence on rotor vibration was found to be exerted by the imbalanced mass and its distribution radius,as revealed by the results of orthogonal experiments.Through implementing finite element analysis,the imbalance response curve for the rotating packed bed rotor was obtained,and a correlation among rotor imbalance mass,distribution radius of imbalance mass,and bearing residue life was established via data fitting.The predicted value of the bearing life can be used as the reference basis for an early safety warning of a rotating packed bed to effectively avoid accidents.
基金financially supported by the National Natural Science Foundation of China(Grant 52175099)the China Postdoctoral Science Foundation(Grant No.2020M671494)+1 种基金the Jiangsu Planned Projects for Postdoctoral Research Funds(Grant No.2020Z179)the Nanjing University of Science and Technology Independent Research Program(Grant No.30920021105)。
文摘To improve the hit probability of tank at high speed,a prediction method of projectile-target intersection based on adaptive robust constraint-following control and interval uncertainty analysis is proposed.The method proposed provides a novel way to predict the impact point of projectile for moving tank.First,bidirectional stability constraints and stability constraint-following error are constructed using the Udwadia-Kalaba theory,and an adaptive robust constraint-following controller is designed considering uncertainties.Second,the exterior ballistic ordinary differential equation with uncertainties is integrated into the controller,and the pointing control of stability system is extended to the impact-point control of projectile.Third,based on the interval uncertainty analysis method combining Chebyshev polynomial expansion and affine arithmetic,a prediction method of projectile-target intersection is proposed.Finally,the co-simulation experiment is performed by establishing the multi-body system dynamic model of tank and mathematical model of control system.The results demonstrate that the prediction method of projectile-target intersection based on uncertainty analysis can effectively decrease the uncertainties of system,improve the prediction accuracy,and increase the hit probability.The adaptive robust constraint-following control can effectively restrain the uncertainties caused by road excitation and model error.
基金NationalNatural Science Foundation of China,Grant/AwardNumber:61867004National Natural Science Foundation of China Youth Fund,Grant/Award Number:41801288.
文摘Cross-Project Defect Prediction(CPDP)is a method that utilizes historical data from other source projects to train predictive models for defect prediction in the target project.However,existing CPDP methods only consider linear correlations between features(indicators)of the source and target projects.These models are not capable of evaluating non-linear correlations between features when they exist,for example,when there are differences in data distributions between the source and target projects.As a result,the performance of such CPDP models is compromised.In this paper,this paper proposes a novel CPDP method based on Synthetic Minority Oversampling Technique(SMOTE)and Deep Canonical Correlation Analysis(DCCA),referred to as S-DCCA.Canonical Correlation Analysis(CCA)is employed to address the issue of non-linear correlations between features of the source and target projects.S-DCCA extends CCA by incorporating the MlpNet model for feature extraction from the dataset.The redundant features are then eliminated by maximizing the correlated feature subset using the CCA loss function.Finally,cross-project defect prediction is achieved through the application of the SMOTE data sampling technique.Area Under Curve(AUC)and F1 scores(F1)are used as evaluation metrics.This paper conducted experiments on 27 projects from four public datasets to validate the proposed method.The results demonstrate that,on average,our method outperforms all baseline approaches by at least 1.2%in AUC and 5.5%in F1 score.This indicates that the proposed method exhibits favorable performance characteristics.
基金funded by the Beijing Municipal Natural Science Foundation(Grant No.4212026)the Foundation Enhancement Program(Grant No.2021-JCJQ-JJ-0059).
文摘Predicting election outcomes is a crucial undertaking,and various methods are employed for this purpose,such as traditional opinion polling,and social media analysis.However,traditional polling approaches often struggle to capture the intricate nuances of voter sentiment at local levels,resulting in a limited depth of analysis and understanding.In light of this challenge,this study focuses on predicting elections at the state/regional level along with the country level,intending to offer a comprehensive analysis and deeper insights into the electoral process.To achieve this,the study introduces the Location-Based Election Prediction Model(LEPM),which utilizes social media data,specifically Twitter,and integrates location-aware sentiment analysis techniques at both the state/region and country levels.LEPM predicts the support and opposing strength of each political party/candidate.To determine the location of users/voters who have not disclosed their location information in tweets,the model utilizes a Voter Location Detection(VotLocaDetect)approach,which leverages recent tweets/posts.The sentiment analysis techniques employed in this study include rule-based sentiment analysis,Valence Aware Dictionary and Sentiment Reasoner(VADER)as well as transformers-based sentiment analysis such as Bidirectional Encoder Representations from Transformers(BERT),BERTweet,and Election based BERT(ElecBERT).This study uses the 2020 United States(US)Presidential Election as a case study.By applying the LEPM model to the election,the study demonstrates its ability to accurately predict outcomes in forty-one states,achieving an 0.84 accuracy rate at the state level.Moreover,at the country level,the LEPM model outperforms traditional polling results.With a low Mean Absolute Error(MAE)of 0.87,the model exhibits more precise predictions and serves as a successful alternative to conventional polls and other methodologies.Leveraging the extensive social media data,the LEPM model provides nuanced insights into voter behavior,enabling policymakers to make informed decisions and facilitating in-depth analyses of elections.The study emphasizes the importance of using social media data for reliable election prediction and offers implications for enhancing prediction accuracy and understanding voter sentiment and behavior.
文摘The price prediction task is a well-studied problem due to its impact on the business domain.There are several research studies that have been conducted to predict the future price of items by capturing the patterns of price change,but there is very limited work to study the price prediction of seasonal goods(e.g.,Christmas gifts).Seasonal items’prices have different patterns than normal items;this can be linked to the offers and discounted prices of seasonal items.This lack of research studies motivates the current work to investigate the problem of seasonal items’prices as a time series task.We proposed utilizing two different approaches to address this problem,namely,1)machine learning(ML)-based models and 2)deep learning(DL)-based models.Thus,this research tuned a set of well-known predictive models on a real-life dataset.Those models are ensemble learning-based models,random forest,Ridge,Lasso,and Linear regression.Moreover,two new DL architectures based on gated recurrent unit(GRU)and long short-term memory(LSTM)models are proposed.Then,the performance of the utilized ensemble learning and classic ML models are compared against the proposed two DL architectures on different accuracy metrics,where the evaluation includes both numerical and visual comparisons of the examined models.The obtained results show that the ensemble learning models outperformed the classic machine learning-based models(e.g.,linear regression and random forest)and the DL-based models.
文摘Machine learning algorithms (MLs) can potentially improve disease diagnostics, leading to early detection and treatment of these diseases. As a malignant tumor whose primary focus is located in the bronchial mucosal epithelium, lung cancer has the highest mortality and morbidity among cancer types, threatening health and life of patients suffering from the disease. Machine learning algorithms such as Random Forest (RF), Support Vector Machine (SVM), K-Nearest Neighbor (KNN) and Naïve Bayes (NB) have been used for lung cancer prediction. However they still face challenges such as high dimensionality of the feature space, over-fitting, high computational complexity, noise and missing data, low accuracies, low precision and high error rates. Ensemble learning, which combines classifiers, may be helpful to boost prediction on new data. However, current ensemble ML techniques rarely consider comprehensive evaluation metrics to evaluate the performance of individual classifiers. The main purpose of this study was to develop an ensemble classifier that improves lung cancer prediction. An ensemble machine learning algorithm is developed based on RF, SVM, NB, and KNN. Feature selection is done based on Principal Component Analysis (PCA) and Analysis of Variance (ANOVA). This algorithm is then executed on lung cancer data and evaluated using execution time, true positives (TP), true negatives (TN), false positives (FP), false negatives (FN), false positive rate (FPR), recall (R), precision (P) and F-measure (FM). Experimental results show that the proposed ensemble classifier has the best classification of 0.9825% with the lowest error rate of 0.0193. This is followed by SVM in which the probability of having the best classification is 0.9652% at an error rate of 0.0206. On the other hand, NB had the worst performance of 0.8475% classification at 0.0738 error rate.
基金This work is funded by the National Natural Science Foundation of China(Grant Nos.42377164 and 52079062)the National Science Fund for Distinguished Young Scholars of China(Grant No.52222905).
文摘In the existing landslide susceptibility prediction(LSP)models,the influences of random errors in landslide conditioning factors on LSP are not considered,instead the original conditioning factors are directly taken as the model inputs,which brings uncertainties to LSP results.This study aims to reveal the influence rules of the different proportional random errors in conditioning factors on the LSP un-certainties,and further explore a method which can effectively reduce the random errors in conditioning factors.The original conditioning factors are firstly used to construct original factors-based LSP models,and then different random errors of 5%,10%,15% and 20%are added to these original factors for con-structing relevant errors-based LSP models.Secondly,low-pass filter-based LSP models are constructed by eliminating the random errors using low-pass filter method.Thirdly,the Ruijin County of China with 370 landslides and 16 conditioning factors are used as study case.Three typical machine learning models,i.e.multilayer perceptron(MLP),support vector machine(SVM)and random forest(RF),are selected as LSP models.Finally,the LSP uncertainties are discussed and results show that:(1)The low-pass filter can effectively reduce the random errors in conditioning factors to decrease the LSP uncertainties.(2)With the proportions of random errors increasing from 5%to 20%,the LSP uncertainty increases continuously.(3)The original factors-based models are feasible for LSP in the absence of more accurate conditioning factors.(4)The influence degrees of two uncertainty issues,machine learning models and different proportions of random errors,on the LSP modeling are large and basically the same.(5)The Shapley values effectively explain the internal mechanism of machine learning model predicting landslide sus-ceptibility.In conclusion,greater proportion of random errors in conditioning factors results in higher LSP uncertainty,and low-pass filter can effectively reduce these random errors.
文摘Cardiovascular Diseases (CVDs) pose a significant global health challenge, necessitating accurate risk prediction for effective preventive measures. This comprehensive comparative study explores the performance of traditional Machine Learning (ML) and Deep Learning (DL) models in predicting CVD risk, utilizing a meticulously curated dataset derived from health records. Rigorous preprocessing, including normalization and outlier removal, enhances model robustness. Diverse ML models (Logistic Regression, Random Forest, Support Vector Machine, K-Nearest Neighbor, Decision Tree, and Gradient Boosting) are compared with a Long Short-Term Memory (LSTM) neural network for DL. Evaluation metrics include accuracy, ROC AUC, computation time, and memory usage. Results identify the Gradient Boosting Classifier and LSTM as top performers, demonstrating high accuracy and ROC AUC scores. Comparative analyses highlight model strengths and limitations, contributing valuable insights for optimizing predictive strategies. This study advances predictive analytics for cardiovascular health, with implications for personalized medicine. The findings underscore the versatility of intelligent systems in addressing health challenges, emphasizing the broader applications of ML and DL in disease identification beyond cardiovascular health.
文摘In recent decades, Urban Heat Island Effects have become more pronounced and more widely examined. Despite great technological advances, our current societies still experience great spatial disparity in urban forest access. Urban Heat Island Effects are measurable phenomenon that are being experienced by the world’s most urbanized areas, including increased summer high temperatures and lower evapotranspiration from having impervious surfaces instead of vegetation and trees. Tree canopy cover is our natural mitigation tool that absorbs sunlight for photosynthesis, protects humans from incoming radiation, and releases cooling moisture into the air. Unfortunately, urban areas typically have low levels of vegetation. Vulnerable urban communities are lower-income areas of inner cities with less access to heat protection like air conditioners. This study uses mean evapotranspiration levels to assess the variability of urban heat island effects across the state of Tennessee. Results show that increased developed land surface cover in Tennessee creates measurable changes in atmospheric evapotranspiration. As a result, the mean evapotranspiration levels in areas with less tree vegetation are significantly lower than the surrounding forested areas. Central areas of urban cities in Tennessee had lower mean evapotranspiration recordings than surrounding areas with less development. This work demonstrates the need for increased tree canopy coverage.
基金The National Natural Science Foundation of China under contract No.51379002the Fundamental Research Funds for the Central Universities of China under contract Nos 3132016322 and 3132016314the Applied Basic Research Project Fund of the Chinese Ministry of Transport of China under contract No.2014329225010
文摘An efficient and accurate prediction of a precise tidal level in estuaries and coastal areas is indispensable for the management and decision-making of human activity in the field wok of marine engineering. The variation of the tidal level is a time-varying process. The time-varying factors including interference from the external environment that cause the change of tides are fairly complicated. Furthermore, tidal variations are affected not only by periodic movement of celestial bodies but also by time-varying interference from the external environment. Consequently, for the efficient and precise tidal level prediction, a neuro-fuzzy hybrid technology based on the combination of harmonic analysis and adaptive network-based fuzzy inference system(ANFIS)model is utilized to construct a precise tidal level prediction system, which takes both advantages of the harmonic analysis method and the ANFIS network. The proposed prediction model is composed of two modules: the astronomical tide module caused by celestial bodies' movement and the non-astronomical tide module caused by various meteorological and other environmental factors. To generate a fuzzy inference system(FIS) structure,three approaches which include grid partition(GP), fuzzy c-means(FCM) and sub-clustering(SC) are used in the ANFIS network constructing process. Furthermore, to obtain the optimal ANFIS based prediction model, large numbers of simulation experiments are implemented for each FIS generating approach. In this tidal prediction study, the optimal ANFIS model is used to predict the non-astronomical tide module, while the conventional harmonic analysis model is used to predict the astronomical tide module. The final prediction result is performed by combining the estimation outputs of the harmonious analysis model and the optimal ANFIS model. To demonstrate the applicability and capability of the proposed novel prediction model, measured tidal level samples of Fort Pulaski tidal station are selected as the testing database. Simulation and experimental results confirm that the proposed prediction approach can achieve precise predictions for the tidal level with high accuracy, satisfactory convergence and stability.
基金This research was supported by Basic Science Research Program through the National Research Foundation of Korea(NRF)funded by the Ministry of Education(NRF-2021R1A6A1A03039493).
文摘Due to the advances of intelligent transportation system(ITSs),traffic forecasting has gained significant interest as robust traffic prediction acts as an important part in different ITSs namely traffic signal control,navigation,route mapping,etc.The traffic prediction model aims to predict the traffic conditions based on the past traffic data.For more accurate traffic prediction,this study proposes an optimal deep learning-enabled statistical analysis model.This study offers the design of optimal convolutional neural network with attention long short term memory(OCNN-ALSTM)model for traffic prediction.The proposed OCNN-ALSTM technique primarily preprocesses the traffic data by the use of min-max normalization technique.Besides,OCNN-ALSTM technique was executed for classifying and predicting the traffic data in real time cases.For enhancing the predictive outcomes of the OCNN-ALSTM technique,the bird swarm algorithm(BSA)is employed to it and thereby overall efficacy of the network gets improved.The design of BSA for optimal hyperparameter tuning of the CNN-ALSTM model shows the novelty of the work.The experimental validation of the OCNNALSTM technique is performed using benchmark datasets and the results are examined under several aspects.The simulation results reported the enhanced outcomes of the OCNN-ALSTM model over the recent methods under several dimensions.
基金supported by the Natural Science Foundation of Shaanxi Province under Grant 2019JQ206in part by the Science and Technology Department of Shaanxi Province under Grant 2020CGXNG-009in part by the Education Department of Shaanxi Province under Grant 17JK0346。
文摘An accurate landslide displacement prediction is an important part of landslide warning system. Aiming at the dynamic characteristics of landslide evolution and the shortcomings of traditional static prediction models, this paper proposes a dynamic prediction model of landslide displacement based on singular spectrum analysis(SSA) and stack long short-term memory(SLSTM) network. The SSA is used to decompose the landslide accumulated displacement time series data into trend term and periodic term displacement subsequences. A cubic polynomial function is used to predict the trend term displacement subsequence, and the SLSTM neural network is used to predict the periodic term displacement subsequence. At the same time, the Bayesian optimization algorithm is used to determine that the SLSTM network input sequence length is 12 and the number of hidden layer nodes is 18. The SLSTM network is updated by adding predicted values to the training set to achieve dynamic displacement prediction. Finally, the accumulated landslide displacement is obtained by superimposing the predicted value of each displacement subsequence. The proposed model was verified on the Xintan landslide in Hubei Province, China. The results show that when predicting the displacement of the periodic term, the SLSTM network has higher prediction accuracy than the support vector machine(SVM) and auto regressive integrated moving average(ARIMA). The mean relative error(MRE) is reduced by 4.099% and 3.548% respectively, while the root mean square error(RMSE) is reduced by 5.830 mm and 3.854 mm respectively. It is concluded that the SLSTM network model can better simulate the dynamic characteristics of landslides.
基金National Natural Science Funds for Distinguished Young Scholar under Grant No.51009086Hubei Key Laboratory of Roadway Bridge and Structure Engineering under Grant No.DQJJ201313Major State Basic Research Development Program of China(973 Program)under Grant No.2010CB732001
文摘Blast-induced dominant vibration frequency(DVF)involves a complex,nonlinear and small sample system considering rock properties,blasting parameters and topography.In this study,a combination of grey relational analysis and dimensional analysis procedures for prediction of dominant vibration frequency are presented.Six factors are selected from extensive effect factor sequences based on grey relational analysis,and then a novel blast-induced dominant vibration frequency prediction is obtained by dimensional analysis.In addition,the prediction is simplified by sensitivity analysis with 195 experimental blast records.Validation is carried out for the proposed formula based on the site test database of the firstperiod blasting excavation in the Guangdong Lufeng Nuclear Power Plant(GLNPP).The results show the proposed approach has a higher fitting degree and smaller mean error when compared with traditional predictions.
基金supported by the key laboratory fund of Hubei province (Grant No. 2015KLA0,DZ-2016-01-H )graduate research innovation Project of NCIAE (No. YKY2016-08 )the science and technology research projects of Hebei province (Grant No. ZD 2016 106 )
文摘Despite of the small amount in the atmosphere,ozone is one of the most critical atmospheric component as it protects human beings and any other life on the earth from the sun's high frequency ultraviolet radiation. In recent decades,the global ozone depletion caused by human activities is w ell know n and produces an " ozone hole",the most direct consequence of w hich is the increase in ultraviolet radiation,w hich w ill affect human survival,climatic environment,ecological environment and other important adverse impacts. Due to the implementation of the M ontreal protocol and other agreement,the total amount of ozone depleting substance in the atmosphere has been prominent reduced,w hich w ill lead to a new round of regional climate change.Therefore,predicting the changes of the total ozone in the future w ill have an important guiding significance for predicting the future climate change and making reasonable measures to deal w ith the climate change. In this paper,based on the ozone data of 1979 to 2016 in the southern hemisphere and ARIM A model algorithm,using time series analysis,w e obtain prediction effect of ARIM A model is good by Ljung-Box Q-test and R^2,and the model can be used to predict the future ozone change. With the help of SPSS softw are,the future trend of the total ozone can be predicted in the future 50 years. Based on the above experiment results,the global ozone change in the future 50 years can be forecasted,namely the atmospheric ozone layer w ill return to its 1980's standard by the middle of this century at the global scale.
基金This work is supported by the Natural Science Foundation of the Jiangsu Higher Education Institutions of China(19KJB520028)the Collaborative Innovation Center of Jiangsu Maritime Institute。
文摘Electricity consumption forecasting is one of the most important tasks for power system workers,and plays an important role in regional power systems.Due to the difference in the trend of power load and the past in the new normal,the influencing factors are more diversified,which makes it more difficult to predict the current electricity consumption.In this paper,the grey system theory and BP neural network are combined to predict the annual electricity consumption in Jiangsu.According to the historical data of annual electricity consumption and the six factors affecting electricity consumption,the gray correlation analysis method is used to screen the important factors,and three factors with large correlation degree are selected as the input parameters of BP neural network.The power forecasting model uses nearly 18 years of data to train and validate the model.The results show that the gray correlation analysis and BP neural network method have higher accuracy in power consumption prediction,and the calculation is more convenient than traditional methods.
基金supported by the National Hi-tech Research and Development Program of China (No.2006BAK03B02-04) the New Century Excellent Talent Support Plan of Ministry of Education of China (No.NCET-06-0477)
文摘Based on the principle of Bayesian discriminant analysis, we established a model of Bayesian discriminant analysis for predicting coal and gas outbursts. We selected five major indices which affect outbursts, i.e., initial speed of methane diffusion, a consistent coal coefficient, gas pressure, destructive style of coal and mining depth, as discriminating factors of the model. In our model, we divided the type of coal and gas outbursts into four grades regarded as four normal populations. We then obtained the corresponding discriminant functions through training a set of data from engineering examples as learning samples and evaluated their criteria by a back substitution method to verify the optimal properties of the model. Finally, we applied the model to the prediction of coal and gas outbursts in the Yunnan Enhong Mine. Our results coincided completely with the actual situation. These results show that a model of Bayesian discriminant analysis has excellent recognition performance, high prediction accuracy and a low error rate and is an effective method to predict coal and gas outbursts.
文摘The first automobile engine remanufacture company in China, Jinan Fuqiang Power Co, Ltd, was introduced. The engine remanufacturing technological process of this company was described. The benefit statistic of remanufacturing 10000 Styer engines were analyzed, and the contribution of engine remanufacturing to cycle economy was predicted. The results show that remanufacturing engineering could use the maximal additional values of obsolete engines, and make contributions to materials conservation, capital saving, energy conservation and environment protection. 10000 engines are supposed to be remanufactured per year, the following benefits would be obtained: reclaiming additional values of RMB3.23 hundred million, saving metallic materials about 7.65 thousand tons, saving energy of 16 million kilowatt-hours, reducing emission of CO2 about 11.3[CD*2]15.3 thousand tons, and providing employment for 500 persons. According to the survey and analysis, tremendous benefits will be gained by the year of 2020. For example, reclaiming additional values per year of 1424[CD*2]2236 hundred million RMB, saving energy per year of 60[CD*2]90 hundred million kilowatt-hours, reducing emission of CO2 about 6.67[CD*2]9.69 million tons. It can be deduced that developing remanufacturing will play an important role in enriching the cycle economy and accelerating the development of national economy.
基金Key project of China Meteorological Administration "Studies on the cause of formation and the application of prediction for the heavy rainstorms over Yangtze River and Nenjiang River Basins in 1998" and the sub-project II (96-908-02-05) of National Key
文摘By comparing with ENSO events that ever happened in the history, the basic features and probable causes of the anomalous sea surface temperature of the tropical Pacific Ocean during 1997 and 1998 have been analyzed diagnostically. It is found that the 1997/1998 El Nino had significant abnormalities and peculiarities. It differs from the previous El Ni駉 events falling into the simple eastern pattern or western pattern. The predictions of 1997/1998 El Ni駉 event have also been tested with an intermediate ocean-atmosphere coupled dynamic model. The results show that the skills of the 0~24 lead month forecasts for the warm event are all above 0.5. The predictions of the mature phase and the later stages of the warm event are better than those of the beginning phase.