当前建筑业迅速发展,但随之而来的是频频发生的建筑安全事故,造成不可逆转的损失和伤害。虽然近些年来在建筑安全事故控制方面的研究已取得一定的成果,但建筑安全事故仍未得到有效控制。针对建筑业市政工程安全事故总数和死亡人数,探究...当前建筑业迅速发展,但随之而来的是频频发生的建筑安全事故,造成不可逆转的损失和伤害。虽然近些年来在建筑安全事故控制方面的研究已取得一定的成果,但建筑安全事故仍未得到有效控制。针对建筑业市政工程安全事故总数和死亡人数,探究二者之间的关系,构建灰狼优化算法-支持向量回归机(Grey Wolf Optimization and Support Vactor Regression,GWO-SVR)组合模型,收集2008—2020年每个月的建筑安全事故数据及死亡人数数据集,发现二者之间成正向相关关系,以建筑安全事故数为特征对建筑死亡人数进行预测,精度达到95%以上,对建筑安全资源与人力投入有较大参考价值,有助于提升建筑安全管理水平。展开更多
Purpose:The purpose of this study is to develop and compare model choice strategies in context of logistic regression.Model choice means the choice of the covariates to be included in the model.Design/methodology/appr...Purpose:The purpose of this study is to develop and compare model choice strategies in context of logistic regression.Model choice means the choice of the covariates to be included in the model.Design/methodology/approach:The study is based on Monte Carlo simulations.The methods are compared in terms of three measures of accuracy:specificity and two kinds of sensitivity.A loss function combining sensitivity and specificity is introduced and used for a final comparison.Findings:The choice of method depends on how much the users emphasize sensitivity against specificity.It also depends on the sample size.For a typical logistic regression setting with a moderate sample size and a small to moderate effect size,either BIC,BICc or Lasso seems to be optimal.Research limitations:Numerical simulations cannot cover the whole range of data-generating processes occurring with real-world data.Thus,more simulations are needed.Practical implications:Researchers can refer to these results if they believe that their data-generating process is somewhat similar to some of the scenarios presented in this paper.Alternatively,they could run their own simulations and calculate the loss function.Originality/value:This is a systematic comparison of model choice algorithms and heuristics in context of logistic regression.The distinction between two types of sensitivity and a comparison based on a loss function are methodological novelties.展开更多
The Extensible Markup Language(XML)files,widely used for storing and exchanging information on the web require efficient parsing mechanisms to improve the performance of the applications.With the existing Document Obj...The Extensible Markup Language(XML)files,widely used for storing and exchanging information on the web require efficient parsing mechanisms to improve the performance of the applications.With the existing Document Object Model(DOM)based parsing,the performance degrades due to sequential processing and large memory requirements,thereby requiring an efficient XML parser to mitigate these issues.In this paper,we propose a Parallel XML Tree Generator(PXTG)algorithm for accelerating the parsing of XML files and a Regression-based XML Parsing Framework(RXPF)that analyzes and predicts performance through profiling,regression,and code generation for efficient parsing.The PXTG algorithm is based on dividing the XML file into n parts and producing n trees in parallel.The profiling phase of the RXPF framework produces a dataset by measuring the performance of various parsing models including StAX,SAX,DOM,JDOM,and PXTG on different cores by using multiple file sizes.The regression phase produces the prediction model,based on which the final code for efficient parsing of XML files is produced through the code generation phase.The RXPF framework has shown a significant improvement in performance varying from 9.54%to 32.34%over other existing models used for parsing XML files.展开更多
In view of the composition analysis and identification of ancient glass products, L1 regularization, K-Means cluster analysis, elbow rule and other methods were comprehensively used to build logical regression, cluste...In view of the composition analysis and identification of ancient glass products, L1 regularization, K-Means cluster analysis, elbow rule and other methods were comprehensively used to build logical regression, cluster analysis, hyper-parameter test and other models, and SPSS, Python and other tools were used to obtain the classification rules of glass products under different fluxes, sub classification under different chemical compositions, hyper-parameter K value test and rationality analysis. Research can provide theoretical support for the protection and restoration of ancient glass relics.展开更多
Concentrate copper grade(CCG)is one of the important production indicators of copper flotation processes,and keeping the CCG at the set value is of great significance to the economic benefit of copper flotation indust...Concentrate copper grade(CCG)is one of the important production indicators of copper flotation processes,and keeping the CCG at the set value is of great significance to the economic benefit of copper flotation industrial processes.This paper addresses the fluctuation problem of CCG through an operational optimization method.Firstly,a density-based affinity propagationalgorithm is proposed so that more ideal working condition categories can be obtained for the complex raw ore properties.Next,a Bayesian network(BN)is applied to explore the relationship between the operational variables and the CCG.Based on the analysis results of BN,a weighted Gaussian process regression model is constructed to predict the CCG that a higher prediction accuracy can be obtained.To ensure the predicted CCG is close to the set value with a smaller magnitude of the operation adjustments and a smaller uncertainty of the prediction results,an index-oriented adaptive differential evolution(IOADE)algorithm is proposed,and the convergence performance of IOADE is superior to the traditional differential evolution and adaptive differential evolution methods.Finally,the effectiveness and feasibility of the proposed methods are verified by the experiments on a copper flotation industrial process.展开更多
Reliable long-term settlement prediction of a high embankment relates to mountain infrastructure safety.This study developed a novel hybrid model(NHM)that combines a joint denoising technique with an enhanced gray wol...Reliable long-term settlement prediction of a high embankment relates to mountain infrastructure safety.This study developed a novel hybrid model(NHM)that combines a joint denoising technique with an enhanced gray wolf optimizer(EGWO)-n-support vector regression(n-SVR)method.High-embankment field measurements were preprocessed using the joint denoising technique,which in-cludes complete ensemble empirical mode decomposition,singular value decomposition,and wavelet packet transform.Furthermore,high-embankment settlements were predicted using the EGWO-n-SVR method.In this method,the standard gray wolf optimizer(GWO)was improved to obtain the EGWO to better tune the n-SVR model hyperparameters.The proposed NHM was then tested in two case studies.Finally,the influences of the data division ratio and kernel function on the EGWO-n-SVR forecasting performance and prediction efficiency were investigated.The results indicate that the NHM suppresses noise and restores details in high-embankment field measurements.Simultaneously,the NHM out-performs other alternative prediction methods in prediction accuracy and robustness.This demonstrates that the proposed NHM is effective in predicting high-embankment settlements with noisy field mea-surements.Moreover,the appropriate data division ratio and kernel function for EGWO-n-SVR are 7:3 and radial basis function,respectively.展开更多
Accurately estimating blasting vibration during rock blasting is the foundation of blasting vibration management.In this study,Tuna Swarm Optimization(TSO),Whale Optimization Algorithm(WOA),and Cuckoo Search(CS)were u...Accurately estimating blasting vibration during rock blasting is the foundation of blasting vibration management.In this study,Tuna Swarm Optimization(TSO),Whale Optimization Algorithm(WOA),and Cuckoo Search(CS)were used to optimize two hyperparameters in support vector regression(SVR).Based on these methods,three hybrid models to predict peak particle velocity(PPV)for bench blasting were developed.Eighty-eight samples were collected to establish the PPV database,eight initial blasting parameters were chosen as input parameters for the predictionmodel,and the PPV was the output parameter.As predictive performance evaluation indicators,the coefficient of determination(R2),rootmean square error(RMSE),mean absolute error(MAE),and a10-index were selected.The normalizedmutual information value is then used to evaluate the impact of various input parameters on the PPV prediction outcomes.According to the research findings,TSO,WOA,and CS can all enhance the predictive performance of the SVR model.The TSO-SVR model provides the most accurate predictions.The performances of the optimized hybrid SVR models are superior to the unoptimized traditional prediction model.The maximum charge per delay impacts the PPV prediction value the most.展开更多
文摘当前建筑业迅速发展,但随之而来的是频频发生的建筑安全事故,造成不可逆转的损失和伤害。虽然近些年来在建筑安全事故控制方面的研究已取得一定的成果,但建筑安全事故仍未得到有效控制。针对建筑业市政工程安全事故总数和死亡人数,探究二者之间的关系,构建灰狼优化算法-支持向量回归机(Grey Wolf Optimization and Support Vactor Regression,GWO-SVR)组合模型,收集2008—2020年每个月的建筑安全事故数据及死亡人数数据集,发现二者之间成正向相关关系,以建筑安全事故数为特征对建筑死亡人数进行预测,精度达到95%以上,对建筑安全资源与人力投入有较大参考价值,有助于提升建筑安全管理水平。
文摘Purpose:The purpose of this study is to develop and compare model choice strategies in context of logistic regression.Model choice means the choice of the covariates to be included in the model.Design/methodology/approach:The study is based on Monte Carlo simulations.The methods are compared in terms of three measures of accuracy:specificity and two kinds of sensitivity.A loss function combining sensitivity and specificity is introduced and used for a final comparison.Findings:The choice of method depends on how much the users emphasize sensitivity against specificity.It also depends on the sample size.For a typical logistic regression setting with a moderate sample size and a small to moderate effect size,either BIC,BICc or Lasso seems to be optimal.Research limitations:Numerical simulations cannot cover the whole range of data-generating processes occurring with real-world data.Thus,more simulations are needed.Practical implications:Researchers can refer to these results if they believe that their data-generating process is somewhat similar to some of the scenarios presented in this paper.Alternatively,they could run their own simulations and calculate the loss function.Originality/value:This is a systematic comparison of model choice algorithms and heuristics in context of logistic regression.The distinction between two types of sensitivity and a comparison based on a loss function are methodological novelties.
文摘The Extensible Markup Language(XML)files,widely used for storing and exchanging information on the web require efficient parsing mechanisms to improve the performance of the applications.With the existing Document Object Model(DOM)based parsing,the performance degrades due to sequential processing and large memory requirements,thereby requiring an efficient XML parser to mitigate these issues.In this paper,we propose a Parallel XML Tree Generator(PXTG)algorithm for accelerating the parsing of XML files and a Regression-based XML Parsing Framework(RXPF)that analyzes and predicts performance through profiling,regression,and code generation for efficient parsing.The PXTG algorithm is based on dividing the XML file into n parts and producing n trees in parallel.The profiling phase of the RXPF framework produces a dataset by measuring the performance of various parsing models including StAX,SAX,DOM,JDOM,and PXTG on different cores by using multiple file sizes.The regression phase produces the prediction model,based on which the final code for efficient parsing of XML files is produced through the code generation phase.The RXPF framework has shown a significant improvement in performance varying from 9.54%to 32.34%over other existing models used for parsing XML files.
文摘In view of the composition analysis and identification of ancient glass products, L1 regularization, K-Means cluster analysis, elbow rule and other methods were comprehensively used to build logical regression, cluster analysis, hyper-parameter test and other models, and SPSS, Python and other tools were used to obtain the classification rules of glass products under different fluxes, sub classification under different chemical compositions, hyper-parameter K value test and rationality analysis. Research can provide theoretical support for the protection and restoration of ancient glass relics.
基金supported in part by the National Key Research and Development Program of China(2021YFC2902703)the National Natural Science Foundation of China(62173078,61773105,61533007,61873049,61873053,61703085,61374147)。
文摘Concentrate copper grade(CCG)is one of the important production indicators of copper flotation processes,and keeping the CCG at the set value is of great significance to the economic benefit of copper flotation industrial processes.This paper addresses the fluctuation problem of CCG through an operational optimization method.Firstly,a density-based affinity propagationalgorithm is proposed so that more ideal working condition categories can be obtained for the complex raw ore properties.Next,a Bayesian network(BN)is applied to explore the relationship between the operational variables and the CCG.Based on the analysis results of BN,a weighted Gaussian process regression model is constructed to predict the CCG that a higher prediction accuracy can be obtained.To ensure the predicted CCG is close to the set value with a smaller magnitude of the operation adjustments and a smaller uncertainty of the prediction results,an index-oriented adaptive differential evolution(IOADE)algorithm is proposed,and the convergence performance of IOADE is superior to the traditional differential evolution and adaptive differential evolution methods.Finally,the effectiveness and feasibility of the proposed methods are verified by the experiments on a copper flotation industrial process.
基金We acknowledge the funding support from the National Natural Science Foundation of China(Grant No.51808462)the Natural Science Foundation Project of Sichuan Province,China(Grant No.2023NSFSC0346)the Science and Technology Project of Inner Mongolia Transportation Department,China(Grant No.NJ-2022-14).
文摘Reliable long-term settlement prediction of a high embankment relates to mountain infrastructure safety.This study developed a novel hybrid model(NHM)that combines a joint denoising technique with an enhanced gray wolf optimizer(EGWO)-n-support vector regression(n-SVR)method.High-embankment field measurements were preprocessed using the joint denoising technique,which in-cludes complete ensemble empirical mode decomposition,singular value decomposition,and wavelet packet transform.Furthermore,high-embankment settlements were predicted using the EGWO-n-SVR method.In this method,the standard gray wolf optimizer(GWO)was improved to obtain the EGWO to better tune the n-SVR model hyperparameters.The proposed NHM was then tested in two case studies.Finally,the influences of the data division ratio and kernel function on the EGWO-n-SVR forecasting performance and prediction efficiency were investigated.The results indicate that the NHM suppresses noise and restores details in high-embankment field measurements.Simultaneously,the NHM out-performs other alternative prediction methods in prediction accuracy and robustness.This demonstrates that the proposed NHM is effective in predicting high-embankment settlements with noisy field mea-surements.Moreover,the appropriate data division ratio and kernel function for EGWO-n-SVR are 7:3 and radial basis function,respectively.
基金financially supported by the NationalNatural Science Foundation of China(Grant No.42072309)the Fundamental Research Funds for National University,China University of Geosciences(Wuhan)(Grant No.CUGDCJJ202217)+1 种基金the Knowledge Innovation Program of Wuhan-Basic Research(Grant No.2022020801010199)the Hubei Key Laboratory of Blasting Engineering Foundation(HKLBEF202002).
文摘Accurately estimating blasting vibration during rock blasting is the foundation of blasting vibration management.In this study,Tuna Swarm Optimization(TSO),Whale Optimization Algorithm(WOA),and Cuckoo Search(CS)were used to optimize two hyperparameters in support vector regression(SVR).Based on these methods,three hybrid models to predict peak particle velocity(PPV)for bench blasting were developed.Eighty-eight samples were collected to establish the PPV database,eight initial blasting parameters were chosen as input parameters for the predictionmodel,and the PPV was the output parameter.As predictive performance evaluation indicators,the coefficient of determination(R2),rootmean square error(RMSE),mean absolute error(MAE),and a10-index were selected.The normalizedmutual information value is then used to evaluate the impact of various input parameters on the PPV prediction outcomes.According to the research findings,TSO,WOA,and CS can all enhance the predictive performance of the SVR model.The TSO-SVR model provides the most accurate predictions.The performances of the optimized hybrid SVR models are superior to the unoptimized traditional prediction model.The maximum charge per delay impacts the PPV prediction value the most.