Predicting molecular properties is essential for advancing for advancing drug discovery and design. Recently, Graph Neural Networks (GNNs) have gained prominence due to their ability to capture the complex structural ...Predicting molecular properties is essential for advancing for advancing drug discovery and design. Recently, Graph Neural Networks (GNNs) have gained prominence due to their ability to capture the complex structural and relational information inherent in molecular graphs. Despite their effectiveness, the “black-box” nature of GNNs remains a significant obstacle to their widespread adoption in chemistry, as it hinders interpretability and trust. In this context, several explanation methods based on factual reasoning have emerged. These methods aim to interpret the predictions made by GNNs by analyzing the key features contributing to the prediction. However, these approaches fail to answer critical questions: “How to ensure that the structure-property mapping learned by GNNs is consistent with established domain knowledge”. In this paper, we propose MMGCF, a novel counterfactual explanation framework designed specifically for the prediction of GNN-based molecular properties. MMGCF constructs a hierarchical tree structure on molecular motifs, enabling the systematic generation of counterfactuals through motif perturbations. This framework identifies causally significant motifs and elucidates their impact on model predictions, offering insights into the relationship between structural modifications and predicted properties. Our method demonstrates its effectiveness through comprehensive quantitative and qualitative evaluations of four real-world molecular datasets.展开更多
Collaborative Filtering(CF) is a leading approach to build recommender systems which has gained considerable development and popularity. A predominant approach to CF is rating prediction recommender algorithm, aiming ...Collaborative Filtering(CF) is a leading approach to build recommender systems which has gained considerable development and popularity. A predominant approach to CF is rating prediction recommender algorithm, aiming to predict a user's rating for those items which were not rated yet by the user. However, with the increasing number of items and users, thedata is sparse.It is difficult to detectlatent closely relation among the items or users for predicting the user behaviors. In this paper,we enhance the rating prediction approach leading to substantial improvement of prediction accuracy by categorizing according to the genres of movies. Then the probabilities that users are interested in the genres are computed to integrate the prediction of each genre cluster. A novel probabilistic approach based on the sentiment analysis of the user reviews is also proposed to give intuitional explanations of why an item is recommended.To test the novel recommendation approach, a new corpus of user reviews on movies obtained from the Internet Movies Database(IMDB) has been generated. Experimental results show that the proposed framework is effective and achieves a better prediction performance.展开更多
Sinter is the core raw material for blast furnaces.Flue pressure,which is an important state parameter,affects sinter quality.In this paper,flue pressure prediction and optimization were studied based on the shapley a...Sinter is the core raw material for blast furnaces.Flue pressure,which is an important state parameter,affects sinter quality.In this paper,flue pressure prediction and optimization were studied based on the shapley additive explanation(SHAP)to predict the flue pressure and take targeted adjustment measures.First,the sintering process data were collected and processed.A flue pressure prediction model was then constructed after comparing different feature selection methods and model algorithms using SHAP+extremely random-ized trees(ET).The prediction accuracy of the model within the error range of±0.25 kPa was 92.63%.SHAP analysis was employed to improve the interpretability of the prediction model.The effects of various sintering operation parameters on flue pressure,the relation-ship between the numerical range of key operation parameters and flue pressure,the effect of operation parameter combinations on flue pressure,and the prediction process of the flue pressure prediction model on a single sample were analyzed.A flue pressure optimization module was also constructed and analyzed when the prediction satisfied the judgment conditions.The operating parameter combination was then pushed.The flue pressure was increased by 5.87%during the verification process,achieving a good optimization effect.展开更多
BACKGROUND Colorectal polyps are precancerous diseases of colorectal cancer.Early detection and resection of colorectal polyps can effectively reduce the mortality of colorectal cancer.Endoscopic mucosal resection(EMR...BACKGROUND Colorectal polyps are precancerous diseases of colorectal cancer.Early detection and resection of colorectal polyps can effectively reduce the mortality of colorectal cancer.Endoscopic mucosal resection(EMR)is a common polypectomy proce-dure in clinical practice,but it has a high postoperative recurrence rate.Currently,there is no predictive model for the recurrence of colorectal polyps after EMR.AIM To construct and validate a machine learning(ML)model for predicting the risk of colorectal polyp recurrence one year after EMR.METHODS This study retrospectively collected data from 1694 patients at three medical centers in Xuzhou.Additionally,a total of 166 patients were collected to form a prospective validation set.Feature variable screening was conducted using uni-variate and multivariate logistic regression analyses,and five ML algorithms were used to construct the predictive models.The optimal models were evaluated based on different performance metrics.Decision curve analysis(DCA)and SHapley Additive exPlanation(SHAP)analysis were performed to assess clinical applicability and predictor importance.RESULTS Multivariate logistic regression analysis identified 8 independent risk factors for colorectal polyp recurrence one year after EMR(P<0.05).Among the models,eXtreme Gradient Boosting(XGBoost)demonstrated the highest area under the curve(AUC)in the training set,internal validation set,and prospective validation set,with AUCs of 0.909(95%CI:0.89-0.92),0.921(95%CI:0.90-0.94),and 0.963(95%CI:0.94-0.99),respectively.DCA indicated favorable clinical utility for the XGBoost model.SHAP analysis identified smoking history,family history,and age as the top three most important predictors in the model.CONCLUSION The XGBoost model has the best predictive performance and can assist clinicians in providing individualized colonoscopy follow-up recommendations.展开更多
BACKGROUND Severe dengue children with critical complications have been attributed to high mortality rates,varying from approximately 1%to over 20%.To date,there is a lack of data on machine-learning-based algorithms ...BACKGROUND Severe dengue children with critical complications have been attributed to high mortality rates,varying from approximately 1%to over 20%.To date,there is a lack of data on machine-learning-based algorithms for predicting the risk of inhospital mortality in children with dengue shock syndrome(DSS).AIM To develop machine-learning models to estimate the risk of death in hospitalized children with DSS.METHODS This single-center retrospective study was conducted at tertiary Children’s Hospital No.2 in Viet Nam,between 2013 and 2022.The primary outcome was the in-hospital mortality rate in children with DSS admitted to the pediatric intensive care unit(PICU).Nine significant features were predetermined for further analysis using machine learning models.An oversampling method was used to enhance the model performance.Supervised models,including logistic regression,Naïve Bayes,Random Forest(RF),K-nearest neighbors,Decision Tree and Extreme Gradient Boosting(XGBoost),were employed to develop predictive models.The Shapley Additive Explanation was used to determine the degree of contribution of the features.RESULTS In total,1278 PICU-admitted children with complete data were included in the analysis.The median patient age was 8.1 years(interquartile range:5.4-10.7).Thirty-nine patients(3%)died.The RF and XGboost models demonstrated the highest performance.The Shapley Addictive Explanations model revealed that the most important predictive features included younger age,female patients,presence of underlying diseases,severe transaminitis,severe bleeding,low platelet counts requiring platelet transfusion,elevated levels of international normalized ratio,blood lactate and serum creatinine,large volume of resuscitation fluid and a high vasoactive inotropic score(>30).CONCLUSION We developed robust machine learning-based models to estimate the risk of death in hospitalized children with DSS.The study findings are applicable to the design of management schemes to enhance survival outcomes of patients with DSS.展开更多
Accurate prediction of shield tunneling-induced settlement is a complex problem that requires consideration of many influential parameters.Recent studies reveal that machine learning(ML)algorithms can predict the sett...Accurate prediction of shield tunneling-induced settlement is a complex problem that requires consideration of many influential parameters.Recent studies reveal that machine learning(ML)algorithms can predict the settlement caused by tunneling.However,well-performing ML models are usually less interpretable.Irrelevant input features decrease the performance and interpretability of an ML model.Nonetheless,feature selection,a critical step in the ML pipeline,is usually ignored in most studies that focused on predicting tunneling-induced settlement.This study applies four techniques,i.e.Pearson correlation method,sequential forward selection(SFS),sequential backward selection(SBS)and Boruta algorithm,to investigate the effect of feature selection on the model’s performance when predicting the tunneling-induced maximum surface settlement(S_(max)).The data set used in this study was compiled from two metro tunnel projects excavated in Hangzhou,China using earth pressure balance(EPB)shields and consists of 14 input features and a single output(i.e.S_(max)).The ML model that is trained on features selected from the Boruta algorithm demonstrates the best performance in both the training and testing phases.The relevant features chosen from the Boruta algorithm further indicate that tunneling-induced settlement is affected by parameters related to tunnel geometry,geological conditions and shield operation.The recently proposed Shapley additive explanations(SHAP)method explores how the input features contribute to the output of a complex ML model.It is observed that the larger settlements are induced during shield tunneling in silty clay.Moreover,the SHAP analysis reveals that the low magnitudes of face pressure at the top of the shield increase the model’s output。展开更多
Two kinds of iterative methods are designed to solve the linear system of equations, we obtain a new interpretation in terms of a geometric concept. Therefore, we have a better insight into the essence of the iterativ...Two kinds of iterative methods are designed to solve the linear system of equations, we obtain a new interpretation in terms of a geometric concept. Therefore, we have a better insight into the essence of the iterative methods and provide a reference for further study and design. Finally, a new iterative method is designed named as the diverse relaxation parameter of the SOR method which, in particular, demonstrates the geometric characteristics. Many examples prove that the method is quite effective.展开更多
Recently,convolutional neural network(CNN)-based visual inspec-tion has been developed to detect defects on building surfaces automatically.The CNN model demonstrates remarkable accuracy in image data analysis;however...Recently,convolutional neural network(CNN)-based visual inspec-tion has been developed to detect defects on building surfaces automatically.The CNN model demonstrates remarkable accuracy in image data analysis;however,the predicted results have uncertainty in providing accurate informa-tion to users because of the“black box”problem in the deep learning model.Therefore,this study proposes a visual explanation method to overcome the uncertainty limitation of CNN-based defect identification.The visual repre-sentative gradient-weights class activation mapping(Grad-CAM)method is adopted to provide visually explainable information.A visualizing evaluation index is proposed to quantitatively analyze visual representations;this index reflects a rough estimate of the concordance rate between the visualized heat map and intended defects.In addition,an ablation study,adopting three-branch combinations with the VGG16,is implemented to identify perfor-mance variations by visualizing predicted results.Experiments reveal that the proposed model,combined with hybrid pooling,batch normalization,and multi-attention modules,achieves the best performance with an accuracy of 97.77%,corresponding to an improvement of 2.49%compared with the baseline model.Consequently,this study demonstrates that reliable results from an automatic defect classification model can be provided to an inspector through the visual representation of the predicted results using CNN models.展开更多
This paper puts forward a new framework of multi-agent based equipment bidding system on the Web (MAEBSW). The purpose of the paper is to make the process of enterprise equipment bidding, especially the decision proce...This paper puts forward a new framework of multi-agent based equipment bidding system on the Web (MAEBSW). The purpose of the paper is to make the process of enterprise equipment bidding, especially the decision process of enterprise equipment bidding, more intelligent. The system discussed in this paper is an efficient method for the enterprises doing business electronically. We also present our opinion of the approaches to defining agents. We emphasize the importance of agent being able to provide explanation about its behavior for the user or other agents. Based on this opinion we design a new common architecture of most agents in MAEBSW.展开更多
The Kirk test has good precision for measuring stray light in optical lithography and is the usual method of measuring stray light.However,Kirk did not provide a theoretical explanation to his simulation model.We atte...The Kirk test has good precision for measuring stray light in optical lithography and is the usual method of measuring stray light.However,Kirk did not provide a theoretical explanation to his simulation model.We attempt to give Kirk's model a kind of theoretical explanation and a little improvement based on the model of point spread function of scattering and the theory of statistical optics.It is indicated by simulation that the improved model fits Kirk's measurement data better.展开更多
There is a puzzling astrophysical result concerning the latest observation of the absorption profile of the redshifted radio line 21 cm from the early Universe(as described in Bowman et al.). The amplitude of the prof...There is a puzzling astrophysical result concerning the latest observation of the absorption profile of the redshifted radio line 21 cm from the early Universe(as described in Bowman et al.). The amplitude of the profile was more than a factor of two greater than the largest predictions. This could mean that the primordial hydrogen gas was much cooler than expected. Some explanations in the literature suggested a possible cooling of baryons either by unspecified dark matter particles or by some exotic dark matter particles with a charge a million times smaller than the electron charge. Other explanations required an additional radio background. In the present paper, we entertain a possible different explanation for the above puzzling observational result: the explanation is based on the alternative kind of hydrogen atoms(AKHA),whose existence was previously demonstrated theoretically, as well as by the analysis of atomic experiments. Namely, the AKHA are expected to decouple from the cosmic microwave background(CMB) much earlier(in the course of the Universe expansion) than usual hydrogen atoms, so that the AKHA temperature is significantly lower than that of usual hydrogen atoms. This seems to lower the excitation(spin) temperature of the hyperfine doublet(responsible for the 21 cm line) sufficiently enough for explaining the above puzzling observational result. This possible explanation appears to be more specific and natural than the previous possible explanations. Further observational studies of the redshifted 21 cm radio line from the early Universe could help to verify which explanation is the most relevant.展开更多
[Objective] The research aimed to analyze explanation effect of the European numerical prediction on temperature. [Method] Based on CMSVM regression method, by using 850 hPa grid point data of the European numerical p...[Objective] The research aimed to analyze explanation effect of the European numerical prediction on temperature. [Method] Based on CMSVM regression method, by using 850 hPa grid point data of the European numerical prediction from 2003 to 2009 and actual data of the maximum and minimum temperatures at 8 automatic stations in Qingyang City, prediction model of the temperature was established, and running effect of the business from 2008 to 2010 was tested and evaluated. [Result] The method had very good guidance role in real-time business running of the temperature prediction. Test and evaluation found that as forecast time prolonged, prediction accuracies of the maximum and minimum temperatures declined. When temperature anomaly was higher (actual temperature was higher than historical mean), prediction accuracy increased. Influence of the European numerical prediction was bigger. [Conclusion] Compared with other methods, operation of the prediction method was convenient, modeling was automatic, running time was short, system was stable, and prediction accuracy was high. It was suitable for implementing of the explanation work for numerical prediction product at meteorological station.展开更多
With large-scale engineering projects being carried out in China, a large number of fossil localities have been discovered and excavated by responsible agencies, but still some important fossils of great value have be...With large-scale engineering projects being carried out in China, a large number of fossil localities have been discovered and excavated by responsible agencies, but still some important fossils of great value have been removed and smuggled into foreign countries. In the last three years, more than 1345 fossil specimens have been intercepted by Customs in Shenzhen, Shanghai, Tianjin, Beijing and elsewhere, and more than 5000 fossils, most of which are listed as key fossils,展开更多
Majorana zero modes in the hybrid semiconductor-superconductornanowire is one of the promising candidates for topologicalquantum computing. Recently, in nanowires with a superconductingisland, the signature of Majoran...Majorana zero modes in the hybrid semiconductor-superconductornanowire is one of the promising candidates for topologicalquantum computing. Recently, in nanowires with a superconductingisland, the signature of Majorana zero modescan be revealed as a subgap state whose energy oscillatesaround zero in magnetic field. This oscillation was interpretedas overlapping Majoranas. However, the oscillation amplitudeeither dies away after an overshoot or decays, sharply oppositeto the theoretically predicted enhanced oscillations for Majoranabound states, as the magnetic field increases. Several theoreticalstudies have tried to address this discrepancy, but arepartially successful. This discrepancy has raised the concernson the conclusive identification of Majorana bound states, andhas even endangered the scheme of Majorana qubits basedon the nanowires.展开更多
The flow regimes of GLCC with horizon inlet and a vertical pipe are investigated in experiments,and the velocities and pressure drops data labeled by the corresponding flow regimes are collected.Combined with the flow...The flow regimes of GLCC with horizon inlet and a vertical pipe are investigated in experiments,and the velocities and pressure drops data labeled by the corresponding flow regimes are collected.Combined with the flow regimes data of other GLCC positions from other literatures in existence,the gas and liquid superficial velocities and pressure drops are used as the input of the machine learning algorithms respectively which are applied to identify the flow regimes.The choosing of input data types takes the availability of data for practical industry fields into consideration,and the twelve machine learning algorithms are chosen from the classical and popular algorithms in the area of classification,including the typical ensemble models,SVM,KNN,Bayesian Model and MLP.The results of flow regimes identification show that gas and liquid superficial velocities are the ideal type of input data for the flow regimes identification by machine learning.Most of the ensemble models can identify the flow regimes of GLCC by gas and liquid velocities with the accuracy of 0.99 and more.For the pressure drops as the input of each algorithm,it is not the suitable as gas and liquid velocities,and only XGBoost and Bagging Tree can identify the GLCC flow regimes accurately.The success and confusion of each algorithm are analyzed and explained based on the experimental phenomena of flow regimes evolution processes,the flow regimes map,and the principles of algorithms.The applicability and feasibility of each algorithm according to different types of data for GLCC flow regimes identification are proposed.展开更多
Existing explanation methods for Convolutional Neural Networks(CNNs)lack the pixel-level visualization explanations to generate the reliable fine-grained decision features.Since there are inconsistencies between the e...Existing explanation methods for Convolutional Neural Networks(CNNs)lack the pixel-level visualization explanations to generate the reliable fine-grained decision features.Since there are inconsistencies between the explanation and the actual behavior of the model to be interpreted,we propose a Fine-Grained Visual Explanation for CNN,namely F-GVE,which produces a fine-grained explanation with higher consistency to the decision of the original model.The exact backward class-specific gradients with respect to the input image is obtained to highlight the object-related pixels the model used to make prediction.In addition,for better visualization and less noise,F-GVE selects an appropriate threshold to filter the gradient during the calculation and the explanation map is obtained by element-wise multiplying the gradient and the input image to show fine-grained classification decision features.Experimental results demonstrate that F-GVE has good visual performances and highlights the importance of fine-grained decision features.Moreover,the faithfulness of the explanation in this paper is high and it is effective and practical on troubleshooting and debugging detection.展开更多
DyTiFe_(11) compound is a ferromagnetic substance.It has tetragonal body-centered ThMn_(12)-type crystallographic structure.At room temperature,the easy magnetization direction is the c-axis.A spin reorientation begin...DyTiFe_(11) compound is a ferromagnetic substance.It has tetragonal body-centered ThMn_(12)-type crystallographic structure.At room temperature,the easy magnetization direction is the c-axis.A spin reorientation begins to appear at about 175K.The contribution of Fe sublattice to magnetocrystalline anisotropy was determined by experiments and that of Dy sublattice was obtained by using single ion model calculation.Results show that the spin reorientation arises from the competition of anisotropy between Fe and Dy sublattices.展开更多
文摘Predicting molecular properties is essential for advancing for advancing drug discovery and design. Recently, Graph Neural Networks (GNNs) have gained prominence due to their ability to capture the complex structural and relational information inherent in molecular graphs. Despite their effectiveness, the “black-box” nature of GNNs remains a significant obstacle to their widespread adoption in chemistry, as it hinders interpretability and trust. In this context, several explanation methods based on factual reasoning have emerged. These methods aim to interpret the predictions made by GNNs by analyzing the key features contributing to the prediction. However, these approaches fail to answer critical questions: “How to ensure that the structure-property mapping learned by GNNs is consistent with established domain knowledge”. In this paper, we propose MMGCF, a novel counterfactual explanation framework designed specifically for the prediction of GNN-based molecular properties. MMGCF constructs a hierarchical tree structure on molecular motifs, enabling the systematic generation of counterfactuals through motif perturbations. This framework identifies causally significant motifs and elucidates their impact on model predictions, offering insights into the relationship between structural modifications and predicted properties. Our method demonstrates its effectiveness through comprehensive quantitative and qualitative evaluations of four real-world molecular datasets.
基金supported in part by National Science Foundation of China under Grants No.61303105 and 61402304the Humanity&Social Science general project of Ministry of Education under Grants No.14YJAZH046+2 种基金the Beijing Natural Science Foundation under Grants No.4154065the Beijing Educational Committee Science and Technology Development Planned under Grants No.KM201410028017Academic Degree Graduate Courses group projects
文摘Collaborative Filtering(CF) is a leading approach to build recommender systems which has gained considerable development and popularity. A predominant approach to CF is rating prediction recommender algorithm, aiming to predict a user's rating for those items which were not rated yet by the user. However, with the increasing number of items and users, thedata is sparse.It is difficult to detectlatent closely relation among the items or users for predicting the user behaviors. In this paper,we enhance the rating prediction approach leading to substantial improvement of prediction accuracy by categorizing according to the genres of movies. Then the probabilities that users are interested in the genres are computed to integrate the prediction of each genre cluster. A novel probabilistic approach based on the sentiment analysis of the user reviews is also proposed to give intuitional explanations of why an item is recommended.To test the novel recommendation approach, a new corpus of user reviews on movies obtained from the Internet Movies Database(IMDB) has been generated. Experimental results show that the proposed framework is effective and achieves a better prediction performance.
基金supported by the General Program of the National Natural Science Foundation of China(No.52274326)the China Baowu Low Carbon Metallurgy Innovation Foundation(No.BWLCF202109)the Seventh Batch of Ten Thousand Talents Plan of China(No.ZX20220553).
文摘Sinter is the core raw material for blast furnaces.Flue pressure,which is an important state parameter,affects sinter quality.In this paper,flue pressure prediction and optimization were studied based on the shapley additive explanation(SHAP)to predict the flue pressure and take targeted adjustment measures.First,the sintering process data were collected and processed.A flue pressure prediction model was then constructed after comparing different feature selection methods and model algorithms using SHAP+extremely random-ized trees(ET).The prediction accuracy of the model within the error range of±0.25 kPa was 92.63%.SHAP analysis was employed to improve the interpretability of the prediction model.The effects of various sintering operation parameters on flue pressure,the relation-ship between the numerical range of key operation parameters and flue pressure,the effect of operation parameter combinations on flue pressure,and the prediction process of the flue pressure prediction model on a single sample were analyzed.A flue pressure optimization module was also constructed and analyzed when the prediction satisfied the judgment conditions.The operating parameter combination was then pushed.The flue pressure was increased by 5.87%during the verification process,achieving a good optimization effect.
文摘BACKGROUND Colorectal polyps are precancerous diseases of colorectal cancer.Early detection and resection of colorectal polyps can effectively reduce the mortality of colorectal cancer.Endoscopic mucosal resection(EMR)is a common polypectomy proce-dure in clinical practice,but it has a high postoperative recurrence rate.Currently,there is no predictive model for the recurrence of colorectal polyps after EMR.AIM To construct and validate a machine learning(ML)model for predicting the risk of colorectal polyp recurrence one year after EMR.METHODS This study retrospectively collected data from 1694 patients at three medical centers in Xuzhou.Additionally,a total of 166 patients were collected to form a prospective validation set.Feature variable screening was conducted using uni-variate and multivariate logistic regression analyses,and five ML algorithms were used to construct the predictive models.The optimal models were evaluated based on different performance metrics.Decision curve analysis(DCA)and SHapley Additive exPlanation(SHAP)analysis were performed to assess clinical applicability and predictor importance.RESULTS Multivariate logistic regression analysis identified 8 independent risk factors for colorectal polyp recurrence one year after EMR(P<0.05).Among the models,eXtreme Gradient Boosting(XGBoost)demonstrated the highest area under the curve(AUC)in the training set,internal validation set,and prospective validation set,with AUCs of 0.909(95%CI:0.89-0.92),0.921(95%CI:0.90-0.94),and 0.963(95%CI:0.94-0.99),respectively.DCA indicated favorable clinical utility for the XGBoost model.SHAP analysis identified smoking history,family history,and age as the top three most important predictors in the model.CONCLUSION The XGBoost model has the best predictive performance and can assist clinicians in providing individualized colonoscopy follow-up recommendations.
文摘BACKGROUND Severe dengue children with critical complications have been attributed to high mortality rates,varying from approximately 1%to over 20%.To date,there is a lack of data on machine-learning-based algorithms for predicting the risk of inhospital mortality in children with dengue shock syndrome(DSS).AIM To develop machine-learning models to estimate the risk of death in hospitalized children with DSS.METHODS This single-center retrospective study was conducted at tertiary Children’s Hospital No.2 in Viet Nam,between 2013 and 2022.The primary outcome was the in-hospital mortality rate in children with DSS admitted to the pediatric intensive care unit(PICU).Nine significant features were predetermined for further analysis using machine learning models.An oversampling method was used to enhance the model performance.Supervised models,including logistic regression,Naïve Bayes,Random Forest(RF),K-nearest neighbors,Decision Tree and Extreme Gradient Boosting(XGBoost),were employed to develop predictive models.The Shapley Additive Explanation was used to determine the degree of contribution of the features.RESULTS In total,1278 PICU-admitted children with complete data were included in the analysis.The median patient age was 8.1 years(interquartile range:5.4-10.7).Thirty-nine patients(3%)died.The RF and XGboost models demonstrated the highest performance.The Shapley Addictive Explanations model revealed that the most important predictive features included younger age,female patients,presence of underlying diseases,severe transaminitis,severe bleeding,low platelet counts requiring platelet transfusion,elevated levels of international normalized ratio,blood lactate and serum creatinine,large volume of resuscitation fluid and a high vasoactive inotropic score(>30).CONCLUSION We developed robust machine learning-based models to estimate the risk of death in hospitalized children with DSS.The study findings are applicable to the design of management schemes to enhance survival outcomes of patients with DSS.
基金support provided by The Science and Technology Development Fund,Macao SAR,China(File Nos.0057/2020/AGJ and SKL-IOTSC-2021-2023)Science and Technology Program of Guangdong Province,China(Grant No.2021A0505080009).
文摘Accurate prediction of shield tunneling-induced settlement is a complex problem that requires consideration of many influential parameters.Recent studies reveal that machine learning(ML)algorithms can predict the settlement caused by tunneling.However,well-performing ML models are usually less interpretable.Irrelevant input features decrease the performance and interpretability of an ML model.Nonetheless,feature selection,a critical step in the ML pipeline,is usually ignored in most studies that focused on predicting tunneling-induced settlement.This study applies four techniques,i.e.Pearson correlation method,sequential forward selection(SFS),sequential backward selection(SBS)and Boruta algorithm,to investigate the effect of feature selection on the model’s performance when predicting the tunneling-induced maximum surface settlement(S_(max)).The data set used in this study was compiled from two metro tunnel projects excavated in Hangzhou,China using earth pressure balance(EPB)shields and consists of 14 input features and a single output(i.e.S_(max)).The ML model that is trained on features selected from the Boruta algorithm demonstrates the best performance in both the training and testing phases.The relevant features chosen from the Boruta algorithm further indicate that tunneling-induced settlement is affected by parameters related to tunnel geometry,geological conditions and shield operation.The recently proposed Shapley additive explanations(SHAP)method explores how the input features contribute to the output of a complex ML model.It is observed that the larger settlements are induced during shield tunneling in silty clay.Moreover,the SHAP analysis reveals that the low magnitudes of face pressure at the top of the shield increase the model’s output。
基金Supported by the National Natural Science Foundation of China(61272300)
文摘Two kinds of iterative methods are designed to solve the linear system of equations, we obtain a new interpretation in terms of a geometric concept. Therefore, we have a better insight into the essence of the iterative methods and provide a reference for further study and design. Finally, a new iterative method is designed named as the diverse relaxation parameter of the SOR method which, in particular, demonstrates the geometric characteristics. Many examples prove that the method is quite effective.
基金supported by a Korea Agency for Infrastructure Technology Advancement(KAIA)grant funded by the Ministry of Land,Infrastructure,and Transport(Grant 22CTAP-C163951-02).
文摘Recently,convolutional neural network(CNN)-based visual inspec-tion has been developed to detect defects on building surfaces automatically.The CNN model demonstrates remarkable accuracy in image data analysis;however,the predicted results have uncertainty in providing accurate informa-tion to users because of the“black box”problem in the deep learning model.Therefore,this study proposes a visual explanation method to overcome the uncertainty limitation of CNN-based defect identification.The visual repre-sentative gradient-weights class activation mapping(Grad-CAM)method is adopted to provide visually explainable information.A visualizing evaluation index is proposed to quantitatively analyze visual representations;this index reflects a rough estimate of the concordance rate between the visualized heat map and intended defects.In addition,an ablation study,adopting three-branch combinations with the VGG16,is implemented to identify perfor-mance variations by visualizing predicted results.Experiments reveal that the proposed model,combined with hybrid pooling,batch normalization,and multi-attention modules,achieves the best performance with an accuracy of 97.77%,corresponding to an improvement of 2.49%compared with the baseline model.Consequently,this study demonstrates that reliable results from an automatic defect classification model can be provided to an inspector through the visual representation of the predicted results using CNN models.
文摘This paper puts forward a new framework of multi-agent based equipment bidding system on the Web (MAEBSW). The purpose of the paper is to make the process of enterprise equipment bidding, especially the decision process of enterprise equipment bidding, more intelligent. The system discussed in this paper is an efficient method for the enterprises doing business electronically. We also present our opinion of the approaches to defining agents. We emphasize the importance of agent being able to provide explanation about its behavior for the user or other agents. Based on this opinion we design a new common architecture of most agents in MAEBSW.
基金by the National Basic Research Program of China under Grant No 2007AA01Z333the National Special Program of China under Grant No 2009ZX02204-008.
文摘The Kirk test has good precision for measuring stray light in optical lithography and is the usual method of measuring stray light.However,Kirk did not provide a theoretical explanation to his simulation model.We attempt to give Kirk's model a kind of theoretical explanation and a little improvement based on the model of point spread function of scattering and the theory of statistical optics.It is indicated by simulation that the improved model fits Kirk's measurement data better.
文摘There is a puzzling astrophysical result concerning the latest observation of the absorption profile of the redshifted radio line 21 cm from the early Universe(as described in Bowman et al.). The amplitude of the profile was more than a factor of two greater than the largest predictions. This could mean that the primordial hydrogen gas was much cooler than expected. Some explanations in the literature suggested a possible cooling of baryons either by unspecified dark matter particles or by some exotic dark matter particles with a charge a million times smaller than the electron charge. Other explanations required an additional radio background. In the present paper, we entertain a possible different explanation for the above puzzling observational result: the explanation is based on the alternative kind of hydrogen atoms(AKHA),whose existence was previously demonstrated theoretically, as well as by the analysis of atomic experiments. Namely, the AKHA are expected to decouple from the cosmic microwave background(CMB) much earlier(in the course of the Universe expansion) than usual hydrogen atoms, so that the AKHA temperature is significantly lower than that of usual hydrogen atoms. This seems to lower the excitation(spin) temperature of the hyperfine doublet(responsible for the 21 cm line) sufficiently enough for explaining the above puzzling observational result. This possible explanation appears to be more specific and natural than the previous possible explanations. Further observational studies of the redshifted 21 cm radio line from the early Universe could help to verify which explanation is the most relevant.
文摘[Objective] The research aimed to analyze explanation effect of the European numerical prediction on temperature. [Method] Based on CMSVM regression method, by using 850 hPa grid point data of the European numerical prediction from 2003 to 2009 and actual data of the maximum and minimum temperatures at 8 automatic stations in Qingyang City, prediction model of the temperature was established, and running effect of the business from 2008 to 2010 was tested and evaluated. [Result] The method had very good guidance role in real-time business running of the temperature prediction. Test and evaluation found that as forecast time prolonged, prediction accuracies of the maximum and minimum temperatures declined. When temperature anomaly was higher (actual temperature was higher than historical mean), prediction accuracy increased. Influence of the European numerical prediction was bigger. [Conclusion] Compared with other methods, operation of the prediction method was convenient, modeling was automatic, running time was short, system was stable, and prediction accuracy was high. It was suitable for implementing of the explanation work for numerical prediction product at meteorological station.
文摘With large-scale engineering projects being carried out in China, a large number of fossil localities have been discovered and excavated by responsible agencies, but still some important fossils of great value have been removed and smuggled into foreign countries. In the last three years, more than 1345 fossil specimens have been intercepted by Customs in Shenzhen, Shanghai, Tianjin, Beijing and elsewhere, and more than 5000 fossils, most of which are listed as key fossils,
文摘Majorana zero modes in the hybrid semiconductor-superconductornanowire is one of the promising candidates for topologicalquantum computing. Recently, in nanowires with a superconductingisland, the signature of Majorana zero modescan be revealed as a subgap state whose energy oscillatesaround zero in magnetic field. This oscillation was interpretedas overlapping Majoranas. However, the oscillation amplitudeeither dies away after an overshoot or decays, sharply oppositeto the theoretically predicted enhanced oscillations for Majoranabound states, as the magnetic field increases. Several theoreticalstudies have tried to address this discrepancy, but arepartially successful. This discrepancy has raised the concernson the conclusive identification of Majorana bound states, andhas even endangered the scheme of Majorana qubits basedon the nanowires.
文摘The flow regimes of GLCC with horizon inlet and a vertical pipe are investigated in experiments,and the velocities and pressure drops data labeled by the corresponding flow regimes are collected.Combined with the flow regimes data of other GLCC positions from other literatures in existence,the gas and liquid superficial velocities and pressure drops are used as the input of the machine learning algorithms respectively which are applied to identify the flow regimes.The choosing of input data types takes the availability of data for practical industry fields into consideration,and the twelve machine learning algorithms are chosen from the classical and popular algorithms in the area of classification,including the typical ensemble models,SVM,KNN,Bayesian Model and MLP.The results of flow regimes identification show that gas and liquid superficial velocities are the ideal type of input data for the flow regimes identification by machine learning.Most of the ensemble models can identify the flow regimes of GLCC by gas and liquid velocities with the accuracy of 0.99 and more.For the pressure drops as the input of each algorithm,it is not the suitable as gas and liquid velocities,and only XGBoost and Bagging Tree can identify the GLCC flow regimes accurately.The success and confusion of each algorithm are analyzed and explained based on the experimental phenomena of flow regimes evolution processes,the flow regimes map,and the principles of algorithms.The applicability and feasibility of each algorithm according to different types of data for GLCC flow regimes identification are proposed.
基金This work was partially supported by Beijing Natural Science Foundation(No.4222038)by Open Research Project of the State Key Laboratory of Media Convergence and Communication(Communication University of China),by the National Key RD Program of China(No.2021YFF0307600)and by Fundamental Research Funds for the Central Universities.
文摘Existing explanation methods for Convolutional Neural Networks(CNNs)lack the pixel-level visualization explanations to generate the reliable fine-grained decision features.Since there are inconsistencies between the explanation and the actual behavior of the model to be interpreted,we propose a Fine-Grained Visual Explanation for CNN,namely F-GVE,which produces a fine-grained explanation with higher consistency to the decision of the original model.The exact backward class-specific gradients with respect to the input image is obtained to highlight the object-related pixels the model used to make prediction.In addition,for better visualization and less noise,F-GVE selects an appropriate threshold to filter the gradient during the calculation and the explanation map is obtained by element-wise multiplying the gradient and the input image to show fine-grained classification decision features.Experimental results demonstrate that F-GVE has good visual performances and highlights the importance of fine-grained decision features.Moreover,the faithfulness of the explanation in this paper is high and it is effective and practical on troubleshooting and debugging detection.
文摘DyTiFe_(11) compound is a ferromagnetic substance.It has tetragonal body-centered ThMn_(12)-type crystallographic structure.At room temperature,the easy magnetization direction is the c-axis.A spin reorientation begins to appear at about 175K.The contribution of Fe sublattice to magnetocrystalline anisotropy was determined by experiments and that of Dy sublattice was obtained by using single ion model calculation.Results show that the spin reorientation arises from the competition of anisotropy between Fe and Dy sublattices.