The purpose of this study was to establish a method able to accurately estimate the long-term exposure levels of individuals to fine particulate matter (PM2.5) in Jiujiang City (China) by constructing land use regress...The purpose of this study was to establish a method able to accurately estimate the long-term exposure levels of individuals to fine particulate matter (PM2.5) in Jiujiang City (China) by constructing land use regression (LUR) models. Subsequently, the accuracy of models was further verified. PM2.5 concentrations were continuously collected daily from seven monitoring stations for the construction of daily LUR models from September 1 to 14, 2023. The constructed models used PM2.5 concentrations as the dependent variable, while land use, elevation, population density and road length were used as the predictive variables. Subsequently, twenty volunteers were invited to participate, with their daily PM2.5 exposure estimated based on their work address and home address, allowing their average exposure levels to be calculated. Furthermore, volunteers wore portable PM2.5 detectors continuously for a 14-day period and the average measured PM2.5 level was used as a comparative standard. Results showed that the adjusted R2 values for the 14 daily models ranged from 0.85 to 0.94, with the R2 values generated from leave-one-out-cross-validation tests all greater than 0.61, indicating good prediction accuracy. No significant differences were observed in the measurement accuracy of the LUR modeling method and measurements using a portable PM2.5 detector (p > 0.05). This study aimed to develop a novel method for the accurate and convenient measurement of individual long-term PM2.5 exposure levels for epidemiological studies in urban environments comparable to that of Jiujiang city.展开更多
The abstract provided offers a succinct overview of the research paper’s focus on the significance of statistics, specifically regression analysis, across diverse fields. The emphasis on regression analysis indicates...The abstract provided offers a succinct overview of the research paper’s focus on the significance of statistics, specifically regression analysis, across diverse fields. The emphasis on regression analysis indicates its importance as a statistical method that helps researchers understand relationships between variables and make predictions based on data. The inclusion of multiple disciplines, such as health sciences, social sciences, environmental studies, economics, engineering, clinical psychology, social psychology, developmental psychology, cognitive psychology, and education highlights the interdisciplinary relevance of regression analysis. This breadth suggests that the findings and methodologies discussed in the paper may have wide applications, benefiting various sectors by enhancing the quality of research outcomes. The mention of “methodologies and data analysis techniques” indicates that the paper will likely delve into specific statistical approaches, offering a comprehensive examination of how regression analysis is applied in real-world scenarios. This nuance is essential, as it demonstrates the research’s commitment to not only presenting theoretical insights but also practical applications. Furthermore, the abstract states that regression analysis “enhances the validity of findings” and “informs data-driven decision-making.” This assertion underlines the critical role that robust statistical methods play in ensuring that research conclusions are reliable and applicable. The ability of regression analysis to provide clarity and support informed decisions makes it a valuable tool in both academic and professional settings. The abstract effectively outlines the paper’s exploration of regression analysis in various fields, underscoring its importance in enhancing research validity and facilitating informed decision-making. The interdisciplinary nature of the research broadens its appeal and emphasizes the need for rigorous statistical approaches in addressing complex issues across different domains.展开更多
Identification of the ice channel is the basic technology for developing intelligent ships in ice-covered waters,which is important to ensure the safety and economy of navigation.In the Arctic,merchant ships with low ...Identification of the ice channel is the basic technology for developing intelligent ships in ice-covered waters,which is important to ensure the safety and economy of navigation.In the Arctic,merchant ships with low ice class often navigate in channels opened up by icebreakers.Navigation in the ice channel often depends on good maneuverability skills and abundant experience from the captain to a large extent.The ship may get stuck if steered into ice fields off the channel.Under this circumstance,it is very important to study how to identify the boundary lines of ice channels with a reliable method.In this paper,a two-staged ice channel identification method is developed based on image segmentation and corner point regression.The first stage employs the image segmentation method to extract channel regions.In the second stage,an intelligent corner regression network is proposed to extract the channel boundary lines from the channel region.A non-intelligent angle-based filtering and clustering method is proposed and compared with corner point regression network.The training and evaluation of the segmentation method and corner regression network are carried out on the synthetic and real ice channel dataset.The evaluation results show that the accuracy of the method using the corner point regression network in the second stage is achieved as high as 73.33%on the synthetic ice channel dataset and 70.66%on the real ice channel dataset,and the processing speed can reach up to 14.58frames per second.展开更多
Despite the maturity of ensemble numerical weather prediction(NWP),the resulting forecasts are still,more often than not,under-dispersed.As such,forecast calibration tools have become popular.Among those tools,quantil...Despite the maturity of ensemble numerical weather prediction(NWP),the resulting forecasts are still,more often than not,under-dispersed.As such,forecast calibration tools have become popular.Among those tools,quantile regression(QR)is highly competitive in terms of both flexibility and predictive performance.Nevertheless,a long-standing problem of QR is quantile crossing,which greatly limits the interpretability of QR-calibrated forecasts.On this point,this study proposes a non-crossing quantile regression neural network(NCQRNN),for calibrating ensemble NWP forecasts into a set of reliable quantile forecasts without crossing.The overarching design principle of NCQRNN is to add on top of the conventional QRNN structure another hidden layer,which imposes a non-decreasing mapping between the combined output from nodes of the last hidden layer to the nodes of the output layer,through a triangular weight matrix with positive entries.The empirical part of the work considers a solar irradiance case study,in which four years of ensemble irradiance forecasts at seven locations,issued by the European Centre for Medium-Range Weather Forecasts,are calibrated via NCQRNN,as well as via an eclectic mix of benchmarking models,ranging from the naïve climatology to the state-of-the-art deep-learning and other non-crossing models.Formal and stringent forecast verification suggests that the forecasts post-processed via NCQRNN attain the maximum sharpness subject to calibration,amongst all competitors.Furthermore,the proposed conception to resolve quantile crossing is remarkably simple yet general,and thus has broad applicability as it can be integrated with many shallow-and deep-learning-based neural networks.展开更多
In this present time,Human Activity Recognition(HAR)has been of considerable aid in the case of health monitoring and recovery.The exploitation of machine learning with an intelligent agent in the area of health infor...In this present time,Human Activity Recognition(HAR)has been of considerable aid in the case of health monitoring and recovery.The exploitation of machine learning with an intelligent agent in the area of health informatics gathered using HAR augments the decision-making quality and significance.Although many research works conducted on Smart Healthcare Monitoring,there remain a certain number of pitfalls such as time,overhead,and falsification involved during analysis.Therefore,this paper proposes a Statistical Partial Regression and Support Vector Intelligent Agent Learning(SPR-SVIAL)for Smart Healthcare Monitoring.At first,the Statistical Partial Regression Feature Extraction model is used for data preprocessing along with the dimensionality-reduced features extraction process.Here,the input dataset the continuous beat-to-beat heart data,triaxial accelerometer data,and psychological characteristics were acquired from IoT wearable devices.To attain highly accurate Smart Healthcare Monitoring with less time,Partial Least Square helps extract the dimensionality-reduced features.After that,with these resulting features,SVIAL is proposed for Smart Healthcare Monitoring with the help of Machine Learning and Intelligent Agents to minimize both analysis falsification and overhead.Experimental evaluation is carried out for factors such as time,overhead,and false positive rate accuracy concerning several instances.The quantitatively analyzed results indicate the better performance of our proposed SPR-SVIAL method when compared with two state-of-the-art methods.展开更多
Purpose–This study aimed to facilitate a rapid evaluation of track service status and vehicle ride comfort based on car body acceleration.Consequently,a low-cost,data-driven approach was proposed for analyzing speed-...Purpose–This study aimed to facilitate a rapid evaluation of track service status and vehicle ride comfort based on car body acceleration.Consequently,a low-cost,data-driven approach was proposed for analyzing speed-related acceleration limits in metro systems.Design/methodology/approach–A portable sensing terminal was developed to realize easy and efficient detection of car body acceleration.Further,field measurements were performed on a 51.95-km metro line.Data from 272 metro sections were tested as a case study,and a quantile regression method was proposed to fit the control limits of the car body acceleration at different speeds using the measured data.Findings–First,the frequency statistics of the measured data in the speed-acceleration dimension indicated that the car body acceleration was primarily concentrated within the constant speed stage,particularly at speeds of 15.4,18.3,and 20.9 m/s.Second,resampling was performed according to the probability density distribution of car body acceleration for different speed domains to achieve data balance.Finally,combined with the traditional linear relationship between speed and acceleration,the statistical relationships between the speed and car body acceleration under different quantiles were determined.We concluded the lateral/vertical quantiles of 0.8989/0.9895,0.9942/0.997,and 0.9998/0.993 as being excellent,good,and qualified control limits,respectively,for the lateral and vertical acceleration of the car body.In addition,regression lines for the speedrelated acceleration limits at other quantiles(0.5,0.75,2s,and 3s)were obtained.Originality/value–The proposed method is expected to serve as a reference for further studies on speedrelated acceleration limits in rail transit systems.展开更多
Purpose:The purpose of this study is to develop and compare model choice strategies in context of logistic regression.Model choice means the choice of the covariates to be included in the model.Design/methodology/appr...Purpose:The purpose of this study is to develop and compare model choice strategies in context of logistic regression.Model choice means the choice of the covariates to be included in the model.Design/methodology/approach:The study is based on Monte Carlo simulations.The methods are compared in terms of three measures of accuracy:specificity and two kinds of sensitivity.A loss function combining sensitivity and specificity is introduced and used for a final comparison.Findings:The choice of method depends on how much the users emphasize sensitivity against specificity.It also depends on the sample size.For a typical logistic regression setting with a moderate sample size and a small to moderate effect size,either BIC,BICc or Lasso seems to be optimal.Research limitations:Numerical simulations cannot cover the whole range of data-generating processes occurring with real-world data.Thus,more simulations are needed.Practical implications:Researchers can refer to these results if they believe that their data-generating process is somewhat similar to some of the scenarios presented in this paper.Alternatively,they could run their own simulations and calculate the loss function.Originality/value:This is a systematic comparison of model choice algorithms and heuristics in context of logistic regression.The distinction between two types of sensitivity and a comparison based on a loss function are methodological novelties.展开更多
The performance of lithium-ion batteries(LIBs)gradually declines over time,making it critical to predict the battery’s state of health(SOH)in real-time.This paper presents a model that incorporates health indicators ...The performance of lithium-ion batteries(LIBs)gradually declines over time,making it critical to predict the battery’s state of health(SOH)in real-time.This paper presents a model that incorporates health indicators and ensemble Gaussian process regression(EGPR)to predict the SOH of LIBs.Firstly,the degradation process of an LIB is analyzed through indirect health indicators(HIs)derived from voltage and temperature during discharge.Next,the parameters in the EGPR model are optimized using the gannet optimization algorithm(GOA),and the EGPR is employed to estimate the SOH of LIBs.Finally,the proposed model is tested under various experimental scenarios and compared with other machine learning models.The effectiveness of EGPR model is demonstrated using the National Aeronautics and Space Administration(NASA)LIB.The root mean square error(RMSE)is maintained within 0.20%,and the mean absolute error(MAE)is below 0.16%,illustrating the proposed approach’s excellent predictive accuracy and wide applicability.展开更多
In oil and gas exploration,elucidating the complex interdependencies among geological variables is paramount.Our study introduces the application of sophisticated regression analysis method at the forefront,aiming not...In oil and gas exploration,elucidating the complex interdependencies among geological variables is paramount.Our study introduces the application of sophisticated regression analysis method at the forefront,aiming not just at predicting geophysical logging curve values but also innovatively mitigate hydrocarbon depletion observed in geochemical logging.Through a rigorous assessment,we explore the efficacy of eight regression models,bifurcated into linear and nonlinear groups,to accommodate the multifaceted nature of geological datasets.Our linear model suite encompasses the Standard Equation,Ridge Regression,Least Absolute Shrinkage and Selection Operator,and Elastic Net,each presenting distinct advantages.The Standard Equation serves as a foundational benchmark,whereas Ridge Regression implements penalty terms to counteract overfitting,thus bolstering model robustness in the presence of multicollinearity.The Least Absolute Shrinkage and Selection Operator for variable selection functions to streamline models,enhancing their interpretability,while Elastic Net amalgamates the merits of Ridge Regression and Least Absolute Shrinkage and Selection Operator,offering a harmonized solution to model complexity and comprehensibility.On the nonlinear front,Gradient Descent,Kernel Ridge Regression,Support Vector Regression,and Piecewise Function-Fitting methods introduce innovative approaches.Gradient Descent assures computational efficiency in optimizing solutions,Kernel Ridge Regression leverages the kernel trick to navigate nonlinear patterns,and Support Vector Regression is proficient in forecasting extremities,pivotal for exploration risk assessment.The Piecewise Function-Fitting approach,tailored for geological data,facilitates adaptable modeling of variable interrelations,accommodating abrupt data trend shifts.Our analysis identifies Ridge Regression,particularly when augmented by Piecewise Function-Fitting,as superior in recouping hydrocarbon losses,and underscoring its utility in resource quantification refinement.Meanwhile,Kernel Ridge Regression emerges as a noteworthy strategy in ameliorating porosity-logging curve prediction for well A,evidencing its aptness for intricate geological structures.This research attests to the scientific ascendancy and broad-spectrum relevance of these regression techniques over conventional methods while heralding new horizons for their deployment in the oil and gas sector.The insights garnered from these advanced modeling strategies are set to transform geological and engineering practices in hydrocarbon prediction,evaluation,and recovery.展开更多
The Extensible Markup Language(XML)files,widely used for storing and exchanging information on the web require efficient parsing mechanisms to improve the performance of the applications.With the existing Document Obj...The Extensible Markup Language(XML)files,widely used for storing and exchanging information on the web require efficient parsing mechanisms to improve the performance of the applications.With the existing Document Object Model(DOM)based parsing,the performance degrades due to sequential processing and large memory requirements,thereby requiring an efficient XML parser to mitigate these issues.In this paper,we propose a Parallel XML Tree Generator(PXTG)algorithm for accelerating the parsing of XML files and a Regression-based XML Parsing Framework(RXPF)that analyzes and predicts performance through profiling,regression,and code generation for efficient parsing.The PXTG algorithm is based on dividing the XML file into n parts and producing n trees in parallel.The profiling phase of the RXPF framework produces a dataset by measuring the performance of various parsing models including StAX,SAX,DOM,JDOM,and PXTG on different cores by using multiple file sizes.The regression phase produces the prediction model,based on which the final code for efficient parsing of XML files is produced through the code generation phase.The RXPF framework has shown a significant improvement in performance varying from 9.54%to 32.34%over other existing models used for parsing XML files.展开更多
The picking efficiency of seismic first breaks(FBs)has been greatly accelerated by deep learning(DL)technology.However,the picking accuracy and efficiency of DL methods still face huge challenges in low signal-to-nois...The picking efficiency of seismic first breaks(FBs)has been greatly accelerated by deep learning(DL)technology.However,the picking accuracy and efficiency of DL methods still face huge challenges in low signal-to-noise ratio(SNR)situations.To address this issue,we propose a regression approach to pick FBs based on bidirectional long short-term memory(Bi LSTM)neural network by learning the implicit Eikonal equation of 3D inhomogeneous media with rugged topography in the target region.We employ a regressive model that represents the relationships among the elevation of shots,offset and the elevation of receivers with their seismic traveltime to predict the unknown FBs,from common-shot gathers with sparsely distributed traces.Different from image segmentation methods which automatically extract image features and classify FBs from seismic data,the proposed method can learn the inner relationship between field geometry and FBs.In addition,the predicted results by the regressive model are continuous values of FBs rather than the discrete ones of the binary distribution.The picking results of synthetic data shows that the proposed method has low dependence on label data,and can obtain reliable and similar predicted results using two types of label data with large differences.The picking results of9380 shots for 3D seismic data generated by vibroseis indicate that the proposed method can still accurately predict FBs in low SNR data.The subsequent stacked profiles further illustrate the reliability and effectiveness of the proposed method.The results of model data and field seismic data demonstrate that the proposed regression method is a robust first-break picker with high potential for field application.展开更多
To accurately model flows with shock waves using staggered-grid Lagrangian hydrodynamics, the artificial viscosity has to be introduced to convert kinetic energy into internal energy, thereby increasing the entropy ac...To accurately model flows with shock waves using staggered-grid Lagrangian hydrodynamics, the artificial viscosity has to be introduced to convert kinetic energy into internal energy, thereby increasing the entropy across shocks. Determining the appropriate strength of the artificial viscosity is an art and strongly depends on the particular problem and experience of the researcher. The objective of this study is to pose the problem of finding the appropriate strength of the artificial viscosity as an optimization problem and solve this problem using machine learning (ML) tools, specifically using surrogate models based on Gaussian Process regression (GPR) and Bayesian analysis. We describe the optimization method and discuss various practical details of its implementation. The shock-containing problems for which we apply this method all have been implemented in the LANL code FLAG (Burton in Connectivity structures and differencing techniques for staggered-grid free-Lagrange hydrodynamics, Tech. Rep. UCRL-JC-110555, Lawrence Livermore National Laboratory, Livermore, CA, 1992, 1992, in Consistent finite-volume discretization of hydrodynamic conservation laws for unstructured grids, Tech. Rep. CRL-JC-118788, Lawrence Livermore National Laboratory, Livermore, CA, 1992, 1994, Multidimensional discretization of conservation laws for unstructured polyhedral grids, Tech. Rep. UCRL-JC-118306, Lawrence Livermore National Laboratory, Livermore, CA, 1992, 1994, in FLAG, a multi-dimensional, multiple mesh, adaptive free-Lagrange, hydrodynamics code. In: NECDC, 1992). First, we apply ML to find optimal values to isolated shock problems of different strengths. Second, we apply ML to optimize the viscosity for a one-dimensional (1D) propagating detonation problem based on Zel’dovich-von Neumann-Doring (ZND) (Fickett and Davis in Detonation: theory and experiment. Dover books on physics. Dover Publications, Mineola, 2000) detonation theory using a reactive burn model. We compare results for default (currently used values in FLAG) and optimized values of the artificial viscosity for these problems demonstrating the potential for significant improvement in the accuracy of computations.展开更多
Objective This study employs the Geographically and Temporally Weighted Regression(GTWR)model to assess the impact of meteorological elements and imported cases on dengue fever outbreaks,emphasizing the spatial-tempor...Objective This study employs the Geographically and Temporally Weighted Regression(GTWR)model to assess the impact of meteorological elements and imported cases on dengue fever outbreaks,emphasizing the spatial-temporal variability of these factors in border regions.Methods We conducted a descriptive analysis of dengue fever’s temporal-spatial distribution in Yunnan border areas.Utilizing annual data from 2013 to 2019,with each county in the Yunnan border serving as a spatial unit,we constructed a GTWR model to investigate the determinants of dengue fever and their spatio-temporal heterogeneity in this region.Results The GTWR model,proving more effective than Ordinary Least Squares(OLS)analysis,identified significant spatial and temporal heterogeneity in factors influencing dengue fever’s spread along the Yunnan border.Notably,the GTWR model revealed a substantial variation in the relationship between indigenous dengue fever incidence,meteorological variables,and imported cases across different counties.Conclusion In the Yunnan border areas,local dengue incidence is affected by temperature,humidity,precipitation,wind speed,and imported cases,with these factors’influence exhibiting notable spatial and temporal variation.展开更多
Ignimbrites have been widely used as building materials in many historical and touristic structures in the Kayseri region of Türkiye. Their diverse colours and textures make them a popular choice for modern const...Ignimbrites have been widely used as building materials in many historical and touristic structures in the Kayseri region of Türkiye. Their diverse colours and textures make them a popular choice for modern construction as well. However, ignimbrites are particularly vulnerable to atmospheric conditions, such as freeze-thaw cycles, due to their high porosity, which is a result of their formation process. When water enters the pores of the ignimbrites, it can freeze during cold weather. As the water freezes and expands, it generates internal stress within the stone, causing micro-cracks to develop. Over time, repeated freeze-thaw (F-T) cycles lead to the growth of these micro-cracks into larger cracks, compromising the structural integrity of the ignimbrites and eventually making them unsuitable for use as building materials. The determination of the long-term F-T performance of ignimbrites can be established after long F-T experimental processes. Determining the long-term F-T performance of ignimbrites typically requires extensive experimental testing over prolonged freeze-thaw cycles. To streamline this process, developing accurate predictive equations becomes crucial. In this study, such equations were formulated using classical regression analyses and artificial neural networks (ANN) based on data obtained from these experiments, allowing for the prediction of the F-T performance of ignimbrites and other similar building stones without the need for lengthy testing. In this study, uniaxial compressive strength, ultrasonic propagation velocity, apparent porosity and mass loss of ignimbrites after long-term F-T were determined. Following the F-T cycles, the disintegration rate was evaluated using decay function approaches, while uniaxial compressive strength (UCS) values were predicted with minimal input parameters through both regression and ANN analyses. The ANN and regression models created for this purpose were first started with a single input value and then developed with two and three combinations. The predictive performance of the models was assessed by comparing them to regression models using the coefficient of determination (R2) as the evaluation criterion. As a result of the study, higher R2 values (0.87) were obtained in models built with artificial neural network. The results of the study indicate that ANN usage can produce results close to experimental outcomes in predicting the long-term F-T performance of ignimbrite samples.展开更多
The extended kernel ridge regression(EKRR)method with odd-even effects was adopted to improve the description of the nuclear charge radius using five commonly used nuclear models.These are:(i)the isospin-dependent A^(...The extended kernel ridge regression(EKRR)method with odd-even effects was adopted to improve the description of the nuclear charge radius using five commonly used nuclear models.These are:(i)the isospin-dependent A^(1∕3) formula,(ii)relativistic continuum Hartree-Bogoliubov(RCHB)theory,(iii)Hartree-Fock-Bogoliubov(HFB)model HFB25,(iv)the Weizsacker-Skyrme(WS)model WS*,and(v)HFB25*model.In the last two models,the charge radii were calculated using a five-parameter formula with the nuclear shell corrections and deformations obtained from the WS and HFB25 models,respectively.For each model,the resultant root-mean-square deviation for the 1014 nuclei with proton number Z≥8 can be significantly reduced to 0.009-0.013 fm after considering the modification with the EKRR method.The best among them was the RCHB model,with a root-mean-square deviation of 0.0092 fm.The extrapolation abilities of the KRR and EKRR methods for the neutron-rich region were examined,and it was found that after considering the odd-even effects,the extrapolation power was improved compared with that of the original KRR method.The strong odd-even staggering of nuclear charge radii of Ca and Cu isotopes and the abrupt kinks across the neutron N=126 and 82 shell closures were also calculated and could be reproduced quite well by calculations using the EKRR method.展开更多
In view of the composition analysis and identification of ancient glass products, L1 regularization, K-Means cluster analysis, elbow rule and other methods were comprehensively used to build logical regression, cluste...In view of the composition analysis and identification of ancient glass products, L1 regularization, K-Means cluster analysis, elbow rule and other methods were comprehensively used to build logical regression, cluster analysis, hyper-parameter test and other models, and SPSS, Python and other tools were used to obtain the classification rules of glass products under different fluxes, sub classification under different chemical compositions, hyper-parameter K value test and rationality analysis. Research can provide theoretical support for the protection and restoration of ancient glass relics.展开更多
This article develops a procedure for screening variables, in ultra high-di- mensional settings, based on their predictive significance. This is achieved by ranking the variables according to the variance of their res...This article develops a procedure for screening variables, in ultra high-di- mensional settings, based on their predictive significance. This is achieved by ranking the variables according to the variance of their respective marginal regression functions (RV-SIS). We show that, under some mild technical conditions, the RV-SIS possesses a sure screening property, which is defined by Fan and Lv (2008). Numerical comparisons suggest that RV-SIS has competitive performance compared to other screening procedures, and outperforms them in many different model settings.展开更多
In the railway system,fasteners have the functions of damping,maintaining the track distance,and adjusting the track level.Therefore,routine maintenance and inspection of fasteners are important to ensure the safe ope...In the railway system,fasteners have the functions of damping,maintaining the track distance,and adjusting the track level.Therefore,routine maintenance and inspection of fasteners are important to ensure the safe operation of track lines.Currently,assessment methods for fastener tightness include manual observation,acoustic wave detection,and image detection.There are limitations such as low accuracy and efficiency,easy interference and misjudgment,and a lack of accurate,stable,and fast detection methods.Aiming at the small deformation characteristics and large elastic change of fasteners from full loosening to full tightening,this study proposes high-precision surface-structured light technology for fastener detection and fastener deformation feature extraction based on the center-line projection distance and a fastener tightness regression method based on neural networks.First,the method uses a 3D camera to obtain a fastener point cloud and then segments the elastic rod area based on the iterative closest point algorithm registration.Principal component analysis is used to calculate the normal vector of the segmented elastic rod surface and extract the point on the centerline of the elastic rod.The point is projected onto the upper surface of the bolt to calculate the projection distance.Subsequently,the mapping relationship between the projection distance sequence and fastener tightness is established,and the influence of each parameter on the fastener tightness prediction is analyzed.Finally,by setting up a fastener detection scene in the track experimental base,collecting data,and completing the algorithm verification,the results showed that the deviation between the fastener tightness regression value obtained after the algorithm processing and the actual measured value RMSE was 0.2196 mm,which significantly improved the effect compared with other tightness detection methods,and realized an effective fastener tightness regression.展开更多
Objective Previous studies on the association between lipid profiles and chronic kidney disease(CKD)have yielded inconsistent results and no defined thresholds for blood lipids.Methods A prospective cohort study inclu...Objective Previous studies on the association between lipid profiles and chronic kidney disease(CKD)have yielded inconsistent results and no defined thresholds for blood lipids.Methods A prospective cohort study including 32,351 subjects who completed baseline and follow-up surveys over 5 years was conducted.Restricted cubic splines and Cox models were used to examine the association between the lipid profiles and CKD.A regression discontinuity design was used to determine the cutoff value of lipid profiles that was significantly associated with increased the risk of CKD.Results Over a median follow-up time of 2.2(0.5,4.2)years,648(2.00%)subjects developed CKD.The lipid profiles that were significantly and linearly related to CKD included total cholesterol(TC),triglycerides(TG),high-density lipoprotein cholesterol(HDL-C),TC/HDL-C,and TG/HDL-C,whereas lowdensity lipoprotein cholesterol(LDL-C)and LDL-C/HDL-C were nonlinearly correlated with CKD.TC,TG,TC/HDL-C,and TG/HDL-C showed an upward jump at the cutoff value,increasing the risk of CKD by 0.90%,1.50%,2.30%,and 1.60%,respectively,whereas HDL-C showed a downward jump at the cutoff value,reducing this risk by 1.0%.Female and participants with dyslipidemia had a higher risk of CKD,while the cutoff values for the different characteristics of the population were different.Conclusion There was a significant association between lipid profiles and CKD in a prospective cohort from Northwest China,while TG,TC/HDL-C,and TG/HDL-C showed a stronger risk association.The specific cutoff values of lipid profiles may provide a clinical reference for screening or diagnosing CKD risk.展开更多
This research introduces a novel approach to improve and optimize the predictive capacity of consumer purchase behaviors on e-commerce platforms. This study presented an introduction to the fundamental concepts of the...This research introduces a novel approach to improve and optimize the predictive capacity of consumer purchase behaviors on e-commerce platforms. This study presented an introduction to the fundamental concepts of the logistic regression algorithm. In addition, it analyzed user data obtained from an e-commerce platform. The original data were preprocessed, and a consumer purchase prediction model was developed for the e-commerce platform using the logistic regression method. The comparison study used the classic random forest approach, further enhanced by including the K-fold cross-validation method. Evaluation of the accuracy of the model’s classification was conducted using performance indicators that included the accuracy rate, the precision rate, the recall rate, and the F1 score. A visual examination determined the significance of the findings. The findings suggest that employing the logistic regression algorithm to forecast customer purchase behaviors on e-commerce platforms can improve the efficacy of the approach and yield more accurate predictions. This study serves as a valuable resource for improving the precision of forecasting customers’ purchase behaviors on e-commerce platforms. It has significant practical implications for optimizing the operational efficiency of e-commerce platforms.展开更多
文摘The purpose of this study was to establish a method able to accurately estimate the long-term exposure levels of individuals to fine particulate matter (PM2.5) in Jiujiang City (China) by constructing land use regression (LUR) models. Subsequently, the accuracy of models was further verified. PM2.5 concentrations were continuously collected daily from seven monitoring stations for the construction of daily LUR models from September 1 to 14, 2023. The constructed models used PM2.5 concentrations as the dependent variable, while land use, elevation, population density and road length were used as the predictive variables. Subsequently, twenty volunteers were invited to participate, with their daily PM2.5 exposure estimated based on their work address and home address, allowing their average exposure levels to be calculated. Furthermore, volunteers wore portable PM2.5 detectors continuously for a 14-day period and the average measured PM2.5 level was used as a comparative standard. Results showed that the adjusted R2 values for the 14 daily models ranged from 0.85 to 0.94, with the R2 values generated from leave-one-out-cross-validation tests all greater than 0.61, indicating good prediction accuracy. No significant differences were observed in the measurement accuracy of the LUR modeling method and measurements using a portable PM2.5 detector (p > 0.05). This study aimed to develop a novel method for the accurate and convenient measurement of individual long-term PM2.5 exposure levels for epidemiological studies in urban environments comparable to that of Jiujiang city.
文摘The abstract provided offers a succinct overview of the research paper’s focus on the significance of statistics, specifically regression analysis, across diverse fields. The emphasis on regression analysis indicates its importance as a statistical method that helps researchers understand relationships between variables and make predictions based on data. The inclusion of multiple disciplines, such as health sciences, social sciences, environmental studies, economics, engineering, clinical psychology, social psychology, developmental psychology, cognitive psychology, and education highlights the interdisciplinary relevance of regression analysis. This breadth suggests that the findings and methodologies discussed in the paper may have wide applications, benefiting various sectors by enhancing the quality of research outcomes. The mention of “methodologies and data analysis techniques” indicates that the paper will likely delve into specific statistical approaches, offering a comprehensive examination of how regression analysis is applied in real-world scenarios. This nuance is essential, as it demonstrates the research’s commitment to not only presenting theoretical insights but also practical applications. Furthermore, the abstract states that regression analysis “enhances the validity of findings” and “informs data-driven decision-making.” This assertion underlines the critical role that robust statistical methods play in ensuring that research conclusions are reliable and applicable. The ability of regression analysis to provide clarity and support informed decisions makes it a valuable tool in both academic and professional settings. The abstract effectively outlines the paper’s exploration of regression analysis in various fields, underscoring its importance in enhancing research validity and facilitating informed decision-making. The interdisciplinary nature of the research broadens its appeal and emphasizes the need for rigorous statistical approaches in addressing complex issues across different domains.
基金financially supported by the National Key Research and Development Program(Grant No.2022YFE0107000)the General Projects of the National Natural Science Foundation of China(Grant No.52171259)the High-Tech Ship Research Project of the Ministry of Industry and Information Technology(Grant No.[2021]342)。
文摘Identification of the ice channel is the basic technology for developing intelligent ships in ice-covered waters,which is important to ensure the safety and economy of navigation.In the Arctic,merchant ships with low ice class often navigate in channels opened up by icebreakers.Navigation in the ice channel often depends on good maneuverability skills and abundant experience from the captain to a large extent.The ship may get stuck if steered into ice fields off the channel.Under this circumstance,it is very important to study how to identify the boundary lines of ice channels with a reliable method.In this paper,a two-staged ice channel identification method is developed based on image segmentation and corner point regression.The first stage employs the image segmentation method to extract channel regions.In the second stage,an intelligent corner regression network is proposed to extract the channel boundary lines from the channel region.A non-intelligent angle-based filtering and clustering method is proposed and compared with corner point regression network.The training and evaluation of the segmentation method and corner regression network are carried out on the synthetic and real ice channel dataset.The evaluation results show that the accuracy of the method using the corner point regression network in the second stage is achieved as high as 73.33%on the synthetic ice channel dataset and 70.66%on the real ice channel dataset,and the processing speed can reach up to 14.58frames per second.
基金supported by the National Natural Science Foundation of China (Project No.42375192)the China Meteorological Administration Climate Change Special Program (CMA-CCSP+1 种基金Project No.QBZ202315)support by the Vector Stiftung through the Young Investigator Group"Artificial Intelligence for Probabilistic Weather Forecasting."
文摘Despite the maturity of ensemble numerical weather prediction(NWP),the resulting forecasts are still,more often than not,under-dispersed.As such,forecast calibration tools have become popular.Among those tools,quantile regression(QR)is highly competitive in terms of both flexibility and predictive performance.Nevertheless,a long-standing problem of QR is quantile crossing,which greatly limits the interpretability of QR-calibrated forecasts.On this point,this study proposes a non-crossing quantile regression neural network(NCQRNN),for calibrating ensemble NWP forecasts into a set of reliable quantile forecasts without crossing.The overarching design principle of NCQRNN is to add on top of the conventional QRNN structure another hidden layer,which imposes a non-decreasing mapping between the combined output from nodes of the last hidden layer to the nodes of the output layer,through a triangular weight matrix with positive entries.The empirical part of the work considers a solar irradiance case study,in which four years of ensemble irradiance forecasts at seven locations,issued by the European Centre for Medium-Range Weather Forecasts,are calibrated via NCQRNN,as well as via an eclectic mix of benchmarking models,ranging from the naïve climatology to the state-of-the-art deep-learning and other non-crossing models.Formal and stringent forecast verification suggests that the forecasts post-processed via NCQRNN attain the maximum sharpness subject to calibration,amongst all competitors.Furthermore,the proposed conception to resolve quantile crossing is remarkably simple yet general,and thus has broad applicability as it can be integrated with many shallow-and deep-learning-based neural networks.
基金supported by Princess Nourah bint Abdulrahman University Researchers Supporting Project Number(PNURSP2022R194)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.
文摘In this present time,Human Activity Recognition(HAR)has been of considerable aid in the case of health monitoring and recovery.The exploitation of machine learning with an intelligent agent in the area of health informatics gathered using HAR augments the decision-making quality and significance.Although many research works conducted on Smart Healthcare Monitoring,there remain a certain number of pitfalls such as time,overhead,and falsification involved during analysis.Therefore,this paper proposes a Statistical Partial Regression and Support Vector Intelligent Agent Learning(SPR-SVIAL)for Smart Healthcare Monitoring.At first,the Statistical Partial Regression Feature Extraction model is used for data preprocessing along with the dimensionality-reduced features extraction process.Here,the input dataset the continuous beat-to-beat heart data,triaxial accelerometer data,and psychological characteristics were acquired from IoT wearable devices.To attain highly accurate Smart Healthcare Monitoring with less time,Partial Least Square helps extract the dimensionality-reduced features.After that,with these resulting features,SVIAL is proposed for Smart Healthcare Monitoring with the help of Machine Learning and Intelligent Agents to minimize both analysis falsification and overhead.Experimental evaluation is carried out for factors such as time,overhead,and false positive rate accuracy concerning several instances.The quantitatively analyzed results indicate the better performance of our proposed SPR-SVIAL method when compared with two state-of-the-art methods.
基金the National Natural Science Foundation of China(NSFC)under No.52308473the National KeyR&DProgram under No.2022YFB2603301the China Postdoctoral Science Foundation funded project(Certificate Number:2023M743895).
文摘Purpose–This study aimed to facilitate a rapid evaluation of track service status and vehicle ride comfort based on car body acceleration.Consequently,a low-cost,data-driven approach was proposed for analyzing speed-related acceleration limits in metro systems.Design/methodology/approach–A portable sensing terminal was developed to realize easy and efficient detection of car body acceleration.Further,field measurements were performed on a 51.95-km metro line.Data from 272 metro sections were tested as a case study,and a quantile regression method was proposed to fit the control limits of the car body acceleration at different speeds using the measured data.Findings–First,the frequency statistics of the measured data in the speed-acceleration dimension indicated that the car body acceleration was primarily concentrated within the constant speed stage,particularly at speeds of 15.4,18.3,and 20.9 m/s.Second,resampling was performed according to the probability density distribution of car body acceleration for different speed domains to achieve data balance.Finally,combined with the traditional linear relationship between speed and acceleration,the statistical relationships between the speed and car body acceleration under different quantiles were determined.We concluded the lateral/vertical quantiles of 0.8989/0.9895,0.9942/0.997,and 0.9998/0.993 as being excellent,good,and qualified control limits,respectively,for the lateral and vertical acceleration of the car body.In addition,regression lines for the speedrelated acceleration limits at other quantiles(0.5,0.75,2s,and 3s)were obtained.Originality/value–The proposed method is expected to serve as a reference for further studies on speedrelated acceleration limits in rail transit systems.
文摘Purpose:The purpose of this study is to develop and compare model choice strategies in context of logistic regression.Model choice means the choice of the covariates to be included in the model.Design/methodology/approach:The study is based on Monte Carlo simulations.The methods are compared in terms of three measures of accuracy:specificity and two kinds of sensitivity.A loss function combining sensitivity and specificity is introduced and used for a final comparison.Findings:The choice of method depends on how much the users emphasize sensitivity against specificity.It also depends on the sample size.For a typical logistic regression setting with a moderate sample size and a small to moderate effect size,either BIC,BICc or Lasso seems to be optimal.Research limitations:Numerical simulations cannot cover the whole range of data-generating processes occurring with real-world data.Thus,more simulations are needed.Practical implications:Researchers can refer to these results if they believe that their data-generating process is somewhat similar to some of the scenarios presented in this paper.Alternatively,they could run their own simulations and calculate the loss function.Originality/value:This is a systematic comparison of model choice algorithms and heuristics in context of logistic regression.The distinction between two types of sensitivity and a comparison based on a loss function are methodological novelties.
基金supported by Fundamental Research Program of Shanxi Province(No.202203021211088)Shanxi Provincial Natural Science Foundation(No.202204021301049).
文摘The performance of lithium-ion batteries(LIBs)gradually declines over time,making it critical to predict the battery’s state of health(SOH)in real-time.This paper presents a model that incorporates health indicators and ensemble Gaussian process regression(EGPR)to predict the SOH of LIBs.Firstly,the degradation process of an LIB is analyzed through indirect health indicators(HIs)derived from voltage and temperature during discharge.Next,the parameters in the EGPR model are optimized using the gannet optimization algorithm(GOA),and the EGPR is employed to estimate the SOH of LIBs.Finally,the proposed model is tested under various experimental scenarios and compared with other machine learning models.The effectiveness of EGPR model is demonstrated using the National Aeronautics and Space Administration(NASA)LIB.The root mean square error(RMSE)is maintained within 0.20%,and the mean absolute error(MAE)is below 0.16%,illustrating the proposed approach’s excellent predictive accuracy and wide applicability.
文摘In oil and gas exploration,elucidating the complex interdependencies among geological variables is paramount.Our study introduces the application of sophisticated regression analysis method at the forefront,aiming not just at predicting geophysical logging curve values but also innovatively mitigate hydrocarbon depletion observed in geochemical logging.Through a rigorous assessment,we explore the efficacy of eight regression models,bifurcated into linear and nonlinear groups,to accommodate the multifaceted nature of geological datasets.Our linear model suite encompasses the Standard Equation,Ridge Regression,Least Absolute Shrinkage and Selection Operator,and Elastic Net,each presenting distinct advantages.The Standard Equation serves as a foundational benchmark,whereas Ridge Regression implements penalty terms to counteract overfitting,thus bolstering model robustness in the presence of multicollinearity.The Least Absolute Shrinkage and Selection Operator for variable selection functions to streamline models,enhancing their interpretability,while Elastic Net amalgamates the merits of Ridge Regression and Least Absolute Shrinkage and Selection Operator,offering a harmonized solution to model complexity and comprehensibility.On the nonlinear front,Gradient Descent,Kernel Ridge Regression,Support Vector Regression,and Piecewise Function-Fitting methods introduce innovative approaches.Gradient Descent assures computational efficiency in optimizing solutions,Kernel Ridge Regression leverages the kernel trick to navigate nonlinear patterns,and Support Vector Regression is proficient in forecasting extremities,pivotal for exploration risk assessment.The Piecewise Function-Fitting approach,tailored for geological data,facilitates adaptable modeling of variable interrelations,accommodating abrupt data trend shifts.Our analysis identifies Ridge Regression,particularly when augmented by Piecewise Function-Fitting,as superior in recouping hydrocarbon losses,and underscoring its utility in resource quantification refinement.Meanwhile,Kernel Ridge Regression emerges as a noteworthy strategy in ameliorating porosity-logging curve prediction for well A,evidencing its aptness for intricate geological structures.This research attests to the scientific ascendancy and broad-spectrum relevance of these regression techniques over conventional methods while heralding new horizons for their deployment in the oil and gas sector.The insights garnered from these advanced modeling strategies are set to transform geological and engineering practices in hydrocarbon prediction,evaluation,and recovery.
文摘The Extensible Markup Language(XML)files,widely used for storing and exchanging information on the web require efficient parsing mechanisms to improve the performance of the applications.With the existing Document Object Model(DOM)based parsing,the performance degrades due to sequential processing and large memory requirements,thereby requiring an efficient XML parser to mitigate these issues.In this paper,we propose a Parallel XML Tree Generator(PXTG)algorithm for accelerating the parsing of XML files and a Regression-based XML Parsing Framework(RXPF)that analyzes and predicts performance through profiling,regression,and code generation for efficient parsing.The PXTG algorithm is based on dividing the XML file into n parts and producing n trees in parallel.The profiling phase of the RXPF framework produces a dataset by measuring the performance of various parsing models including StAX,SAX,DOM,JDOM,and PXTG on different cores by using multiple file sizes.The regression phase produces the prediction model,based on which the final code for efficient parsing of XML files is produced through the code generation phase.The RXPF framework has shown a significant improvement in performance varying from 9.54%to 32.34%over other existing models used for parsing XML files.
基金financially supported by the National Key R&D Program of China(2018YFA0702504)the National Natural Science Foundation of China(42174152)+1 种基金the Strategic Cooperation Technology Projects of China National Petroleum Corporation(CNPC)and China University of Petroleum-Beijing(CUPB)(ZLZX2020-03)the R&D Department of China National Petroleum Corporation(2022DQ0604-01)。
文摘The picking efficiency of seismic first breaks(FBs)has been greatly accelerated by deep learning(DL)technology.However,the picking accuracy and efficiency of DL methods still face huge challenges in low signal-to-noise ratio(SNR)situations.To address this issue,we propose a regression approach to pick FBs based on bidirectional long short-term memory(Bi LSTM)neural network by learning the implicit Eikonal equation of 3D inhomogeneous media with rugged topography in the target region.We employ a regressive model that represents the relationships among the elevation of shots,offset and the elevation of receivers with their seismic traveltime to predict the unknown FBs,from common-shot gathers with sparsely distributed traces.Different from image segmentation methods which automatically extract image features and classify FBs from seismic data,the proposed method can learn the inner relationship between field geometry and FBs.In addition,the predicted results by the regressive model are continuous values of FBs rather than the discrete ones of the binary distribution.The picking results of synthetic data shows that the proposed method has low dependence on label data,and can obtain reliable and similar predicted results using two types of label data with large differences.The picking results of9380 shots for 3D seismic data generated by vibroseis indicate that the proposed method can still accurately predict FBs in low SNR data.The subsequent stacked profiles further illustrate the reliability and effectiveness of the proposed method.The results of model data and field seismic data demonstrate that the proposed regression method is a robust first-break picker with high potential for field application.
基金This work was performed under the auspices of the National Nuclear Security Administration of the US Department of Energy at Los Alamos National Laboratory under Contract No.89233218CNA000001The Authors gratefully acknowledge the support of the US Department of Energy National Nuclear Security Administration Advanced Simulation and Computing Program.LA-UR-22-33159.
文摘To accurately model flows with shock waves using staggered-grid Lagrangian hydrodynamics, the artificial viscosity has to be introduced to convert kinetic energy into internal energy, thereby increasing the entropy across shocks. Determining the appropriate strength of the artificial viscosity is an art and strongly depends on the particular problem and experience of the researcher. The objective of this study is to pose the problem of finding the appropriate strength of the artificial viscosity as an optimization problem and solve this problem using machine learning (ML) tools, specifically using surrogate models based on Gaussian Process regression (GPR) and Bayesian analysis. We describe the optimization method and discuss various practical details of its implementation. The shock-containing problems for which we apply this method all have been implemented in the LANL code FLAG (Burton in Connectivity structures and differencing techniques for staggered-grid free-Lagrange hydrodynamics, Tech. Rep. UCRL-JC-110555, Lawrence Livermore National Laboratory, Livermore, CA, 1992, 1992, in Consistent finite-volume discretization of hydrodynamic conservation laws for unstructured grids, Tech. Rep. CRL-JC-118788, Lawrence Livermore National Laboratory, Livermore, CA, 1992, 1994, Multidimensional discretization of conservation laws for unstructured polyhedral grids, Tech. Rep. UCRL-JC-118306, Lawrence Livermore National Laboratory, Livermore, CA, 1992, 1994, in FLAG, a multi-dimensional, multiple mesh, adaptive free-Lagrange, hydrodynamics code. In: NECDC, 1992). First, we apply ML to find optimal values to isolated shock problems of different strengths. Second, we apply ML to optimize the viscosity for a one-dimensional (1D) propagating detonation problem based on Zel’dovich-von Neumann-Doring (ZND) (Fickett and Davis in Detonation: theory and experiment. Dover books on physics. Dover Publications, Mineola, 2000) detonation theory using a reactive burn model. We compare results for default (currently used values in FLAG) and optimized values of the artificial viscosity for these problems demonstrating the potential for significant improvement in the accuracy of computations.
基金supported by National Science and Technology Infrastructure Platform National Population and Health Science Data Sharing Service Platform Public Health Science Data Center[NCMI-ZB01N-201905]。
文摘Objective This study employs the Geographically and Temporally Weighted Regression(GTWR)model to assess the impact of meteorological elements and imported cases on dengue fever outbreaks,emphasizing the spatial-temporal variability of these factors in border regions.Methods We conducted a descriptive analysis of dengue fever’s temporal-spatial distribution in Yunnan border areas.Utilizing annual data from 2013 to 2019,with each county in the Yunnan border serving as a spatial unit,we constructed a GTWR model to investigate the determinants of dengue fever and their spatio-temporal heterogeneity in this region.Results The GTWR model,proving more effective than Ordinary Least Squares(OLS)analysis,identified significant spatial and temporal heterogeneity in factors influencing dengue fever’s spread along the Yunnan border.Notably,the GTWR model revealed a substantial variation in the relationship between indigenous dengue fever incidence,meteorological variables,and imported cases across different counties.Conclusion In the Yunnan border areas,local dengue incidence is affected by temperature,humidity,precipitation,wind speed,and imported cases,with these factors’influence exhibiting notable spatial and temporal variation.
文摘Ignimbrites have been widely used as building materials in many historical and touristic structures in the Kayseri region of Türkiye. Their diverse colours and textures make them a popular choice for modern construction as well. However, ignimbrites are particularly vulnerable to atmospheric conditions, such as freeze-thaw cycles, due to their high porosity, which is a result of their formation process. When water enters the pores of the ignimbrites, it can freeze during cold weather. As the water freezes and expands, it generates internal stress within the stone, causing micro-cracks to develop. Over time, repeated freeze-thaw (F-T) cycles lead to the growth of these micro-cracks into larger cracks, compromising the structural integrity of the ignimbrites and eventually making them unsuitable for use as building materials. The determination of the long-term F-T performance of ignimbrites can be established after long F-T experimental processes. Determining the long-term F-T performance of ignimbrites typically requires extensive experimental testing over prolonged freeze-thaw cycles. To streamline this process, developing accurate predictive equations becomes crucial. In this study, such equations were formulated using classical regression analyses and artificial neural networks (ANN) based on data obtained from these experiments, allowing for the prediction of the F-T performance of ignimbrites and other similar building stones without the need for lengthy testing. In this study, uniaxial compressive strength, ultrasonic propagation velocity, apparent porosity and mass loss of ignimbrites after long-term F-T were determined. Following the F-T cycles, the disintegration rate was evaluated using decay function approaches, while uniaxial compressive strength (UCS) values were predicted with minimal input parameters through both regression and ANN analyses. The ANN and regression models created for this purpose were first started with a single input value and then developed with two and three combinations. The predictive performance of the models was assessed by comparing them to regression models using the coefficient of determination (R2) as the evaluation criterion. As a result of the study, higher R2 values (0.87) were obtained in models built with artificial neural network. The results of the study indicate that ANN usage can produce results close to experimental outcomes in predicting the long-term F-T performance of ignimbrite samples.
基金This work was supported by the National Natural Science Foundation of China(Nos.11875027,11975096).
文摘The extended kernel ridge regression(EKRR)method with odd-even effects was adopted to improve the description of the nuclear charge radius using five commonly used nuclear models.These are:(i)the isospin-dependent A^(1∕3) formula,(ii)relativistic continuum Hartree-Bogoliubov(RCHB)theory,(iii)Hartree-Fock-Bogoliubov(HFB)model HFB25,(iv)the Weizsacker-Skyrme(WS)model WS*,and(v)HFB25*model.In the last two models,the charge radii were calculated using a five-parameter formula with the nuclear shell corrections and deformations obtained from the WS and HFB25 models,respectively.For each model,the resultant root-mean-square deviation for the 1014 nuclei with proton number Z≥8 can be significantly reduced to 0.009-0.013 fm after considering the modification with the EKRR method.The best among them was the RCHB model,with a root-mean-square deviation of 0.0092 fm.The extrapolation abilities of the KRR and EKRR methods for the neutron-rich region were examined,and it was found that after considering the odd-even effects,the extrapolation power was improved compared with that of the original KRR method.The strong odd-even staggering of nuclear charge radii of Ca and Cu isotopes and the abrupt kinks across the neutron N=126 and 82 shell closures were also calculated and could be reproduced quite well by calculations using the EKRR method.
文摘In view of the composition analysis and identification of ancient glass products, L1 regularization, K-Means cluster analysis, elbow rule and other methods were comprehensively used to build logical regression, cluster analysis, hyper-parameter test and other models, and SPSS, Python and other tools were used to obtain the classification rules of glass products under different fluxes, sub classification under different chemical compositions, hyper-parameter K value test and rationality analysis. Research can provide theoretical support for the protection and restoration of ancient glass relics.
文摘This article develops a procedure for screening variables, in ultra high-di- mensional settings, based on their predictive significance. This is achieved by ranking the variables according to the variance of their respective marginal regression functions (RV-SIS). We show that, under some mild technical conditions, the RV-SIS possesses a sure screening property, which is defined by Fan and Lv (2008). Numerical comparisons suggest that RV-SIS has competitive performance compared to other screening procedures, and outperforms them in many different model settings.
基金Supported by Fundamental Research Funds for the Central Universities of China(Grant No.2023JBMC014).
文摘In the railway system,fasteners have the functions of damping,maintaining the track distance,and adjusting the track level.Therefore,routine maintenance and inspection of fasteners are important to ensure the safe operation of track lines.Currently,assessment methods for fastener tightness include manual observation,acoustic wave detection,and image detection.There are limitations such as low accuracy and efficiency,easy interference and misjudgment,and a lack of accurate,stable,and fast detection methods.Aiming at the small deformation characteristics and large elastic change of fasteners from full loosening to full tightening,this study proposes high-precision surface-structured light technology for fastener detection and fastener deformation feature extraction based on the center-line projection distance and a fastener tightness regression method based on neural networks.First,the method uses a 3D camera to obtain a fastener point cloud and then segments the elastic rod area based on the iterative closest point algorithm registration.Principal component analysis is used to calculate the normal vector of the segmented elastic rod surface and extract the point on the centerline of the elastic rod.The point is projected onto the upper surface of the bolt to calculate the projection distance.Subsequently,the mapping relationship between the projection distance sequence and fastener tightness is established,and the influence of each parameter on the fastener tightness prediction is analyzed.Finally,by setting up a fastener detection scene in the track experimental base,collecting data,and completing the algorithm verification,the results showed that the deviation between the fastener tightness regression value obtained after the algorithm processing and the actual measured value RMSE was 0.2196 mm,which significantly improved the effect compared with other tightness detection methods,and realized an effective fastener tightness regression.
基金supported by the Municipal Science and Technology Program of Wuwei City,China(WW2202RPZ037)the Fundamental Research Funds for the Central Universities in China(Grant No.lzujbky-2018-69).
文摘Objective Previous studies on the association between lipid profiles and chronic kidney disease(CKD)have yielded inconsistent results and no defined thresholds for blood lipids.Methods A prospective cohort study including 32,351 subjects who completed baseline and follow-up surveys over 5 years was conducted.Restricted cubic splines and Cox models were used to examine the association between the lipid profiles and CKD.A regression discontinuity design was used to determine the cutoff value of lipid profiles that was significantly associated with increased the risk of CKD.Results Over a median follow-up time of 2.2(0.5,4.2)years,648(2.00%)subjects developed CKD.The lipid profiles that were significantly and linearly related to CKD included total cholesterol(TC),triglycerides(TG),high-density lipoprotein cholesterol(HDL-C),TC/HDL-C,and TG/HDL-C,whereas lowdensity lipoprotein cholesterol(LDL-C)and LDL-C/HDL-C were nonlinearly correlated with CKD.TC,TG,TC/HDL-C,and TG/HDL-C showed an upward jump at the cutoff value,increasing the risk of CKD by 0.90%,1.50%,2.30%,and 1.60%,respectively,whereas HDL-C showed a downward jump at the cutoff value,reducing this risk by 1.0%.Female and participants with dyslipidemia had a higher risk of CKD,while the cutoff values for the different characteristics of the population were different.Conclusion There was a significant association between lipid profiles and CKD in a prospective cohort from Northwest China,while TG,TC/HDL-C,and TG/HDL-C showed a stronger risk association.The specific cutoff values of lipid profiles may provide a clinical reference for screening or diagnosing CKD risk.
文摘This research introduces a novel approach to improve and optimize the predictive capacity of consumer purchase behaviors on e-commerce platforms. This study presented an introduction to the fundamental concepts of the logistic regression algorithm. In addition, it analyzed user data obtained from an e-commerce platform. The original data were preprocessed, and a consumer purchase prediction model was developed for the e-commerce platform using the logistic regression method. The comparison study used the classic random forest approach, further enhanced by including the K-fold cross-validation method. Evaluation of the accuracy of the model’s classification was conducted using performance indicators that included the accuracy rate, the precision rate, the recall rate, and the F1 score. A visual examination determined the significance of the findings. The findings suggest that employing the logistic regression algorithm to forecast customer purchase behaviors on e-commerce platforms can improve the efficacy of the approach and yield more accurate predictions. This study serves as a valuable resource for improving the precision of forecasting customers’ purchase behaviors on e-commerce platforms. It has significant practical implications for optimizing the operational efficiency of e-commerce platforms.