To model the strain-inducedγ→α′phase transformation for the Cr-Mn metastable austenitic stainless steel,the 201Cu steel was chosen as the analytical material and the cylindrical samples of this steel with size ofϕ...To model the strain-inducedγ→α′phase transformation for the Cr-Mn metastable austenitic stainless steel,the 201Cu steel was chosen as the analytical material and the cylindrical samples of this steel with size ofϕ5 mm×10 mm were compressed at strains of 0.2–0.6 in the temperature range of 25–150℃ and in the strain rate range of 0.1–5.0 s^(−1).The flaky samples were prepared by wire cutting from the cylindrical samples and the volume fraction of the strain-inducedα′phase was detected in the test point of the flaky samples.The volume fraction changing with the process parameters was modeled,and the critical temperatures and the critical strains to preventγ→α′phase transformation were calculated as other different process parameters changed.The linear fitting goodness of the model between the calculated volume fraction values and the tested ones is 0.986 and the validity of the model was verified by application in cold and warm rolling experiments.展开更多
Artificial Intelligence(AI)and Computer Vision(CV)advancements have led to many useful methodologies in recent years,particularly to help visually-challenged people.Object detection includes a variety of challenges,fo...Artificial Intelligence(AI)and Computer Vision(CV)advancements have led to many useful methodologies in recent years,particularly to help visually-challenged people.Object detection includes a variety of challenges,for example,handlingmultiple class images,images that get augmented when captured by a camera and so on.The test images include all these variants as well.These detection models alert them about their surroundings when they want to walk independently.This study compares four CNN-based pre-trainedmodels:ResidualNetwork(ResNet-50),Inception v3,DenseConvolutional Network(DenseNet-121),and SqueezeNet,predominantly used in image recognition applications.Based on the analysis performed on these test images,the study infers that Inception V3 outperformed other pre-trained models in terms of accuracy and speed.To further improve the performance of the Inception v3 model,the thermal exchange optimization(TEO)algorithm is applied to tune the hyperparameters(number of epochs,batch size,and learning rate)showing the novelty of the work.Better accuracy was achieved owing to the inclusion of an auxiliary classifier as a regularizer,hyperparameter optimizer,and factorization approach.Additionally,Inception V3 can handle images of different sizes.This makes Inception V3 the optimum model for assisting visually challenged people in real-world communication when integrated with Internet of Things(IoT)-based devices.展开更多
COVID-19 has created a panic all around the globe.It is a contagious dis-ease caused by Severe Acute Respiratory Syndrome Coronavirus 2(SARS-CoV-2),originated from Wuhan in December 2019 and spread quickly all over th...COVID-19 has created a panic all around the globe.It is a contagious dis-ease caused by Severe Acute Respiratory Syndrome Coronavirus 2(SARS-CoV-2),originated from Wuhan in December 2019 and spread quickly all over the world.The healthcare sector of the world is facing great challenges tackling COVID cases.One of the problems many have witnessed is the misdiagnosis of COVID-19 cases with that of healthy and pneumonia cases.In this article,we propose a deep Convo-lutional Neural Network(CNN)based approach to detect COVID+(i.e.,patients with COVID-19),pneumonia and normal cases,from the chest X-ray images.COVID-19 detection from chest X-ray is suitable considering all aspects in compar-ison to Reverse Transcription Polymerase Chain Reaction(RT-PCR)and Computed Tomography(CT)scan.Several deep CNN models including VGG16,InceptionV3,DenseNet121,DenseNet201 and InceptionResNetV2 have been adopted in this pro-posed work.They have been trained individually to make particular predictions.Empirical results demonstrate that DenseNet201 provides overall better performance with accuracy,recall,F1-score and precision of 94.75%,96%,95%and 95%respec-tively.After careful comparison with results available in the literature,we have found to develop models with a higher reliability.All the studies were carried out using a publicly available chest X-ray(CXR)image data-set.展开更多
基金the financial support from the Open Project provided by the State Key Laboratory of Development and Application Technology of Automotive Steels(Baowu Steel Group)(Grant No.Y21ECEQ17Y).
文摘To model the strain-inducedγ→α′phase transformation for the Cr-Mn metastable austenitic stainless steel,the 201Cu steel was chosen as the analytical material and the cylindrical samples of this steel with size ofϕ5 mm×10 mm were compressed at strains of 0.2–0.6 in the temperature range of 25–150℃ and in the strain rate range of 0.1–5.0 s^(−1).The flaky samples were prepared by wire cutting from the cylindrical samples and the volume fraction of the strain-inducedα′phase was detected in the test point of the flaky samples.The volume fraction changing with the process parameters was modeled,and the critical temperatures and the critical strains to preventγ→α′phase transformation were calculated as other different process parameters changed.The linear fitting goodness of the model between the calculated volume fraction values and the tested ones is 0.986 and the validity of the model was verified by application in cold and warm rolling experiments.
基金Princess Nourah bint Abdulrahman University Researchers Supporting Project number(PNURSP2023R191)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.The authors would like to thank the Deanship of Scientific Research at Umm Al-Qura University for supporting this work by Grant Code:(22UQU4310373DSR61)This study is supported via funding from Prince Sattam bin Abdulaziz University project number(PSAU/2023/R/1444).
文摘Artificial Intelligence(AI)and Computer Vision(CV)advancements have led to many useful methodologies in recent years,particularly to help visually-challenged people.Object detection includes a variety of challenges,for example,handlingmultiple class images,images that get augmented when captured by a camera and so on.The test images include all these variants as well.These detection models alert them about their surroundings when they want to walk independently.This study compares four CNN-based pre-trainedmodels:ResidualNetwork(ResNet-50),Inception v3,DenseConvolutional Network(DenseNet-121),and SqueezeNet,predominantly used in image recognition applications.Based on the analysis performed on these test images,the study infers that Inception V3 outperformed other pre-trained models in terms of accuracy and speed.To further improve the performance of the Inception v3 model,the thermal exchange optimization(TEO)algorithm is applied to tune the hyperparameters(number of epochs,batch size,and learning rate)showing the novelty of the work.Better accuracy was achieved owing to the inclusion of an auxiliary classifier as a regularizer,hyperparameter optimizer,and factorization approach.Additionally,Inception V3 can handle images of different sizes.This makes Inception V3 the optimum model for assisting visually challenged people in real-world communication when integrated with Internet of Things(IoT)-based devices.
文摘COVID-19 has created a panic all around the globe.It is a contagious dis-ease caused by Severe Acute Respiratory Syndrome Coronavirus 2(SARS-CoV-2),originated from Wuhan in December 2019 and spread quickly all over the world.The healthcare sector of the world is facing great challenges tackling COVID cases.One of the problems many have witnessed is the misdiagnosis of COVID-19 cases with that of healthy and pneumonia cases.In this article,we propose a deep Convo-lutional Neural Network(CNN)based approach to detect COVID+(i.e.,patients with COVID-19),pneumonia and normal cases,from the chest X-ray images.COVID-19 detection from chest X-ray is suitable considering all aspects in compar-ison to Reverse Transcription Polymerase Chain Reaction(RT-PCR)and Computed Tomography(CT)scan.Several deep CNN models including VGG16,InceptionV3,DenseNet121,DenseNet201 and InceptionResNetV2 have been adopted in this pro-posed work.They have been trained individually to make particular predictions.Empirical results demonstrate that DenseNet201 provides overall better performance with accuracy,recall,F1-score and precision of 94.75%,96%,95%and 95%respec-tively.After careful comparison with results available in the literature,we have found to develop models with a higher reliability.All the studies were carried out using a publicly available chest X-ray(CXR)image data-set.