Unmanned Aerial Vehicle(UAV)stands as a burgeoning electric transportation carrier,holding substantial promise for the logistics sector.A reinforcement learning framework Centralized-S Proximal Policy Optimization(C-S...Unmanned Aerial Vehicle(UAV)stands as a burgeoning electric transportation carrier,holding substantial promise for the logistics sector.A reinforcement learning framework Centralized-S Proximal Policy Optimization(C-SPPO)based on centralized decision process and considering policy entropy(S)is proposed.The proposed framework aims to plan the best scheduling scheme with the objective of minimizing both the timeout of order requests and the flight impact of UAVs that may lead to conflicts.In this framework,the intents of matching act are generated through the observations of UAV agents,and the ultimate conflict-free matching results are output under the guidance of a centralized decision maker.Concurrently,a pre-activation operation is introduced to further enhance the cooperation among UAV agents.Simulation experiments based on real-world data from New York City are conducted.The results indicate that the proposed CSPPO outperforms the baseline algorithms in the Average Delay Time(ADT),the Maximum Delay Time(MDT),the Order Delay Rate(ODR),the Average Flight Distance(AFD),and the Flight Impact Ratio(FIR).Furthermore,the framework demonstrates scalability to scenarios of different sizes without requiring additional training.展开更多
The knapsack problem is a classical combinatorial optimization problem widely encountered in areas such as logistics,resource allocation,and portfolio optimization.Traditional methods,including dynamic program-ming(DP...The knapsack problem is a classical combinatorial optimization problem widely encountered in areas such as logistics,resource allocation,and portfolio optimization.Traditional methods,including dynamic program-ming(DP)and greedy algorithms,have been effective in solving small problem instances but often struggle with scalability and efficiency as the problem size increases.DP,for instance,has exponential time complexity and can become computationally prohibitive for large problem instances.On the other hand,greedy algorithms offer faster solutions but may not always yield the optimal results,especially when the problem involves complex constraints or large numbers of items.This paper introduces a novel reinforcement learning(RL)approach to solve the knapsack problem by enhancing the state representation within the learning environment.We propose a representation where item weights and volumes are expressed as ratios relative to the knapsack’s capacity,and item values are normalized to represent their percentage of the total value across all items.This novel state modification leads to a 5%improvement in accuracy compared to the state-of-the-art RL-based algorithms,while significantly reducing execution time.Our RL-based method outperforms DP by over 9000 times in terms of speed,making it highly scalable for larger problem instances.Furthermore,we improve the performance of the RL model by incorporating Noisy layers into the neural network architecture.The addition of Noisy layers enhances the exploration capabilities of the agent,resulting in an additional accuracy boost of 0.2%–0.5%.The results demonstrate that our approach not only outperforms existing RL techniques,such as the Transformer model in terms of accuracy,but also provides a substantial improvement than DP in computational efficiency.This combination of enhanced accuracy and speed presents a promising solution for tackling large-scale optimization problems in real-world applications,where both precision and time are critical factors.展开更多
Recent studies employing deep learning to solve the traveling salesman problem(TSP)have mainly focused on learning construction heuristics.Such methods can improve TSP solutions,but still depend on additional programs...Recent studies employing deep learning to solve the traveling salesman problem(TSP)have mainly focused on learning construction heuristics.Such methods can improve TSP solutions,but still depend on additional programs.However,methods that focus on learning improvement heuristics to iteratively refine solutions remain insufficient.Traditional improvement heuristics are guided by a manually designed search strategy and may only achieve limited improvements.This paper proposes a novel framework for learning improvement heuristics,which automatically discovers better improvement policies for heuristics to iteratively solve the TSP.Our framework first designs a new architecture based on a transformer model to make the policy network parameterized,which introduces an action-dropout layer to prevent action selection from overfitting.It then proposes a deep reinforcement learning approach integrating a simulated annealing mechanism(named RL-SA)to learn the pairwise selected policy,aiming to improve the 2-opt algorithm's performance.The RL-SA leverages the whale optimization algorithm to generate initial solutions for better sampling efficiency and uses the Gaussian perturbation strategy to tackle the sparse reward problem of reinforcement learning.The experiment results show that the proposed approach is significantly superior to the state-of-the-art learning-based methods,and further reduces the gap between learning-based methods and highly optimized solvers in the benchmark datasets.Moreover,our pre-trained model M can be applied to guide the SA algorithm(named M-SA(ours)),which performs better than existing deep models in small-,medium-,and large-scale TSPLIB datasets.Additionally,the M-SA(ours)achieves excellent generalization performance in a real-world dataset on global liner shipping routes,with the optimization percentages in distance reduction ranging from3.52%to 17.99%.展开更多
For unachievable tracking problems, where the system output cannot precisely track a given reference, achieving the best possible approximation for the reference trajectory becomes the objective. This study aims to in...For unachievable tracking problems, where the system output cannot precisely track a given reference, achieving the best possible approximation for the reference trajectory becomes the objective. This study aims to investigate solutions using the Ptype learning control scheme. Initially, we demonstrate the necessity of gradient information for achieving the best approximation.Subsequently, we propose an input-output-driven learning gain design to handle the imprecise gradients of a class of uncertain systems. However, it is discovered that the desired performance may not be attainable when faced with incomplete information.To address this issue, an extended iterative learning control scheme is introduced. In this scheme, the tracking errors are modified through output data sampling, which incorporates lowmemory footprints and offers flexibility in learning gain design.The input sequence is shown to converge towards the desired input, resulting in an output that is closest to the given reference in the least square sense. Numerical simulations are provided to validate the theoretical findings.展开更多
This paper presents an end-to-end deep learning method to solve geometry problems via feature learning and contrastive learning of multimodal data.A key challenge in solving geometry problems using deep learning is to...This paper presents an end-to-end deep learning method to solve geometry problems via feature learning and contrastive learning of multimodal data.A key challenge in solving geometry problems using deep learning is to automatically adapt to the task of understanding single-modal and multimodal problems.Existing methods either focus on single-modal ormultimodal problems,and they cannot fit each other.A general geometry problem solver shouldobviouslybe able toprocess variousmodalproblems at the same time.Inthispaper,a shared feature-learning model of multimodal data is adopted to learn the unified feature representation of text and image,which can solve the heterogeneity issue between multimodal geometry problems.A contrastive learning model of multimodal data enhances the semantic relevance betweenmultimodal features and maps them into a unified semantic space,which can effectively adapt to both single-modal and multimodal downstream tasks.Based on the feature extraction and fusion of multimodal data,a proposed geometry problem solver uses relation extraction,theorem reasoning,and problem solving to present solutions in a readable way.Experimental results show the effectiveness of the method.展开更多
Artificial rabbits optimization(ARO)is a recently proposed biology-based optimization algorithm inspired by the detour foraging and random hiding behavior of rabbits in nature.However,for solving optimization problems...Artificial rabbits optimization(ARO)is a recently proposed biology-based optimization algorithm inspired by the detour foraging and random hiding behavior of rabbits in nature.However,for solving optimization problems,the ARO algorithm shows slow convergence speed and can fall into local minima.To overcome these drawbacks,this paper proposes chaotic opposition-based learning ARO(COARO),an improved version of the ARO algorithm that incorporates opposition-based learning(OBL)and chaotic local search(CLS)techniques.By adding OBL to ARO,the convergence speed of the algorithm increases and it explores the search space better.Chaotic maps in CLS provide rapid convergence by scanning the search space efficiently,since their ergodicity and non-repetitive properties.The proposed COARO algorithm has been tested using thirty-three distinct benchmark functions.The outcomes have been compared with the most recent optimization algorithms.Additionally,the COARO algorithm’s problem-solving capabilities have been evaluated using six different engineering design problems and compared with various other algorithms.This study also introduces a binary variant of the continuous COARO algorithm,named BCOARO.The performance of BCOARO was evaluated on the breast cancer dataset.The effectiveness of BCOARO has been compared with different feature selection algorithms.The proposed BCOARO outperforms alternative algorithms,according to the findings obtained for real applications in terms of accuracy performance,and fitness value.Extensive experiments show that the COARO and BCOARO algorithms achieve promising results compared to other metaheuristic algorithms.展开更多
Recently,Problem-Based Learning(PBL)has been regarded as a possible way towards effective educational changes in Chinese universities.However,problems have been exposed in the adoption of PBL,such as choosing effectiv...Recently,Problem-Based Learning(PBL)has been regarded as a possible way towards effective educational changes in Chinese universities.However,problems have been exposed in the adoption of PBL,such as choosing effective PBL problems.The purpose of this paper is to provide a possible solution to the formulation of PBL problems for computer science courses,which is to reimplement open-source projects(ROSP).A case is demonstrated by showing how ROSP was adopted in a practical intercourse-level PBL course module.This paper contributes to a new PBL problem formulation method for promoting PBL in a practical way for Chinese universities.展开更多
To solve the sparse reward problem of job-shop scheduling by deep reinforcement learning,a deep reinforcement learning framework considering sparse reward problem is proposed.The job shop scheduling problem is transfo...To solve the sparse reward problem of job-shop scheduling by deep reinforcement learning,a deep reinforcement learning framework considering sparse reward problem is proposed.The job shop scheduling problem is transformed into Markov decision process,and six state features are designed to improve the state feature representation by using two-way scheduling method,including four state features that distinguish the optimal action and two state features that are related to the learning goal.An extended variant of graph isomorphic network GIN++is used to encode disjunction graphs to improve the performance and generalization ability of the model.Through iterative greedy algorithm,random strategy is generated as the initial strategy,and the action with the maximum information gain is selected to expand it to optimize the exploration ability of Actor-Critic algorithm.Through validation of the trained policy model on multiple public test data sets and comparison with other advanced DRL methods and scheduling rules,the proposed method reduces the minimum average gap by 3.49%,5.31%and 4.16%,respectively,compared with the priority rule-based method,and 5.34%compared with the learning-based method.11.97%and 5.02%,effectively improving the accuracy of DRL to solve the approximate solution of JSSP minimum completion time.展开更多
The emotional factors play an important role in the process of language learning. Exploration of the emotional problems in the process of English learning can promote the emotional communication between teachers and s...The emotional factors play an important role in the process of language learning. Exploration of the emotional problems in the process of English learning can promote the emotional communication between teachers and students. The organic combination of the two aspects of cognitive and emotional can help to finish the promotion of cognitive and education with emotion. People abroad have achieved many results about the research, while further research in domestic is still rare. This paper mainly analyzes the main emotional problems influencing the English leaming, clarifies the importance of using emotional strategy in English learning, points out the way of improving learning effect by affective strategies, thus to guide the students form the corresponding emotion strategy, effectively use in English learning, and improve the English learning.展开更多
English is widely taught as an FL (foreign language) in most middle schools in China. Chinese students have improved their English reading and writing these years, but they are relatively poor in listening and speak...English is widely taught as an FL (foreign language) in most middle schools in China. Chinese students have improved their English reading and writing these years, but they are relatively poor in listening and speaking. The author, based on the empirical study, made an investigation on Senior Three students in Leshan city. The statistics from the questionnaire indicate the problem of "Mute-crux" in English teaching and learning. This paper explores the reasons that cause the problem: teachers' use of the traditional ways, just emphasizing the transmission of English knowledge, but ignoring the meaningful communication with students; exercise-based and exam-oriented teaching, which cause students' rote-learning. The author proposes that Chinese English teachers ought to change the traditional notions and adopt effective ways to encourage students to speak bravely in order to improve their speaking abilities展开更多
With the rapid development of China's higher education, higher education, popularization of increasingly rapid pace, the future number of people receiving higher education will grow exponentially. In the rapidly grow...With the rapid development of China's higher education, higher education, popularization of increasingly rapid pace, the future number of people receiving higher education will grow exponentially. In the rapidly growing scale of higher education at the same time, the quality of higher education has been a cause for concern This paper analyzes the current problems in the study of college students exist, and improvement measures put forward their own proposals.展开更多
With the rapid development of China's higher education, higher education, popularization of increasingly rapid pace, the future number of people receiving higher education will grow exponentially. In the rapidly grow...With the rapid development of China's higher education, higher education, popularization of increasingly rapid pace, the future number of people receiving higher education will grow exponentially. In the rapidly growing scale of higher education at the same time, the quality of higher education has been a cause for concern~ This paper analyzes the current problems in the study of college students exist, and improvement measures put forward their own proposals.展开更多
Using literature review,questionnaire survey,mathematical statistics and other research methods,combined with the previous research results of the research team,systematically sort out the problems of"learning&qu...Using literature review,questionnaire survey,mathematical statistics and other research methods,combined with the previous research results of the research team,systematically sort out the problems of"learning"in the"learning,training,competition,and use"physical education curriculum,and how to solve the problems.The strategic research of the Chinese Academy of Sciences aims to promote the scientificization of"learning"in the teaching of physical education courses,thereby improving the quality and efficiency of physical education courses.展开更多
Pupils learning ability is limited, so it is difficult to get good learning results in art learning by their own understanding and practice. And art knowledge has a long history, which is not as simple and easy to und...Pupils learning ability is limited, so it is difficult to get good learning results in art learning by their own understanding and practice. And art knowledge has a long history, which is not as simple and easy to understand as it seems. This requires teachers to formulate teaching plans in combination with students physiological and psychological characteristics, and to improve students learning efficiency by means of cooperative learning. Cooperative learning can not only create a good teaching atmosphere, stimulate students interest in inquiry, provide a stage for students to show themselves and share their learning achievements, but also strengthen students sense of responsibility, stimulate students artistic thinking, improve their artistic accomplishment, and promote the progress and promotion of students cooperative learning ability.展开更多
With the popularity of Computer Assisted Language Learning(CALL),autonomous language learning centers have been established in the universities throughout the country.However,there are many problems exist in the pract...With the popularity of Computer Assisted Language Learning(CALL),autonomous language learning centers have been established in the universities throughout the country.However,there are many problems exist in the practice of the autonomous language learning center.This essay tries to discuss the problems and its countermeasures with an aim of improving College English learning and teaching by the example of autonomous language learning center in Nanyang Institute of Technology.展开更多
This paper is concerned with a novel integrated multi-step heuristic dynamic programming(MsHDP)algorithm for solving optimal control problems.It is shown that,initialized by the zero cost function,MsHDP can converge t...This paper is concerned with a novel integrated multi-step heuristic dynamic programming(MsHDP)algorithm for solving optimal control problems.It is shown that,initialized by the zero cost function,MsHDP can converge to the optimal solution of the Hamilton-Jacobi-Bellman(HJB)equation.Then,the stability of the system is analyzed using control policies generated by MsHDP.Also,a general stability criterion is designed to determine the admissibility of the current control policy.That is,the criterion is applicable not only to traditional value iteration and policy iteration but also to MsHDP.Further,based on the convergence and the stability criterion,the integrated MsHDP algorithm using immature control policies is developed to accelerate learning efficiency greatly.Besides,actor-critic is utilized to implement the integrated MsHDP scheme,where neural networks are used to evaluate and improve the iterative policy as the parameter architecture.Finally,two simulation examples are given to demonstrate that the learning effectiveness of the integrated MsHDP scheme surpasses those of other fixed or integrated methods.展开更多
The vehicle routing problem(VRP)is a typical discrete combinatorial optimization problem,and many models and algorithms have been proposed to solve the VRP and its variants.Although existing approaches have contribute...The vehicle routing problem(VRP)is a typical discrete combinatorial optimization problem,and many models and algorithms have been proposed to solve the VRP and its variants.Although existing approaches have contributed significantly to the development of this field,these approaches either are limited in problem size or need manual intervention in choosing parameters.To solve these difficulties,many studies have considered learning-based optimization(LBO)algorithms to solve the VRP.This paper reviews recent advances in this field and divides relevant approaches into end-to-end approaches and step-by-step approaches.We performed a statistical analysis of the reviewed articles from various aspects and designed three experiments to evaluate the performance of four representative LBO algorithms.Finally,we conclude the applicable types of problems for different LBO algorithms and suggest directions in which researchers can improve LBO algorithms.展开更多
With the rapid development of air transportation in recent years,airport operations have attracted a lot of attention.Among them,airport gate assignment problem(AGAP)has become a research hotspot.However,the real-time...With the rapid development of air transportation in recent years,airport operations have attracted a lot of attention.Among them,airport gate assignment problem(AGAP)has become a research hotspot.However,the real-time AGAP algorithm is still an open issue.In this study,a deep reinforcement learning based AGAP(DRL-AGAP)is proposed.The optimization object is to maximize the rate of flights assigned to fixed gates.The real-time AGAP is modeled as a Markov decision process(MDP).The state space,action space,value and rewards have been defined.The DRL-AGAP algorithm is evaluated via simulation and it is compared with the flight pre-assignment results of the optimization software Gurobiand Greedy.Simulation results show that the performance of the proposed DRL-AGAP algorithm is close to that of pre-assignment obtained by the Gurobi optimization solver.Meanwhile,the real-time assignment ability is ensured by the proposed DRL-AGAP algorithm due to the dynamic modeling and lower complexity.展开更多
Objectives:Near misses happen more frequently than actual errors,and highlight system vulnerabilities without causing any harm,thus provide a safe space for organizational learning.Second-order problem solving behavio...Objectives:Near misses happen more frequently than actual errors,and highlight system vulnerabilities without causing any harm,thus provide a safe space for organizational learning.Second-order problem solving behavior offers a new perspective to better understand how nurses promote learning from near misses to improve organizational outcomes.This study aimed to explore frontline nurses’perspectives on using second-order problem solving behavior in learning from near misses to improve patient safety.Methods:A qualitative exploratory study design was employed.This study was conducted in three tertiary hospitals in east China from June to November 2015.Purposive sampling was used to recruit 19 frontline nurses.Semi-structured interviews and a qualitative directed content analysis was undertaken using Crossan’s 4I Framework of Organizational Learning as a coding framework.Results:Second-order problem solving behavior,based on the 4I Framework of Organizational Learning,was referred to as being a leader in exposing near misses,pushing forward the cause analysis within limited capacity,balancing the active and passive role during improvement project,and promoting the continuous improvement with passion while feeling low-powered.Conclusions:4I Framework of Organizational Learning can be an underlying guide to enrich frontline nurses’role in promoting organizations to learn from near misses.In this study,nurses displayed their pivotal role in organizational learning from near misses by using second-order problem solving.However,additional knowledge,skills,and support are needed to maximize the application of second-order problem solving behavior when near misses are recognized.展开更多
基金the support of the Chinese Special Research Project for Civil Aircraft(No.MJZ17N22)the National Natural Science Foundation of China(Nos.U2133207,U2333214)+1 种基金the China Postdoctoral Science Foundation(No.2023M741687)the National Social Science Fund of China(No.22&ZD169)。
文摘Unmanned Aerial Vehicle(UAV)stands as a burgeoning electric transportation carrier,holding substantial promise for the logistics sector.A reinforcement learning framework Centralized-S Proximal Policy Optimization(C-SPPO)based on centralized decision process and considering policy entropy(S)is proposed.The proposed framework aims to plan the best scheduling scheme with the objective of minimizing both the timeout of order requests and the flight impact of UAVs that may lead to conflicts.In this framework,the intents of matching act are generated through the observations of UAV agents,and the ultimate conflict-free matching results are output under the guidance of a centralized decision maker.Concurrently,a pre-activation operation is introduced to further enhance the cooperation among UAV agents.Simulation experiments based on real-world data from New York City are conducted.The results indicate that the proposed CSPPO outperforms the baseline algorithms in the Average Delay Time(ADT),the Maximum Delay Time(MDT),the Order Delay Rate(ODR),the Average Flight Distance(AFD),and the Flight Impact Ratio(FIR).Furthermore,the framework demonstrates scalability to scenarios of different sizes without requiring additional training.
基金supported in part by the Research Start-Up Funds of South-Central Minzu University under Grants YZZ23002,YZY23001,and YZZ18006in part by the Hubei Provincial Natural Science Foundation of China under Grants 2024AFB842 and 2023AFB202+3 种基金in part by the Knowledge Innovation Program of Wuhan Basic Research underGrant 2023010201010151in part by the Spring Sunshine Program of Ministry of Education of the People’s Republic of China under Grant HZKY20220331in part by the Funds for Academic Innovation Teams and Research Platformof South-CentralMinzu University Grant Number:XT224003,PTZ24001in part by the Career Development Fund(CDF)of the Agency for Science,Technology and Research(A*STAR)(Grant Number:C233312007).
文摘The knapsack problem is a classical combinatorial optimization problem widely encountered in areas such as logistics,resource allocation,and portfolio optimization.Traditional methods,including dynamic program-ming(DP)and greedy algorithms,have been effective in solving small problem instances but often struggle with scalability and efficiency as the problem size increases.DP,for instance,has exponential time complexity and can become computationally prohibitive for large problem instances.On the other hand,greedy algorithms offer faster solutions but may not always yield the optimal results,especially when the problem involves complex constraints or large numbers of items.This paper introduces a novel reinforcement learning(RL)approach to solve the knapsack problem by enhancing the state representation within the learning environment.We propose a representation where item weights and volumes are expressed as ratios relative to the knapsack’s capacity,and item values are normalized to represent their percentage of the total value across all items.This novel state modification leads to a 5%improvement in accuracy compared to the state-of-the-art RL-based algorithms,while significantly reducing execution time.Our RL-based method outperforms DP by over 9000 times in terms of speed,making it highly scalable for larger problem instances.Furthermore,we improve the performance of the RL model by incorporating Noisy layers into the neural network architecture.The addition of Noisy layers enhances the exploration capabilities of the agent,resulting in an additional accuracy boost of 0.2%–0.5%.The results demonstrate that our approach not only outperforms existing RL techniques,such as the Transformer model in terms of accuracy,but also provides a substantial improvement than DP in computational efficiency.This combination of enhanced accuracy and speed presents a promising solution for tackling large-scale optimization problems in real-world applications,where both precision and time are critical factors.
基金Project supported by the National Natural Science Foundation of China(Grant Nos.72101046 and 61672128)。
文摘Recent studies employing deep learning to solve the traveling salesman problem(TSP)have mainly focused on learning construction heuristics.Such methods can improve TSP solutions,but still depend on additional programs.However,methods that focus on learning improvement heuristics to iteratively refine solutions remain insufficient.Traditional improvement heuristics are guided by a manually designed search strategy and may only achieve limited improvements.This paper proposes a novel framework for learning improvement heuristics,which automatically discovers better improvement policies for heuristics to iteratively solve the TSP.Our framework first designs a new architecture based on a transformer model to make the policy network parameterized,which introduces an action-dropout layer to prevent action selection from overfitting.It then proposes a deep reinforcement learning approach integrating a simulated annealing mechanism(named RL-SA)to learn the pairwise selected policy,aiming to improve the 2-opt algorithm's performance.The RL-SA leverages the whale optimization algorithm to generate initial solutions for better sampling efficiency and uses the Gaussian perturbation strategy to tackle the sparse reward problem of reinforcement learning.The experiment results show that the proposed approach is significantly superior to the state-of-the-art learning-based methods,and further reduces the gap between learning-based methods and highly optimized solvers in the benchmark datasets.Moreover,our pre-trained model M can be applied to guide the SA algorithm(named M-SA(ours)),which performs better than existing deep models in small-,medium-,and large-scale TSPLIB datasets.Additionally,the M-SA(ours)achieves excellent generalization performance in a real-world dataset on global liner shipping routes,with the optimization percentages in distance reduction ranging from3.52%to 17.99%.
基金supported by the National Natural Science Foundation of China (62173333, 12271522)Beijing Natural Science Foundation (Z210002)the Research Fund of Renmin University of China (2021030187)。
文摘For unachievable tracking problems, where the system output cannot precisely track a given reference, achieving the best possible approximation for the reference trajectory becomes the objective. This study aims to investigate solutions using the Ptype learning control scheme. Initially, we demonstrate the necessity of gradient information for achieving the best approximation.Subsequently, we propose an input-output-driven learning gain design to handle the imprecise gradients of a class of uncertain systems. However, it is discovered that the desired performance may not be attainable when faced with incomplete information.To address this issue, an extended iterative learning control scheme is introduced. In this scheme, the tracking errors are modified through output data sampling, which incorporates lowmemory footprints and offers flexibility in learning gain design.The input sequence is shown to converge towards the desired input, resulting in an output that is closest to the given reference in the least square sense. Numerical simulations are provided to validate the theoretical findings.
基金supported by the NationalNatural Science Foundation of China (No.62107014,Jian P.,62177025,He B.)the Key R&D and Promotion Projects of Henan Province (No.212102210147,Jian P.)Innovative Education Program for Graduate Students at North China University of Water Resources and Electric Power,China (No.YK-2021-99,Guo F.).
文摘This paper presents an end-to-end deep learning method to solve geometry problems via feature learning and contrastive learning of multimodal data.A key challenge in solving geometry problems using deep learning is to automatically adapt to the task of understanding single-modal and multimodal problems.Existing methods either focus on single-modal ormultimodal problems,and they cannot fit each other.A general geometry problem solver shouldobviouslybe able toprocess variousmodalproblems at the same time.Inthispaper,a shared feature-learning model of multimodal data is adopted to learn the unified feature representation of text and image,which can solve the heterogeneity issue between multimodal geometry problems.A contrastive learning model of multimodal data enhances the semantic relevance betweenmultimodal features and maps them into a unified semantic space,which can effectively adapt to both single-modal and multimodal downstream tasks.Based on the feature extraction and fusion of multimodal data,a proposed geometry problem solver uses relation extraction,theorem reasoning,and problem solving to present solutions in a readable way.Experimental results show the effectiveness of the method.
基金funded by Firat University Scientific Research Projects Management Unit for the scientific research project of Feyza AltunbeyÖzbay,numbered MF.23.49.
文摘Artificial rabbits optimization(ARO)is a recently proposed biology-based optimization algorithm inspired by the detour foraging and random hiding behavior of rabbits in nature.However,for solving optimization problems,the ARO algorithm shows slow convergence speed and can fall into local minima.To overcome these drawbacks,this paper proposes chaotic opposition-based learning ARO(COARO),an improved version of the ARO algorithm that incorporates opposition-based learning(OBL)and chaotic local search(CLS)techniques.By adding OBL to ARO,the convergence speed of the algorithm increases and it explores the search space better.Chaotic maps in CLS provide rapid convergence by scanning the search space efficiently,since their ergodicity and non-repetitive properties.The proposed COARO algorithm has been tested using thirty-three distinct benchmark functions.The outcomes have been compared with the most recent optimization algorithms.Additionally,the COARO algorithm’s problem-solving capabilities have been evaluated using six different engineering design problems and compared with various other algorithms.This study also introduces a binary variant of the continuous COARO algorithm,named BCOARO.The performance of BCOARO was evaluated on the breast cancer dataset.The effectiveness of BCOARO has been compared with different feature selection algorithms.The proposed BCOARO outperforms alternative algorithms,according to the findings obtained for real applications in terms of accuracy performance,and fitness value.Extensive experiments show that the COARO and BCOARO algorithms achieve promising results compared to other metaheuristic algorithms.
基金This research was financially supported by the PBL Research and Application Project of Northeastern University(Grant No.PBL-JX2021yb029,PBL-JX2021yb027).
文摘Recently,Problem-Based Learning(PBL)has been regarded as a possible way towards effective educational changes in Chinese universities.However,problems have been exposed in the adoption of PBL,such as choosing effective PBL problems.The purpose of this paper is to provide a possible solution to the formulation of PBL problems for computer science courses,which is to reimplement open-source projects(ROSP).A case is demonstrated by showing how ROSP was adopted in a practical intercourse-level PBL course module.This paper contributes to a new PBL problem formulation method for promoting PBL in a practical way for Chinese universities.
基金Shaanxi Provincial Key Research and Development Project(2023YBGY095)and Shaanxi Provincial Qin Chuangyuan"Scientist+Engineer"project(2023KXJ247)Fund support.
文摘To solve the sparse reward problem of job-shop scheduling by deep reinforcement learning,a deep reinforcement learning framework considering sparse reward problem is proposed.The job shop scheduling problem is transformed into Markov decision process,and six state features are designed to improve the state feature representation by using two-way scheduling method,including four state features that distinguish the optimal action and two state features that are related to the learning goal.An extended variant of graph isomorphic network GIN++is used to encode disjunction graphs to improve the performance and generalization ability of the model.Through iterative greedy algorithm,random strategy is generated as the initial strategy,and the action with the maximum information gain is selected to expand it to optimize the exploration ability of Actor-Critic algorithm.Through validation of the trained policy model on multiple public test data sets and comparison with other advanced DRL methods and scheduling rules,the proposed method reduces the minimum average gap by 3.49%,5.31%and 4.16%,respectively,compared with the priority rule-based method,and 5.34%compared with the learning-based method.11.97%and 5.02%,effectively improving the accuracy of DRL to solve the approximate solution of JSSP minimum completion time.
文摘The emotional factors play an important role in the process of language learning. Exploration of the emotional problems in the process of English learning can promote the emotional communication between teachers and students. The organic combination of the two aspects of cognitive and emotional can help to finish the promotion of cognitive and education with emotion. People abroad have achieved many results about the research, while further research in domestic is still rare. This paper mainly analyzes the main emotional problems influencing the English leaming, clarifies the importance of using emotional strategy in English learning, points out the way of improving learning effect by affective strategies, thus to guide the students form the corresponding emotion strategy, effectively use in English learning, and improve the English learning.
文摘English is widely taught as an FL (foreign language) in most middle schools in China. Chinese students have improved their English reading and writing these years, but they are relatively poor in listening and speaking. The author, based on the empirical study, made an investigation on Senior Three students in Leshan city. The statistics from the questionnaire indicate the problem of "Mute-crux" in English teaching and learning. This paper explores the reasons that cause the problem: teachers' use of the traditional ways, just emphasizing the transmission of English knowledge, but ignoring the meaningful communication with students; exercise-based and exam-oriented teaching, which cause students' rote-learning. The author proposes that Chinese English teachers ought to change the traditional notions and adopt effective ways to encourage students to speak bravely in order to improve their speaking abilities
文摘With the rapid development of China's higher education, higher education, popularization of increasingly rapid pace, the future number of people receiving higher education will grow exponentially. In the rapidly growing scale of higher education at the same time, the quality of higher education has been a cause for concern This paper analyzes the current problems in the study of college students exist, and improvement measures put forward their own proposals.
文摘With the rapid development of China's higher education, higher education, popularization of increasingly rapid pace, the future number of people receiving higher education will grow exponentially. In the rapidly growing scale of higher education at the same time, the quality of higher education has been a cause for concern~ This paper analyzes the current problems in the study of college students exist, and improvement measures put forward their own proposals.
文摘Using literature review,questionnaire survey,mathematical statistics and other research methods,combined with the previous research results of the research team,systematically sort out the problems of"learning"in the"learning,training,competition,and use"physical education curriculum,and how to solve the problems.The strategic research of the Chinese Academy of Sciences aims to promote the scientificization of"learning"in the teaching of physical education courses,thereby improving the quality and efficiency of physical education courses.
文摘Pupils learning ability is limited, so it is difficult to get good learning results in art learning by their own understanding and practice. And art knowledge has a long history, which is not as simple and easy to understand as it seems. This requires teachers to formulate teaching plans in combination with students physiological and psychological characteristics, and to improve students learning efficiency by means of cooperative learning. Cooperative learning can not only create a good teaching atmosphere, stimulate students interest in inquiry, provide a stage for students to show themselves and share their learning achievements, but also strengthen students sense of responsibility, stimulate students artistic thinking, improve their artistic accomplishment, and promote the progress and promotion of students cooperative learning ability.
文摘With the popularity of Computer Assisted Language Learning(CALL),autonomous language learning centers have been established in the universities throughout the country.However,there are many problems exist in the practice of the autonomous language learning center.This essay tries to discuss the problems and its countermeasures with an aim of improving College English learning and teaching by the example of autonomous language learning center in Nanyang Institute of Technology.
基金the National Key Research and Development Program of China(2021ZD0112302)the National Natural Science Foundation of China(62222301,61890930-5,62021003)the Beijing Natural Science Foundation(JQ19013).
文摘This paper is concerned with a novel integrated multi-step heuristic dynamic programming(MsHDP)algorithm for solving optimal control problems.It is shown that,initialized by the zero cost function,MsHDP can converge to the optimal solution of the Hamilton-Jacobi-Bellman(HJB)equation.Then,the stability of the system is analyzed using control policies generated by MsHDP.Also,a general stability criterion is designed to determine the admissibility of the current control policy.That is,the criterion is applicable not only to traditional value iteration and policy iteration but also to MsHDP.Further,based on the convergence and the stability criterion,the integrated MsHDP algorithm using immature control policies is developed to accelerate learning efficiency greatly.Besides,actor-critic is utilized to implement the integrated MsHDP scheme,where neural networks are used to evaluate and improve the iterative policy as the parameter architecture.Finally,two simulation examples are given to demonstrate that the learning effectiveness of the integrated MsHDP scheme surpasses those of other fixed or integrated methods.
文摘The vehicle routing problem(VRP)is a typical discrete combinatorial optimization problem,and many models and algorithms have been proposed to solve the VRP and its variants.Although existing approaches have contributed significantly to the development of this field,these approaches either are limited in problem size or need manual intervention in choosing parameters.To solve these difficulties,many studies have considered learning-based optimization(LBO)algorithms to solve the VRP.This paper reviews recent advances in this field and divides relevant approaches into end-to-end approaches and step-by-step approaches.We performed a statistical analysis of the reviewed articles from various aspects and designed three experiments to evaluate the performance of four representative LBO algorithms.Finally,we conclude the applicable types of problems for different LBO algorithms and suggest directions in which researchers can improve LBO algorithms.
基金Supported by the National Natural Science Foundation of China(No.U1633115)the Science and Technology Foundation of Beijing Municipal Commission of Education(No.KM201810005027)。
文摘With the rapid development of air transportation in recent years,airport operations have attracted a lot of attention.Among them,airport gate assignment problem(AGAP)has become a research hotspot.However,the real-time AGAP algorithm is still an open issue.In this study,a deep reinforcement learning based AGAP(DRL-AGAP)is proposed.The optimization object is to maximize the rate of flights assigned to fixed gates.The real-time AGAP is modeled as a Markov decision process(MDP).The state space,action space,value and rewards have been defined.The DRL-AGAP algorithm is evaluated via simulation and it is compared with the flight pre-assignment results of the optimization software Gurobiand Greedy.Simulation results show that the performance of the proposed DRL-AGAP algorithm is close to that of pre-assignment obtained by the Gurobi optimization solver.Meanwhile,the real-time assignment ability is ensured by the proposed DRL-AGAP algorithm due to the dynamic modeling and lower complexity.
文摘Objectives:Near misses happen more frequently than actual errors,and highlight system vulnerabilities without causing any harm,thus provide a safe space for organizational learning.Second-order problem solving behavior offers a new perspective to better understand how nurses promote learning from near misses to improve organizational outcomes.This study aimed to explore frontline nurses’perspectives on using second-order problem solving behavior in learning from near misses to improve patient safety.Methods:A qualitative exploratory study design was employed.This study was conducted in three tertiary hospitals in east China from June to November 2015.Purposive sampling was used to recruit 19 frontline nurses.Semi-structured interviews and a qualitative directed content analysis was undertaken using Crossan’s 4I Framework of Organizational Learning as a coding framework.Results:Second-order problem solving behavior,based on the 4I Framework of Organizational Learning,was referred to as being a leader in exposing near misses,pushing forward the cause analysis within limited capacity,balancing the active and passive role during improvement project,and promoting the continuous improvement with passion while feeling low-powered.Conclusions:4I Framework of Organizational Learning can be an underlying guide to enrich frontline nurses’role in promoting organizations to learn from near misses.In this study,nurses displayed their pivotal role in organizational learning from near misses by using second-order problem solving.However,additional knowledge,skills,and support are needed to maximize the application of second-order problem solving behavior when near misses are recognized.