Click for new scientific resources and news about Corona[COVID-19]

Paper Information

Journal:   NASHRIYYAH -I MUHANDISI -I BARQ VA MUHANDISI -I KAMPYUTAR -I IRAN, B- MUHANDISI -I KAMPYUTAR   SUMMER 2016 , Volume 14 , Number 2; Page(s) 137 To 146.
 
Paper: 

IMPROVING Q-LEARNING USING SIMULTANEOUS UPDATING AND ADAPTIVE POLICY BASED ON OPPOSITE ACTION

 
 
Author(s):  POUYAN M.*, GOLZARI S., MOUSAVI A., HATAM A.
 
* DEPT. OF ELEC. AND COMP. ENG., UNIVERSITY OF HORMOZGAN, BANDAR ABBAS, I.R. IRAN
 
Abstract: 

Q-learning is a one of the most popular and frequently used model-free reinforcement learning method. Among the advantages of this method is independent in its prior knowledge and there is a proof for its convergence to the optimal policy. One of the main limitations of this method is its low convergence speed, especially when the dimension is high. Accelerating convergence of this method is a challenge. Q-learning can be accelerated the convergence by the notion of opposite action. Since two Q-values are updated simultaneously at each learning step. In this paper, adaptive policy and the notion of opposite action are used to speed up the learning process by integrated approach. The methods are simulated for the grid world problem. The results demonstrate a great advance in the learning in terms of success rate, the percent of optimal states, the number of steps to goal, and average reward.

 
Keyword(s): ADAPTIVE POLICY, CONVERGENCE SPEED, OPPOSITE ACTION, SIMULTANEOUS UPDATING, Q-LEARNING
 
 
References: 
  • Not Registered.
  •  
  •  
 
Citations: 
  • Not Registered.
 
+ Click to Cite.
APA: Copy

POUYAN, M., & GOLZARI, S., & MOUSAVI, A., & HATAM, A. (2016). IMPROVING Q-LEARNING USING SIMULTANEOUS UPDATING AND ADAPTIVE POLICY BASED ON OPPOSITE ACTION. NASHRIYYAH -I MUHANDISI -I BARQ VA MUHANDISI -I KAMPYUTAR -I IRAN, B- MUHANDISI -I KAMPYUTAR, 14(2), 137-146. https://www.sid.ir/en/journal/ViewPaper.aspx?id=510938



Vancouver: Copy

POUYAN M., GOLZARI S., MOUSAVI A., HATAM A.. IMPROVING Q-LEARNING USING SIMULTANEOUS UPDATING AND ADAPTIVE POLICY BASED ON OPPOSITE ACTION. NASHRIYYAH -I MUHANDISI -I BARQ VA MUHANDISI -I KAMPYUTAR -I IRAN, B- MUHANDISI -I KAMPYUTAR. 2016 [cited 2021July27];14(2):137-146. Available from: https://www.sid.ir/en/journal/ViewPaper.aspx?id=510938



IEEE: Copy

POUYAN, M., GOLZARI, S., MOUSAVI, A., HATAM, A., 2016. IMPROVING Q-LEARNING USING SIMULTANEOUS UPDATING AND ADAPTIVE POLICY BASED ON OPPOSITE ACTION. NASHRIYYAH -I MUHANDISI -I BARQ VA MUHANDISI -I KAMPYUTAR -I IRAN, B- MUHANDISI -I KAMPYUTAR, [online] 14(2), pp.137-146. Available: https://www.sid.ir/en/journal/ViewPaper.aspx?id=510938.



 
 
Persian Abstract Yearly Visit 57
 
 
Latest on Blog
Enter SID Blog