Skip to main content

Questions tagged [dqn]

The tag has no summary, but it has a tag wiki.

0 votes
0 answers
34 views

Let's say , after taking an action in an environment. I receive the next state, which contains certain KPIs/metrics. Let's say each KPI in that state vector of shape (1,n) is on a different scale. I ...
sandeep kumar's user avatar
0 votes
0 answers
26 views

There is a production floor with W workstations and N jobs with M operations( different processing times per operation ). A job is completed only if its M Operations are completed. Objective is to ...
ArchanaR's user avatar
3 votes
1 answer
274 views

Trying to pick up basics of reinforcement learning by self-study from some blogs and texts. Forgive me if the question is too basic and different bits that I understand are a bit messy, but even after ...
Della's user avatar
  • 465
1 vote
1 answer
77 views

I am new in the area of RL and currently trying to train an online DQN model. Can an online model overfit since its always learning? and how can I tell if that happens?
user125612's user avatar
1 vote
1 answer
239 views

In DQN, why not use target network to predict current state Q values, and not only next state q values? In doing a basic dq learning algorithm with nn from scratch, with replay memory, and minibatch ...
Lorenzo Tinfena's user avatar
0 votes
0 answers
101 views

I am new to deep reinforcement learning! I am following this code for my adaptation problem (doing actions) https://github.com/jaromiru/AI-blog/blob/master/CartPole-DQN.py I am wondering how I can ...
imen kanzali's user avatar
1 vote
0 answers
314 views

I'm doing a course on reinforcement learning, and one of our tasks is to implement an agent on the Lunar lander continuous V2 environment from openAI gym. In order to solve the continuous problem, I ...
user113367's user avatar
3 votes
2 answers
608 views

I created my custom environment in gym, which is a maze. I use a DQN model with ...
Marci's user avatar
  • 31

15 30 50 per page
1
2 3 4 5 6