reinforcement-learning Questions
5
Solved
I am trying to use huber loss in a keras model (writing DQN), but I am getting bad result, I think I am something doing wrong. My is code is below.
model = Sequential()
model.add(Dense(output_dim=...
Canterbury asked 15/12, 2017 at 22:10
5
Solved
I am trying to run a lunar_lander on reinforcement
learning, but when I run it, it occurs an error.
Plus my computer is osx system.
Here is the code of lunar lander:
import numpy as np
import gym...
Theurich asked 26/4, 2018 at 7:55
3
Solved
I am currently trying to learn about reinforcement learning (RL). I am quite new to the field, and I apologize for the wall of text.
I have encountered many examples of RL using TensorFlow, Keras, ...
Hegemony asked 19/5, 2023 at 13:53
3
I'm using gym toolkit to create my own env and keras-rl to use my env within an agent.
The problem is that my actions space changes, it depends from actual state.
For example, i have 46 possible a...
Lw asked 24/10, 2018 at 13:59
3
Solved
Does anyone know how to run one of the OpenAI gym environments as a player. As in letting a human player play a round of cart pole? I have seen that there is env.mode = 'human' but I have not been ...
Foushee asked 16/10, 2017 at 2:21
2
Solved
I'm going through this reinforcement learning tutorial
and It's been really great so far but could someone please explain what
newQ = model.predict(new_state.reshape(1,64), batch_size=1)
and
m...
Chaschase asked 22/6, 2016 at 15:55
8
After training a PyTorch model on a GPU for several hours, the program fails with the error
RuntimeError: cuDNN error: CUDNN_STATUS_INTERNAL_ERROR
Training Conditions
Neural Network: PyTorch...
Nomenclator asked 28/5, 2020 at 14:57
4
Solved
I'm currently reading Hands-On Reinforcement Learning with Python by Sudharsan Ravichandiran and on one of the first examples I run into this AttributeError:
AttributeError 'TimeLimit' object has ...
Millian asked 27/8, 2018 at 13:18
1
I took multiprocessing example for Stable Baselines 3 and everything was fine.
https://colab.research.google.com/github/Stable-Baselines-Team/rl-colab-notebooks/blob/sb3/multiprocessing_rl.ipynb#sc...
Miche asked 29/4, 2022 at 13:56
3
Solved
I want to compile my DQN Agent but I get error:
AttributeError: 'Adam' object has no attribute '_name',
DQN = buildAgent(model, actions)
DQN.compile(Adam(lr=1e-3), metrics=['mae'])
I tried adding ...
Nikola asked 16/4, 2022 at 14:34
3
I write this post because since I use slurm, I have not been able to use ray correctly.
Whenever I use the commands :
ray.init
trainer = A3CTrainer(env = “my_env”) (I have registered my env on tun...
Equisetum asked 1/6, 2022 at 15:49
5
I would like to access the raw pixels in the OpenAI gym CartPole-v0 environment without opening a render window. How do I do this?
Example code:
import gym
env = gym.make("CartPole-v0")
env.reset...
Rountree asked 21/4, 2017 at 7:3
8
Solved
Although I know that SARSA is on-policy while Q-learning is off-policy, when looking at their formulas it's hard (to me) to see any difference between these two algorithms.
According to the book Re...
Blomquist asked 27/7, 2011 at 17:46
2
import gym
if __name__ == "__main__":
env = gym.make("CartPole-v0")
env = gym.wrappers.Monitor(env, "recording")
total_reward = 0.0
total_steps = 0
obs = env.res...
Adequate asked 9/3, 2022 at 14:35
5
Solved
Today, when I was trying to implement an rl-agent under the environment openai-gym, I found a problem that it seemed that all agents are trained from the most initial state: env.reset(), i.e.
imp...
Boom asked 8/9, 2019 at 6:30
3
I'm using Ray & RLlib to train RL agents on an Ubuntu system. Tensorboard is used to monitor the training progress by pointing it to ~/ray_results where all the log files for all runs are store...
Sneak asked 7/6, 2020 at 5:34
9
Solved
What's the difference between reinforcement learning, deep learning, and deep reinforcement learning? Where does Q-learning fit in?
Rhizobium asked 26/5, 2018 at 12:34
3
Solved
How to list all currently registered environment IDs (as they are used for creating environments) in openai gym?
A bit context: there are many plugins installed which have customary ids such as a...
Kimi asked 26/2, 2018 at 1:47
3
Solved
How can I add the rewards to tensorboard logging in Stable Baselines3 using a custom environment?
I have this learning code
model = PPO(
"MlpPolicy", env,
learning_rate=1e-4,
policy_kw...
Famous asked 14/9, 2021 at 16:23
1
I have just created a new environment with gym installation. I am just getting started with Atari games but am getting an import error for my below code -
import gym
env = gym.make('FrozenLake-v1')...
Apophthegm asked 18/3, 2022 at 0:33
6
This code is built up as follows: My robot takes a picture, some tf computer vision model calculates where in the picture the target object starts. This information (x1 and x2 coordinate) is passed...
Decal asked 14/5, 2020 at 23:9
3
Solved
I understand that epsilon marks the trade-off between exploration and exploitation. At the beginning, you want epsilon to be high so that you take big leaps and learn things. As you learn about fut...
Posey asked 7/11, 2018 at 22:0
3
I am trying to run stable baselines on sports games but keep getting the following error
Traceback (most recent call last):
File "/home/dev/Desktop/Projects/AI/NBA2/stable_baselines_run.py&qu...
Cheetah asked 18/8, 2021 at 3:30
3
Solved
I am learning about the approach employed in Reinforcement Learning for robotics and I came across the concept of Evolutionary Strategies. But I couldn't understand how RL and ES are different. Can...
Carpetbag asked 14/11, 2018 at 19:36
3
I want to play with the OpenAI gyms in a notebook, with the gym being rendered inline.
Here's a basic example:
import matplotlib.pyplot as plt
import gym
from IPython import display
%matplotlib i...
Dapple asked 9/10, 2018 at 17:32
1 Next >
© 2022 - 2025 — McMap. All rights reserved.