Carla-ppo
Carla-ppo copied to clipboard
This repository hosts a customized PPO based agent for Carla. The goal of this project is to make it easier to interact with and experiment in Carla with reinforcement learning based agents -- this, b...
hi @bitsauce, in your model,the input state of RL agent is the vae output and measurement(steer,acc,break),but not include the high-level command,such as turn left,turn right,go straight,fllow the road.I want to...
ImportError: cannot import name 'naming'
Hello, I have attempted to implement your trained agent in maps other than Town07. It seems that the agent runs in the environment for a few seconds then the environment...
I would like to ask about the VAE training, I only used your RGB data set and got the following data roughly, but will the loss be too big. ...
Hi, wonderful and very interesting case. But ,why the car looks very slow when running with " the python run_eval.py --model_name pretrained_agent -start_carla"? Actually, the speed tells me about 15km/h....
Hello. When I run run_eval.py it shows the error "Exception: ${CARLA_ROOT} has not been set!"  Now, I use win 10. I followed some instructions but could not fix it....
In the env reset ticks never increase more than zero and the loop runs forever without moving forward. Seems the wait_for _tick id not at all receiving any input back...
Hi! First,In ppo.py `self.policy = self.loss = -self.policy_loss + self.value_loss - self.entropy_loss` you said ' Reduce sum over all sub-policies (where only the active sub-policy will be non-zero due to...