Carla-ppo icon indicating copy to clipboard operation
Carla-ppo copied to clipboard

This repository hosts a customized PPO based agent for Carla. The goal of this project is to make it easier to interact with and experiment in Carla with reinforcement learning based agents -- this, b...

Results 16 Carla-ppo issues
Sort by recently updated
recently updated
newest added

hi @bitsauce, in your model,the input state of RL agent is the vae output and measurement(steer,acc,break),but not include the high-level command,such as turn left,turn right,go straight,fllow the road.I want to...

ImportError: cannot import name 'naming'

Hello, I have attempted to implement your trained agent in maps other than Town07. It seems that the agent runs in the environment for a few seconds then the environment...

I would like to ask about the VAE training, I only used your RGB data set and got the following data roughly, but will the loss be too big. ![image](https://user-images.githubusercontent.com/76544411/148502749-87d682f9-481b-44f1-996e-a19541201d3e.png)...

Hi, wonderful and very interesting case. But ,why the car looks very slow when running with " the python run_eval.py --model_name pretrained_agent -start_carla"? Actually, the speed tells me about 15km/h....

Hello. When I run run_eval.py it shows the error "Exception: ${CARLA_ROOT} has not been set!" ![image](https://user-images.githubusercontent.com/82990411/135705158-464df84b-73d3-4ade-a213-ad07e865bba9.png) Now, I use win 10. I followed some instructions but could not fix it....

In the env reset ticks never increase more than zero and the loop runs forever without moving forward. Seems the wait_for _tick id not at all receiving any input back...

Hi! First,In ppo.py `self.policy = self.loss = -self.policy_loss + self.value_loss - self.entropy_loss` you said ' Reduce sum over all sub-policies (where only the active sub-policy will be non-zero due to...