rrt-algorithms icon indicating copy to clipboard operation
rrt-algorithms copied to clipboard

dwa_algorithm

Open ss48 opened this issue 1 year ago • 1 comments

ss48 avatar Aug 31 '24 09:08 ss48

Adding machine learning (ML) to improve the performance of the robot's routing and scheduling via using the information collected (such as path length, optimised path, computation time, etc.) to train a model that can predict better paths or tune the parameters for optimization. The basis implementation approach to incorporate ML are shortest path, least computation time.

State: The robot's current position, distance to the goal, distance to obstacles, etc. Action: Possible movements (e.g., velocity, yaw rate). Reward: A function that rewards shorter paths, lower computation time, avoiding obstacles, etc.

We used the algorithm - Deep Q-Networks (DQN) to create an environment class that encapsulates the state, actions, and rewards. This environment interacts with your RRT and DWA algorithms. The DQN will be responsible for learning the optimal policy based on the robot's state (current position, distance to the goal, distance to obstacles, etc.), actions (possible movements), and rewards (based on path length, computation time, avoiding obstacles, etc.). Distance to the goal: Encourage the drone to get closer to the goal. Obstacle clearance: Penalize the drone for getting too close to obstacles. Path smoothness: Encourage smoother paths by penalizing sharp turns. Energy usage: Penalize higher energy consumption to encourage efficient paths.

ss48 avatar Aug 31 '24 10:08 ss48