Nalazite se na CroRIS probnoj okolini. Ovdje evidentirani podaci neće biti pohranjeni u Informacijskom sustavu znanosti RH. Ako je ovo greška, CroRIS produkcijskoj okolini moguće je pristupi putem poveznice www.croris.hr
izvor podataka: crosbi

Automated Aerial Suspended Cargo Delivery through Reinforcement Learning (CROSBI ID 212334)

Prilog u časopisu | izvorni znanstveni rad | međunarodna recenzija

Faust, Aleksandra ; Palunko, Ivana ; Cruz, Patricio ; Fierro, Rafael ; Tapia, Lydia Automated Aerial Suspended Cargo Delivery through Reinforcement Learning // Artificial intelligence, 247 (2017), 381-398. doi: 10.1016/j.artint.2014.11.009

Podaci o odgovornosti

Faust, Aleksandra ; Palunko, Ivana ; Cruz, Patricio ; Fierro, Rafael ; Tapia, Lydia

engleski

Automated Aerial Suspended Cargo Delivery through Reinforcement Learning

Cargo-bearing Unmanned aerial vehicles (UAVs) have tremendous potential to assist humans in food, medicine, and supply deliveries. For time-critical cargo delivery tasks, UAVs need to be able to navigate their environments and deliver suspended payloads with bounded load displacement. As a constraint balancing task for joint UAV- suspended load system dynamics, this task poses a challenge. This article presents a reinforcement learning approach to aerial cargo delivery tasks in environments with static obstacles. We first learn a minimal residual oscillations task policy in obstacle- free environments that find trajectories with minimized residual load displacement with a specifically designed feature vector for value function approximation. With insights of learning from the cargo delivery problem, we define a set of formal criteria for class of robotics problems where learning can occur in a simplified problem space and transfer to a broader problem space. Exploiting this property, we create a path tracking method that suppresses load displacement. As an extension to tasks in environments with static obstacles where the load displacement needs to be bounded throughout the trajectory, sampling-based motion planning generates collision-free paths. Next, a reinforcement learning agent transforms these paths into trajectories that maintain the bound on the load displacement while following the collision-free path in a timely manner. We verify the approach both in simulation and in experiments on quadrotor with suspended load.

Reinforcement learning ; aerial load transportation ; quadrotors

nije evidentirano

nije evidentirano

nije evidentirano

nije evidentirano

nije evidentirano

nije evidentirano

Podaci o izdanju

247

2017.

381-398

objavljeno

0004-3702

1872-7921

10.1016/j.artint.2014.11.009

Povezanost rada

Elektrotehnika, Računarstvo

Poveznice
Indeksiranost