TY - JOUR
T1 - Continuous Quantum Reinforcement Learning for Robot Navigation
AU - Drăgan, Theodora Augustina
AU - Künzner, Alexander
AU - Wille, Robert
AU - Lorenz, Jeanette Miriam
N1 - Publisher Copyright:
© 2025 by SCITEPRESS– Science and Technology Publications, Lda.
PY - 2025
Y1 - 2025
N2 - Oneofthemultiple facets of quantum reinforcement learning (QRL) is enhancing reinforcement learning (RL) algorithms with quantum submodules, namely with variational quantum circuits (VQC) as function approx imators. QRL solutions are empirically proven to require fewer training iterations or adjustable parameters than their classical counterparts, but are usually restricted to applications that have a discrete action space and thus limited industrial relevance. We propose a hybrid quantum-classical (HQC) deep deterministic policy gradient (DDPG) approach for a robot to navigate through a maze using continuous states, continuous actions and using local observations from the robot’s LiDAR sensors. We show that this HQC method can lead to models of comparable test results to the neural network (NN)-based DDPG algorithm, that need around 200 times fewer weights. We also study the scalability of our solution with respect to the number of VQC layers and qubits, and find that in general results improve as the layer and qubit counts increase. The best rewards among all similarly sized HQC and classical DDPG methods correspond to a VQC of 8 qubits and 5 layers with no other NN. This work is another step towards continuous QRL, where literature has been sparse.
AB - Oneofthemultiple facets of quantum reinforcement learning (QRL) is enhancing reinforcement learning (RL) algorithms with quantum submodules, namely with variational quantum circuits (VQC) as function approx imators. QRL solutions are empirically proven to require fewer training iterations or adjustable parameters than their classical counterparts, but are usually restricted to applications that have a discrete action space and thus limited industrial relevance. We propose a hybrid quantum-classical (HQC) deep deterministic policy gradient (DDPG) approach for a robot to navigate through a maze using continuous states, continuous actions and using local observations from the robot’s LiDAR sensors. We show that this HQC method can lead to models of comparable test results to the neural network (NN)-based DDPG algorithm, that need around 200 times fewer weights. We also study the scalability of our solution with respect to the number of VQC layers and qubits, and find that in general results improve as the layer and qubit counts increase. The best rewards among all similarly sized HQC and classical DDPG methods correspond to a VQC of 8 qubits and 5 layers with no other NN. This work is another step towards continuous QRL, where literature has been sparse.
KW - Continuous Action Space
KW - LiDAR
KW - Quantum Reinforcement Learning
KW - Robot Navigation
UR - http://www.scopus.com/inward/record.url?scp=105001710824&partnerID=8YFLogxK
U2 - 10.5220/0013371800003890
DO - 10.5220/0013371800003890
M3 - Conference article
AN - SCOPUS:105001710824
SN - 2184-3589
VL - 1
SP - 807
EP - 814
JO - International Conference on Agents and Artificial Intelligence
JF - International Conference on Agents and Artificial Intelligence
T2 - 17th International Conference on Agents and Artificial Intelligence, ICAART 2025
Y2 - 23 February 2025 through 25 February 2025
ER -