Dragan, Theodora-AugustinaTheodora-AugustinaDraganKünzner, AlexanderAlexanderKünznerWille, RobertRobertWilleLorenz, Jeanette MiriamJeanette MiriamLorenz2025-04-072025-04-072025https://publica.fraunhofer.de/handle/publica/48629010.5220/0013371800003890One of the multiple facets of quantum reinforcement learning (QRL) is enhancing reinforcement learning (RL) algorithms with quantum submodules, namely with variational quantum circuits (VQC) as function approx-imators. QRL solutions are empirically proven to require fewer training iterations or adjustable parameters than their classical counterparts, but are usually restricted to applications that have a discrete action space and thus limited industrial relevance. We propose a hybrid quantum-classical (HQC) deep deterministic policy gradient (DDPG) approach for a robot to navigate through a maze using continuous states, continuous actions and using local observations from the robot’s LiDAR sensors. We show that this HQC method can lead to models of comparable test results to the neural network (NN)-based DDPG algorithm, that need around 200 times fewer weights. We also study the scalability of our solution with respect to the number of VQC layers and qubits, and find that in general results improve as the layer and qubit counts increase. The best rewards among all similarly sized HQC and classical DDPG methods correspond to a VQC of 8 qubits and 5 layers with no other NN. This work is another step towards continuous QRL, where literature has been sparse.enquantum reinforcement learningQRLLiDARlight detection and rangingrobot navigationQRLreinforcement learningRLContinuous Quantum Reinforcement Learning for Robot Navigationconference paper