Skip to content

An infrastructure built using PyGame and OpenAI Gymnasium used to train robots within a social navigation context with a wide range of human motion models to simulate crowds of pedestrians.

License

Notifications You must be signed in to change notification settings

TommasoVandermeer/Social-Navigation-PyEnvs

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Social-Navigation-PyEnvs

A framework used to train robots within a social navigation context with a wide range of human motion models to simulate crowds of pedestrians.

Description

This repository contains a framework developed starting from CrowdNav [1] and Python-RVO2 [2] used to train and test learning-based algorithms for Social Navigation.

In order to simulate crowds of pedestrians the following models are implemented:

  • Social Force Model (SFM) [3] and its variations [4], [5]
  • Headed Social Force Model (HSFM) [6]
  • Optimal Reciprocal Collision Avoidance (ORCA) [7]

The CrowdNav module [1] includes the following reinforcement learning algorithms for social robot navigation:

  • Collision Avoidance with Deep RL (CADRL) [8]
  • Long-short term memory RL (LSTM-RL) [9]
  • Social Attentive RL (SARL) [10]

The simulator is built upon Pygame in order to provide a functional visualization tool and OpenAI Gym, which defines the standard API for RL environments. It also implements a laser sensor and a differential drive robot, which allow users to develop sensor-based algorithms.

   
social-nav-overview-1 social-nav-overview-2

Simulation videos: a comparative study of human motion models in reinforcement learning algorithms for social robot navigation

Cite this paper

If this repository or paper turns out to be useful for your research, please cite our paper:

@article{10.1145/3746463,
author = {Van Der Meer, Tommaso and Garulli, Andrea and Giannitrapani, Antonio and Quartullo, Renato},
title = {A Comparative Study of Human Motion Models in Reinforcement Learning Algorithms for Social Robot Navigation},
year = {2025},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
url = {https://doi.org/10.1145/3746463},
doi = {10.1145/3746463},
abstract = {Social robot navigation is an evolving research field that aims to find efficient strategies to safely navigate dynamic environments populated by humans. A critical challenge in this domain is the accurate modeling of human motion, which directly impacts the design and evaluation of navigation algorithms. This paper presents a comparative study of two popular categories of human motion models used in social robot navigation, namely velocity-based models and force-based models. A system-theoretic representation of both model types is presented, which highlights their common feedback structure, although with different state variables. Several navigation policies based on reinforcement learning are trained and tested in various simulated environments involving pedestrian crowds modeled with these approaches. A comparative study is conducted to assess performance across multiple factors, including human motion model, navigation policy, scenario complexity and crowd density. The results highlight advantages and challenges of different approaches to modeling human behavior, as well as their role during training and testing of learning-based navigation policies. The findings offer valuable insights and guidelines for selecting appropriate human motion models when designing socially-aware robot navigation systems.},
note = {Just Accepted},
journal = {J. Hum.-Robot Interact.},
month = jun,
keywords = {Human Motion Models, Social Robot Navigation, Multi-Agent Systems, Motion Planning}
}

Human motion models

ORCA SFM HSFM
social-nav-overview-1 social-nav-overview-1 social-nav-overview-1

Baselines

BP SSP ORCA
social-nav-overview-1 social-nav-overview-1 social-nav-overview-1

Comparing policies

CADRL-HS-HSFM LSTM-RL-HS-HSFM SARL-HS-HSFM
social-nav-overview-1 social-nav-overview-1 social-nav-overview-1

Comparing scenarios

SARL-CC-HSFM SARL-PT-HSFM SARL-HS-HSFM
social-nav-overview-1 social-nav-overview-1 social-nav-overview-1

Comparing human motion models for training and testing

SARL-HS-ORCA SARL-HS-SFM SARL-HS-HSFM
social-nav-overview-1 social-nav-overview-1 social-nav-overview-1

Case study: evaluating scenarios with obstacles

SARL-CC-HSFM
SARL-CC-HSFM_with_static_obstacles

References

  • [1] CrowdNav.
  • [2] Python-RVO2.
  • [3] Helbing, D., Farkas, I., & Vicsek, T. (2000). Simulating dynamical features of escape panic. Nature, 407(6803), 487-490.
  • [4] Moussaïd, M., Helbing, D., Garnier, S., Johansson, A., Combe, M., & Theraulaz, G. (2009). Experimental study of the behavioural mechanisms underlying self-organization in human crowds. Proceedings of the Royal Society B: Biological Sciences, 276(1668), 2755-2762.
  • [5] Guo, R. Y. (2014). Simulation of spatial and temporal separation of pedestrian counter flow through a bottleneck. Physica A: Statistical Mechanics and its Applications, 415, 428-439.
  • [6] Farina, F., Fontanelli, D., Garulli, A., Giannitrapani, A., & Prattichizzo, D. (2017). Walking ahead: The headed social force model. PloS one, 12(1), e0169734.
  • [7] Van Den Berg, J., Snape, J., Guy, S. J., & Manocha, D. (2011, May). Reciprocal collision avoidance with acceleration-velocity obstacles. In 2011 IEEE International Conference on Robotics and Automation (pp. 3475-3482). IEEE.
  • [8] Chen, Y. F., Liu, M., Everett, M., & How, J. P. (2017, May). Decentralized non-communicating multiagent collision avoidance with deep reinforcement learning. In 2017 IEEE international conference on robotics and automation (ICRA) (pp. 285-292). IEEE.
  • [9] Everett, M., Chen, Y. F., & How, J. P. (2018, October). Motion planning among dynamic, decision-making agents with deep reinforcement learning. In 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (pp. 3052-3059). IEEE.
  • [10] Chen, C., Liu, Y., Kreiss, S., & Alahi, A. (2019, May). Crowd-robot interaction: Crowd-aware robot navigation with attention-based deep reinforcement learning. In 2019 international conference on robotics and automation (ICRA) (pp. 6015-6022). IEEE.

About

An infrastructure built using PyGame and OpenAI Gymnasium used to train robots within a social navigation context with a wide range of human motion models to simulate crowds of pedestrians.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages