Logo des Repositoriums
  • English
  • Deutsch
Anmelden
Keine TU-ID? Klicken Sie hier für mehr Informationen.
  1. Startseite
  2. Publikationen
  3. Publikationen der Technischen Universität Darmstadt
  4. Zweitveröffentlichungen (aus DeepGreen)
  5. Entropy based blending of policies for multi-agent coexistence
 
  • Details
2025
Zweitveröffentlichung
Artikel
Verlagsversion

Entropy based blending of policies for multi-agent coexistence

File(s)
Download
Hauptpublikation
10458_2025_Article_9707.pdf
CC BY 4.0 International
Format: Adobe PDF
Size: 1.66 MB
TUDa URI
tuda/14001
URN
urn:nbn:de:tuda-tuprints-306919
DOI
10.26083/tuprints-00030691
Autor:innen
Rother, David
Herbert, Franziska ORCID 0000-0003-4191-9366
Kalter, Fabian
Koert, Dorothea ORCID 0000-0002-3571-6848
Pajarinen, Joni ORCID 0000-0003-4469-8191
Peters, Jan ORCID 0000-0002-5266-8091
Weisswange, Thomas H.
Kurzbeschreibung (Abstract)

Research on multi-agent interaction involving humans is still in its infancy. Most approaches have focused on environments with collaborative human behavior or a small, defined set of situations. When deploying robots in human-inhabited environments in the future, the diversity of interactions surpasses the capabilities of pre-trained collaboration models. ”Coexistence” environments, characterized by agents with varying or partially aligned objectives, present a unique challenge for robotic collaboration. Traditional reinforcement learning methods fall short in these settings. These approaches lack the flexibility to adapt to changing agent counts or task requirements without undergoing retraining. Moreover, existing models do not adequately support scenarios where robots should exhibit helpful behavior toward others without compromising their primary goals. To tackle this issue, we introduce a novel framework that decomposes interaction and task-solving into separate learning problems and blends the resulting policies at inference time using a goal inference model for task estimation. We create impact-aware agents and linearly scale the cost of training agents with the number of agents and available tasks. To this end, a weighting function blending action distributions for individual interactions with the original task action distribution is proposed. To support our claims we demonstrate that our framework scales in task and agent count across several environments and considers collaboration opportunities when present. The new learning paradigm opens the path to more complex multi-robot, multi-human interactions.

Freie Schlagworte

Reinforcement learnin...

Multi-agent systems

Policy blending

Maximum entropy

Cooperative intellige...

Sprache
Englisch
Fachbereich/-gebiet
20 Fachbereich Informatik > Intelligente Autonome Systeme
DDC
000 Allgemeines, Informatik, Informationswissenschaft > 004 Informatik
Institution
Universitäts- und Landesbibliothek Darmstadt
Ort
Darmstadt
Typ des Artikels
Wissenschaftlicher Artikel
Titel der Zeitschrift / Schriftenreihe
Autonomous Agents and Multi-Agent Systems
Jahrgang der Zeitschrift
39
Heftnummer der Zeitschrift
1
ISSN
1573-7454
Verlag
Springer
Ort der Erstveröffentlichung
Dordrecht
Publikationsjahr der Erstveröffentlichung
2025
Verlags-DOI
10.1007/s10458-025-09707-7
PPN
534933424
Artikel-ID
27
Ergänzende Ressourcen (Supplement)
https://github.com/DavidRother/cooking_zoo
https://github.com/DavidRother/lb-foraging

  • TUprints Leitlinien
  • Cookie-Einstellungen
  • Impressum
  • Datenschutzbestimmungen
  • Webseitenanalyse
Diese Webseite wird von der Universitäts- und Landesbibliothek Darmstadt (ULB) betrieben.