{"id":"https://openalex.org/W7130704539","doi":"https://doi.org/10.48550/arxiv.2602.17009","title":"Action-Graph Policies: Learning Action Co-dependencies in Multi-Agent Reinforcement Learning","display_name":"Action-Graph Policies: Learning Action Co-dependencies in Multi-Agent Reinforcement Learning","publication_year":2026,"publication_date":"2026-02-19","ids":{"openalex":"https://openalex.org/W7130704539","doi":"https://doi.org/10.48550/arxiv.2602.17009"},"language":null,"primary_location":{"id":"pmh:doi:10.48550/arxiv.2602.17009","is_oa":true,"landing_page_url":null,"pdf_url":null,"source":{"id":"https://openalex.org/S4406922384","display_name":"Open MIND","issn_l":null,"issn":null,"is_oa":false,"is_in_doaj":false,"is_core":false,"host_organization":null,"host_organization_name":null,"host_organization_lineage":[],"host_organization_lineage_names":[],"type":"repository"},"license":"cc-by","license_id":"https://openalex.org/licenses/cc-by","version":"submittedVersion","is_accepted":false,"is_published":false,"raw_source_name":null,"raw_type":"Article"},"type":"preprint","indexed_in":["datacite"],"open_access":{"is_oa":true,"oa_status":"green","oa_url":null,"any_repository_has_fulltext":true},"authorships":[{"author_position":"first","author":{"id":"https://openalex.org/A5126447284","display_name":"Nikunj Gupta","orcid":null},"institutions":[],"countries":[],"is_corresponding":true,"raw_author_name":"Gupta, Nikunj","raw_affiliation_strings":[],"affiliations":[]},{"author_position":"middle","author":{"id":"https://openalex.org/A5009542299","display_name":"James Z. Hare","orcid":"https://orcid.org/0000-0002-3920-7442"},"institutions":[],"countries":[],"is_corresponding":false,"raw_author_name":"Hare, James Zachary","raw_affiliation_strings":[],"affiliations":[]},{"author_position":"middle","author":{"id":"https://openalex.org/A5126454951","display_name":"Jesse Milzman","orcid":null},"institutions":[],"countries":[],"is_corresponding":false,"raw_author_name":"Milzman, Jesse","raw_affiliation_strings":[],"affiliations":[]},{"author_position":"middle","author":{"id":"https://openalex.org/A5126521206","display_name":"Rajgopal Kannan","orcid":null},"institutions":[],"countries":[],"is_corresponding":false,"raw_author_name":"Kannan, Rajgopal","raw_affiliation_strings":[],"affiliations":[]},{"author_position":"last","author":{"id":"https://openalex.org/A5126483586","display_name":"Viktor Prasanna","orcid":null},"institutions":[],"countries":[],"is_corresponding":false,"raw_author_name":"Prasanna, Viktor","raw_affiliation_strings":[],"affiliations":[]}],"institutions":[],"countries_distinct_count":0,"institutions_distinct_count":5,"corresponding_author_ids":["https://openalex.org/A5126447284"],"corresponding_institution_ids":[],"apc_list":null,"apc_paid":null,"fwci":null,"has_fulltext":false,"cited_by_count":0,"citation_normalized_percentile":null,"cited_by_percentile_year":null,"biblio":{"volume":null,"issue":null,"first_page":null,"last_page":null},"is_retracted":false,"is_paratext":false,"is_xpac":false,"primary_topic":{"id":"https://openalex.org/T10462","display_name":"Reinforcement Learning in Robotics","score":0.7572000026702881,"subfield":{"id":"https://openalex.org/subfields/1702","display_name":"Artificial Intelligence"},"field":{"id":"https://openalex.org/fields/17","display_name":"Computer Science"},"domain":{"id":"https://openalex.org/domains/3","display_name":"Physical Sciences"}},"topics":[{"id":"https://openalex.org/T10462","display_name":"Reinforcement Learning in Robotics","score":0.7572000026702881,"subfield":{"id":"https://openalex.org/subfields/1702","display_name":"Artificial Intelligence"},"field":{"id":"https://openalex.org/fields/17","display_name":"Computer Science"},"domain":{"id":"https://openalex.org/domains/3","display_name":"Physical Sciences"}},{"id":"https://openalex.org/T12026","display_name":"Explainable Artificial Intelligence (XAI)","score":0.049800001084804535,"subfield":{"id":"https://openalex.org/subfields/1702","display_name":"Artificial Intelligence"},"field":{"id":"https://openalex.org/fields/17","display_name":"Computer Science"},"domain":{"id":"https://openalex.org/domains/3","display_name":"Physical Sciences"}},{"id":"https://openalex.org/T11273","display_name":"Advanced Graph Neural Networks","score":0.0406000018119812,"subfield":{"id":"https://openalex.org/subfields/1702","display_name":"Artificial Intelligence"},"field":{"id":"https://openalex.org/fields/17","display_name":"Computer Science"},"domain":{"id":"https://openalex.org/domains/3","display_name":"Physical Sciences"}}],"keywords":[{"id":"https://openalex.org/keywords/observability","display_name":"Observability","score":0.8205999732017517},{"id":"https://openalex.org/keywords/reinforcement-learning","display_name":"Reinforcement learning","score":0.7041000127792358},{"id":"https://openalex.org/keywords/action","display_name":"Action (physics)","score":0.5349000096321106},{"id":"https://openalex.org/keywords/graph","display_name":"Graph","score":0.45419999957084656},{"id":"https://openalex.org/keywords/policy-learning","display_name":"Policy learning","score":0.4043000042438507},{"id":"https://openalex.org/keywords/set","display_name":"Set (abstract data type)","score":0.38909998536109924}],"concepts":[{"id":"https://openalex.org/C36299963","wikidata":"https://www.wikidata.org/wiki/Q1369844","display_name":"Observability","level":2,"score":0.8205999732017517},{"id":"https://openalex.org/C97541855","wikidata":"https://www.wikidata.org/wiki/Q830687","display_name":"Reinforcement learning","level":2,"score":0.7041000127792358},{"id":"https://openalex.org/C41008148","wikidata":"https://www.wikidata.org/wiki/Q21198","display_name":"Computer science","level":0,"score":0.6510999798774719},{"id":"https://openalex.org/C2780791683","wikidata":"https://www.wikidata.org/wiki/Q846785","display_name":"Action (physics)","level":2,"score":0.5349000096321106},{"id":"https://openalex.org/C154945302","wikidata":"https://www.wikidata.org/wiki/Q11660","display_name":"Artificial intelligence","level":1,"score":0.5148000121116638},{"id":"https://openalex.org/C132525143","wikidata":"https://www.wikidata.org/wiki/Q141488","display_name":"Graph","level":2,"score":0.45419999957084656},{"id":"https://openalex.org/C2779436431","wikidata":"https://www.wikidata.org/wiki/Q30672407","display_name":"Policy learning","level":2,"score":0.4043000042438507},{"id":"https://openalex.org/C177264268","wikidata":"https://www.wikidata.org/wiki/Q1514741","display_name":"Set (abstract data type)","level":2,"score":0.38909998536109924},{"id":"https://openalex.org/C18555067","wikidata":"https://www.wikidata.org/wiki/Q8375051","display_name":"Joint (building)","level":2,"score":0.35440000891685486},{"id":"https://openalex.org/C119857082","wikidata":"https://www.wikidata.org/wiki/Q2539","display_name":"Machine learning","level":1,"score":0.3156999945640564},{"id":"https://openalex.org/C51823790","wikidata":"https://www.wikidata.org/wiki/Q504353","display_name":"Greedy algorithm","level":2,"score":0.3052000105381012},{"id":"https://openalex.org/C46686674","wikidata":"https://www.wikidata.org/wiki/Q466303","display_name":"Boosting (machine learning)","level":2,"score":0.2976999878883362},{"id":"https://openalex.org/C175154964","wikidata":"https://www.wikidata.org/wiki/Q380077","display_name":"Task analysis","level":3,"score":0.2596000134944916},{"id":"https://openalex.org/C2775924081","wikidata":"https://www.wikidata.org/wiki/Q55608371","display_name":"Control (management)","level":2,"score":0.25119999051094055}],"mesh":[],"locations_count":2,"locations":[{"id":"pmh:doi:10.48550/arxiv.2602.17009","is_oa":true,"landing_page_url":null,"pdf_url":null,"source":{"id":"https://openalex.org/S4406922384","display_name":"Open MIND","issn_l":null,"issn":null,"is_oa":false,"is_in_doaj":false,"is_core":false,"host_organization":null,"host_organization_name":null,"host_organization_lineage":[],"host_organization_lineage_names":[],"type":"repository"},"license":"cc-by","license_id":"https://openalex.org/licenses/cc-by","version":"submittedVersion","is_accepted":false,"is_published":false,"raw_source_name":null,"raw_type":"Article"},{"id":"doi:10.48550/arxiv.2602.17009","is_oa":true,"landing_page_url":"https://doi.org/10.48550/arxiv.2602.17009","pdf_url":null,"source":{"id":"https://openalex.org/S4306400194","display_name":"arXiv (Cornell University)","issn_l":null,"issn":null,"is_oa":true,"is_in_doaj":false,"is_core":false,"host_organization":"https://openalex.org/I205783295","host_organization_name":"Cornell University","host_organization_lineage":["https://openalex.org/I205783295"],"host_organization_lineage_names":[],"type":"repository"},"license":"cc-by","license_id":"https://openalex.org/licenses/cc-by","version":null,"is_accepted":false,"is_published":null,"raw_source_name":null,"raw_type":"article"}],"best_oa_location":{"id":"pmh:doi:10.48550/arxiv.2602.17009","is_oa":true,"landing_page_url":null,"pdf_url":null,"source":{"id":"https://openalex.org/S4406922384","display_name":"Open MIND","issn_l":null,"issn":null,"is_oa":false,"is_in_doaj":false,"is_core":false,"host_organization":null,"host_organization_name":null,"host_organization_lineage":[],"host_organization_lineage_names":[],"type":"repository"},"license":"cc-by","license_id":"https://openalex.org/licenses/cc-by","version":"submittedVersion","is_accepted":false,"is_published":false,"raw_source_name":null,"raw_type":"Article"},"sustainable_development_goals":[{"display_name":"Peace, Justice and strong institutions","id":"https://metadata.un.org/sdg/16","score":0.5940951704978943}],"awards":[],"funders":[],"has_content":{"pdf":false,"grobid_xml":false},"content_urls":null,"referenced_works_count":0,"referenced_works":[],"related_works":[],"abstract_inverted_index":{"Coordinating":[0],"actions":[1,29,98],"is":[2],"the":[3],"most":[4],"fundamental":[5],"form":[6],"of":[7],"cooperation":[8],"in":[9,146],"multi-agent":[10,148],"reinforcement":[11],"learning":[12],"(MARL).":[13],"Successful":[14],"decentralized":[15],"decision-making":[16],"often":[17],"depends":[18],"not":[19],"only":[20,135],"on":[21,26,72,120],"good":[22],"individual":[23],"actions,":[24],"but":[25],"selecting":[27],"compatible":[28],"across":[30],"agents":[31,67],"to":[32,68,89],"synchronize":[33],"behavior,":[34],"avoid":[35],"conflicts,":[36],"and":[37,93,127],"satisfy":[38],"global":[39,73],"constraints.":[40],"In":[41],"this":[42],"paper,":[43],"we":[44,61,77,113],"propose":[45],"Action":[46],"Graph":[47],"Policies":[48],"(AGP),":[49],"that":[50,65,79,99,115,140],"model":[51],"dependencies":[52],"among":[53],"agents'":[54],"available":[55],"action":[56,74],"choices.":[57],"It":[58],"constructs,":[59],"what":[60],"call,":[62],"\\textit{coordination":[63],"contexts},":[64],"enable":[66],"condition":[69],"their":[70],"decisions":[71],"dependencies.":[75],"Theoretically,":[76],"show":[78,114],"AGPs":[80],"induce":[81],"a":[82],"strictly":[83],"more":[84,102],"expressive":[85],"joint":[86,97],"policy":[87],"compared":[88],"fully":[90],"independent":[91],"policies":[92],"can":[94],"realize":[95],"coordinated":[96],"are":[100],"provably":[101],"optimal":[103],"than":[104],"greedy":[105],"execution":[106],"even":[107],"from":[108],"centralized":[109],"value-decomposition":[110],"methods.":[111],"Empirically,":[112],"AGP":[116,141],"achieves":[117],"80-95\\%":[118],"success":[119],"canonical":[121],"coordination":[122],"tasks":[123],"with":[124],"partial":[125],"observability":[126],"anti-coordination":[128],"penalties,":[129],"where":[130],"other":[131],"MARL":[132],"methods":[133],"reach":[134],"10-25\\%.":[136],"We":[137],"further":[138],"demonstrate":[139],"consistently":[142],"outperforms":[143],"these":[144],"baselines":[145],"diverse":[147],"environments.":[149]},"counts_by_year":[],"updated_date":"2026-04-04T16:13:02.066488","created_date":"2026-02-21T00:00:00"}
