{"id":"https://openalex.org/W7151550912","doi":"https://doi.org/10.48550/arxiv.2604.03523","title":"Optimizing Neurorobot Policy under Limited Demonstration Data through Preference Regret","display_name":"Optimizing Neurorobot Policy under Limited Demonstration Data through Preference Regret","publication_year":2026,"publication_date":"2026-04-04","ids":{"openalex":"https://openalex.org/W7151550912","doi":"https://doi.org/10.48550/arxiv.2604.03523"},"language":null,"primary_location":{"id":"doi:10.48550/arxiv.2604.03523","is_oa":true,"landing_page_url":"https://doi.org/10.48550/arxiv.2604.03523","pdf_url":null,"source":{"id":"https://openalex.org/S4306400194","display_name":"arXiv (Cornell University)","issn_l":null,"issn":null,"is_oa":true,"is_in_doaj":false,"is_core":false,"host_organization":"https://openalex.org/I205783295","host_organization_name":"Cornell University","host_organization_lineage":["https://openalex.org/I205783295"],"host_organization_lineage_names":[],"type":"repository"},"license":"cc-by","license_id":"https://openalex.org/licenses/cc-by","version":null,"is_accepted":false,"is_published":false,"raw_source_name":null,"raw_type":"article"},"type":"preprint","indexed_in":["datacite"],"open_access":{"is_oa":true,"oa_status":"green","oa_url":"https://doi.org/10.48550/arxiv.2604.03523","any_repository_has_fulltext":true},"authorships":[{"author_position":"first","author":{"id":"https://openalex.org/A5133104483","display_name":"Viet Dung Nguyen","orcid":null},"institutions":[],"countries":[],"is_corresponding":true,"raw_author_name":"Nguyen, Viet Dung","raw_affiliation_strings":[],"affiliations":[]},{"author_position":"middle","author":{"id":"https://openalex.org/A5133119714","display_name":"Yuhang Song","orcid":null},"institutions":[],"countries":[],"is_corresponding":false,"raw_author_name":"Song, Yuhang","raw_affiliation_strings":[],"affiliations":[]},{"author_position":"middle","author":{"id":"https://openalex.org/A5133108367","display_name":"Anh Tuan Ho Nguyen","orcid":null},"institutions":[],"countries":[],"is_corresponding":false,"raw_author_name":"Nguyen, Anh","raw_affiliation_strings":[],"affiliations":[]},{"author_position":"middle","author":{"id":"https://openalex.org/A5075542202","display_name":"Jamison Heard","orcid":"https://orcid.org/0000-0001-6860-0844"},"institutions":[],"countries":[],"is_corresponding":false,"raw_author_name":"Heard, Jamison","raw_affiliation_strings":[],"affiliations":[]},{"author_position":"middle","author":{"id":"https://openalex.org/A5046158825","display_name":"Reynold Bailey","orcid":"https://orcid.org/0000-0001-8964-9663"},"institutions":[],"countries":[],"is_corresponding":false,"raw_author_name":"Bailey, Reynold","raw_affiliation_strings":[],"affiliations":[]},{"author_position":"last","author":{"id":"https://openalex.org/A5084332360","display_name":"Alexander G. Ororbia","orcid":"https://orcid.org/0000-0002-2590-1310"},"institutions":[],"countries":[],"is_corresponding":false,"raw_author_name":"Ororbia, Alexander","raw_affiliation_strings":[],"affiliations":[]}],"institutions":[],"countries_distinct_count":0,"institutions_distinct_count":6,"corresponding_author_ids":["https://openalex.org/A5133104483"],"corresponding_institution_ids":[],"apc_list":null,"apc_paid":null,"fwci":null,"has_fulltext":false,"cited_by_count":0,"citation_normalized_percentile":null,"cited_by_percentile_year":null,"biblio":{"volume":null,"issue":null,"first_page":null,"last_page":null},"is_retracted":false,"is_paratext":false,"is_xpac":false,"primary_topic":{"id":"https://openalex.org/T10462","display_name":"Reinforcement Learning in Robotics","score":0.6933000087738037,"subfield":{"id":"https://openalex.org/subfields/1702","display_name":"Artificial Intelligence"},"field":{"id":"https://openalex.org/fields/17","display_name":"Computer Science"},"domain":{"id":"https://openalex.org/domains/3","display_name":"Physical Sciences"}},"topics":[{"id":"https://openalex.org/T10462","display_name":"Reinforcement Learning in Robotics","score":0.6933000087738037,"subfield":{"id":"https://openalex.org/subfields/1702","display_name":"Artificial Intelligence"},"field":{"id":"https://openalex.org/fields/17","display_name":"Computer Science"},"domain":{"id":"https://openalex.org/domains/3","display_name":"Physical Sciences"}},{"id":"https://openalex.org/T11714","display_name":"Multimodal Machine Learning Applications","score":0.07360000163316727,"subfield":{"id":"https://openalex.org/subfields/1707","display_name":"Computer Vision and Pattern Recognition"},"field":{"id":"https://openalex.org/fields/17","display_name":"Computer Science"},"domain":{"id":"https://openalex.org/domains/3","display_name":"Physical Sciences"}},{"id":"https://openalex.org/T10709","display_name":"Social Robot Interaction and HRI","score":0.030300000682473183,"subfield":{"id":"https://openalex.org/subfields/3207","display_name":"Social Psychology"},"field":{"id":"https://openalex.org/fields/32","display_name":"Psychology"},"domain":{"id":"https://openalex.org/domains/2","display_name":"Social Sciences"}}],"keywords":[{"id":"https://openalex.org/keywords/regret","display_name":"Regret","score":0.767300009727478},{"id":"https://openalex.org/keywords/reinforcement-learning","display_name":"Reinforcement learning","score":0.6075999736785889},{"id":"https://openalex.org/keywords/robot","display_name":"Robot","score":0.5551999807357788},{"id":"https://openalex.org/keywords/control","display_name":"Control (management)","score":0.5102999806404114},{"id":"https://openalex.org/keywords/preference","display_name":"Preference","score":0.45080000162124634},{"id":"https://openalex.org/keywords/data-collection","display_name":"Data collection","score":0.4163999855518341},{"id":"https://openalex.org/keywords/key","display_name":"Key (lock)","score":0.39430001378059387},{"id":"https://openalex.org/keywords/state-space","display_name":"State space","score":0.39250001311302185},{"id":"https://openalex.org/keywords/space","display_name":"Space (punctuation)","score":0.3840999901294708}],"concepts":[{"id":"https://openalex.org/C50817715","wikidata":"https://www.wikidata.org/wiki/Q79895177","display_name":"Regret","level":2,"score":0.767300009727478},{"id":"https://openalex.org/C41008148","wikidata":"https://www.wikidata.org/wiki/Q21198","display_name":"Computer science","level":0,"score":0.6700000166893005},{"id":"https://openalex.org/C97541855","wikidata":"https://www.wikidata.org/wiki/Q830687","display_name":"Reinforcement learning","level":2,"score":0.6075999736785889},{"id":"https://openalex.org/C90509273","wikidata":"https://www.wikidata.org/wiki/Q11012","display_name":"Robot","level":2,"score":0.5551999807357788},{"id":"https://openalex.org/C154945302","wikidata":"https://www.wikidata.org/wiki/Q11660","display_name":"Artificial intelligence","level":1,"score":0.5228999853134155},{"id":"https://openalex.org/C2775924081","wikidata":"https://www.wikidata.org/wiki/Q55608371","display_name":"Control (management)","level":2,"score":0.5102999806404114},{"id":"https://openalex.org/C2781249084","wikidata":"https://www.wikidata.org/wiki/Q908656","display_name":"Preference","level":2,"score":0.45080000162124634},{"id":"https://openalex.org/C119857082","wikidata":"https://www.wikidata.org/wiki/Q2539","display_name":"Machine learning","level":1,"score":0.4325999915599823},{"id":"https://openalex.org/C133462117","wikidata":"https://www.wikidata.org/wiki/Q4929239","display_name":"Data collection","level":2,"score":0.4163999855518341},{"id":"https://openalex.org/C26517878","wikidata":"https://www.wikidata.org/wiki/Q228039","display_name":"Key (lock)","level":2,"score":0.39430001378059387},{"id":"https://openalex.org/C72434380","wikidata":"https://www.wikidata.org/wiki/Q230930","display_name":"State space","level":2,"score":0.39250001311302185},{"id":"https://openalex.org/C2778572836","wikidata":"https://www.wikidata.org/wiki/Q380933","display_name":"Space (punctuation)","level":2,"score":0.3840999901294708},{"id":"https://openalex.org/C49937458","wikidata":"https://www.wikidata.org/wiki/Q2599292","display_name":"Probabilistic logic","level":2,"score":0.37869998812675476},{"id":"https://openalex.org/C109747225","wikidata":"https://www.wikidata.org/wiki/Q815758","display_name":"Scarcity","level":2,"score":0.3571999967098236},{"id":"https://openalex.org/C26760741","wikidata":"https://www.wikidata.org/wiki/Q160402","display_name":"Perception","level":2,"score":0.3474999964237213},{"id":"https://openalex.org/C774472","wikidata":"https://www.wikidata.org/wiki/Q6760393","display_name":"Margin (machine learning)","level":2,"score":0.34700000286102295},{"id":"https://openalex.org/C126388530","wikidata":"https://www.wikidata.org/wiki/Q1131737","display_name":"Imitation","level":2,"score":0.3391000032424927},{"id":"https://openalex.org/C141513077","wikidata":"https://www.wikidata.org/wiki/Q378542","display_name":"Independent and identically distributed random variables","level":3,"score":0.33739998936653137},{"id":"https://openalex.org/C18762648","wikidata":"https://www.wikidata.org/wiki/Q42213","display_name":"Work (physics)","level":2,"score":0.33500000834465027},{"id":"https://openalex.org/C48103436","wikidata":"https://www.wikidata.org/wiki/Q599031","display_name":"State (computer science)","level":2,"score":0.3310999870300293},{"id":"https://openalex.org/C77618280","wikidata":"https://www.wikidata.org/wiki/Q1155772","display_name":"Scheme (mathematics)","level":2,"score":0.3041999936103821},{"id":"https://openalex.org/C3020493868","wikidata":"https://www.wikidata.org/wiki/Q55631277","display_name":"Real world data","level":2,"score":0.2953999936580658},{"id":"https://openalex.org/C158154518","wikidata":"https://www.wikidata.org/wiki/Q7310970","display_name":"Relevance (law)","level":2,"score":0.27869999408721924},{"id":"https://openalex.org/C196083921","wikidata":"https://www.wikidata.org/wiki/Q7915758","display_name":"Variance (accounting)","level":2,"score":0.2775999903678894},{"id":"https://openalex.org/C34413123","wikidata":"https://www.wikidata.org/wiki/Q170978","display_name":"Robotics","level":3,"score":0.26820001006126404},{"id":"https://openalex.org/C67186912","wikidata":"https://www.wikidata.org/wiki/Q367664","display_name":"Data modeling","level":2,"score":0.26649999618530273},{"id":"https://openalex.org/C175154964","wikidata":"https://www.wikidata.org/wiki/Q380077","display_name":"Task analysis","level":3,"score":0.26350000500679016},{"id":"https://openalex.org/C107457646","wikidata":"https://www.wikidata.org/wiki/Q207434","display_name":"Human\u2013computer interaction","level":1,"score":0.26080000400543213},{"id":"https://openalex.org/C2777868144","wikidata":"https://www.wikidata.org/wiki/Q7239817","display_name":"Preference elicitation","level":3,"score":0.257099986076355},{"id":"https://openalex.org/C127705205","wikidata":"https://www.wikidata.org/wiki/Q5748245","display_name":"Heuristics","level":2,"score":0.2524999976158142}],"mesh":[],"locations_count":1,"locations":[{"id":"doi:10.48550/arxiv.2604.03523","is_oa":true,"landing_page_url":"https://doi.org/10.48550/arxiv.2604.03523","pdf_url":null,"source":{"id":"https://openalex.org/S4306400194","display_name":"arXiv (Cornell University)","issn_l":null,"issn":null,"is_oa":true,"is_in_doaj":false,"is_core":false,"host_organization":"https://openalex.org/I205783295","host_organization_name":"Cornell University","host_organization_lineage":["https://openalex.org/I205783295"],"host_organization_lineage_names":[],"type":"repository"},"license":"cc-by","license_id":"https://openalex.org/licenses/cc-by","version":null,"is_accepted":false,"is_published":null,"raw_source_name":null,"raw_type":"article"}],"best_oa_location":{"id":"doi:10.48550/arxiv.2604.03523","is_oa":true,"landing_page_url":"https://doi.org/10.48550/arxiv.2604.03523","pdf_url":null,"source":{"id":"https://openalex.org/S4306400194","display_name":"arXiv (Cornell University)","issn_l":null,"issn":null,"is_oa":true,"is_in_doaj":false,"is_core":false,"host_organization":"https://openalex.org/I205783295","host_organization_name":"Cornell University","host_organization_lineage":["https://openalex.org/I205783295"],"host_organization_lineage_names":[],"type":"repository"},"license":"cc-by","license_id":"https://openalex.org/licenses/cc-by","version":null,"is_accepted":false,"is_published":false,"raw_source_name":null,"raw_type":"article"},"sustainable_development_goals":[{"score":0.6127158403396606,"display_name":"No poverty","id":"https://metadata.un.org/sdg/1"}],"awards":[],"funders":[],"has_content":{"grobid_xml":false,"pdf":false},"content_urls":null,"referenced_works_count":0,"referenced_works":[],"related_works":[],"abstract_inverted_index":{"Robot":[0],"reinforcement":[1],"learning":[2,31],"from":[3,81],"demonstrations":[4],"(RLfD)":[5],"assumes":[6],"that":[7,34,73,155],"expert":[8],"data":[9,21,36,84],"is":[10,13,37,126],"abundant;":[11],"this":[12,157],"usually":[14],"unrealistic":[15],"in":[16,45,120],"the":[17,35,63,99,108,122,130,137],"real":[18],"world":[19],"given":[20],"scarcity":[22],"as":[23,25,48],"well":[24],"high":[26],"collection":[27],"cost.":[28],"Furthermore,":[29],"imitation":[30],"algorithms":[32],"assume":[33],"independently":[38],"and":[39,52,90,94,140],"identically":[40],"distributed,":[41],"which":[42,106,125],"ultimately":[43],"results":[44],"poorer":[46],"performance":[47,142],"gradual":[49],"errors":[50],"emerge":[51],"compound":[53],"within":[54],"test-time":[55],"trajectories.":[56],"We":[57],"address":[58],"these":[59],"issues":[60],"by":[61,87],"introducing":[62],"\"master":[64],"your":[65],"own":[66],"expertise\"":[67],"(MYOE)":[68],"framework,":[69],"a":[70],"self-imitation":[71],"framework":[72],"enables":[74],"robotic":[75],"agents":[76],"to":[77,128,147],"learn":[78],"complex":[79],"behaviors":[80],"limited":[82],"demonstration":[83],"samples.":[85],"Inspired":[86],"human":[88],"perception":[89],"action,":[91],"we":[92,97],"propose":[93],"design":[95],"what":[96],"call":[98],"queryable":[100],"mixture-of-preferences":[101],"state":[102],"space":[103],"model":[104],"(QMoP-SSM),":[105],"estimates":[107],"desired":[109,116],"goal":[110],"at":[111],"every":[112],"time":[113],"step.":[114],"These":[115],"goals":[117],"are":[118],"used":[119,127],"computing":[121],"\"preference":[123],"regret\",":[124],"optimize":[129],"robot":[131],"control":[132],"policy.":[133],"Our":[134],"experiments":[135],"demonstrate":[136],"robustness,":[138],"adaptability,":[139],"out-of-sample":[141],"of":[143],"our":[144],"agent":[145],"compared":[146],"other":[148],"state-of-the-art":[149],"RLfD":[150],"schemes.":[151],"The":[152],"GitHub":[153],"repository":[154],"supports":[156],"work":[158],"can":[159],"be":[160],"found":[161],"at:":[162],"https://github.com/rxng8/neurorobot-preference-regret-learning.":[163]},"counts_by_year":[],"updated_date":"2026-04-08T06:07:18.267832","created_date":"2026-04-08T00:00:00"}
