{"id":"https://openalex.org/W4417088305","doi":"https://doi.org/10.48550/arxiv.2505.06601","title":"Learning Guarantee of Reward Modeling Using Deep Neural Networks","display_name":"Learning Guarantee of Reward Modeling Using Deep Neural Networks","publication_year":2025,"publication_date":"2025-05-10","ids":{"openalex":"https://openalex.org/W4417088305","doi":"https://doi.org/10.48550/arxiv.2505.06601"},"language":"en","primary_location":{"id":"pmh:oai:arXiv.org:2505.06601","is_oa":true,"landing_page_url":"http://arxiv.org/abs/2505.06601","pdf_url":"https://arxiv.org/pdf/2505.06601","source":{"id":"https://openalex.org/S4393918464","display_name":"ArXiv.org","issn_l":"2331-8422","issn":["2331-8422"],"is_oa":true,"is_in_doaj":false,"is_core":false,"host_organization":null,"host_organization_name":null,"host_organization_lineage":[],"host_organization_lineage_names":[],"type":"repository"},"license":null,"license_id":null,"version":"submittedVersion","is_accepted":false,"is_published":false,"raw_source_name":null,"raw_type":"text"},"type":"preprint","indexed_in":["arxiv","datacite"],"open_access":{"is_oa":true,"oa_status":"green","oa_url":"https://arxiv.org/pdf/2505.06601","any_repository_has_fulltext":true},"authorships":[{"author_position":"first","author":{"id":"https://openalex.org/A5011780561","display_name":"Yuanhang Luo","orcid":null},"institutions":[],"countries":[],"is_corresponding":true,"raw_author_name":"Luo, Yuanhang","raw_affiliation_strings":[],"affiliations":[]},{"author_position":"middle","author":{"id":"https://openalex.org/A5051771646","display_name":"Yeheng Ge","orcid":null},"institutions":[],"countries":[],"is_corresponding":false,"raw_author_name":"Ge, Yeheng","raw_affiliation_strings":[],"affiliations":[]},{"author_position":"middle","author":{"id":"https://openalex.org/A5028504100","display_name":"Ruijian Han","orcid":"https://orcid.org/0000-0002-9225-2218"},"institutions":[],"countries":[],"is_corresponding":false,"raw_author_name":"Han, Ruijian","raw_affiliation_strings":[],"affiliations":[]},{"author_position":"last","author":{"id":"https://openalex.org/A5058851479","display_name":"Guohao Shen","orcid":"https://orcid.org/0000-0002-5277-0176"},"institutions":[],"countries":[],"is_corresponding":false,"raw_author_name":"Shen, Guohao","raw_affiliation_strings":[],"affiliations":[]}],"institutions":[],"countries_distinct_count":0,"institutions_distinct_count":4,"corresponding_author_ids":["https://openalex.org/A5011780561"],"corresponding_institution_ids":[],"apc_list":null,"apc_paid":null,"fwci":null,"has_fulltext":false,"cited_by_count":0,"citation_normalized_percentile":null,"cited_by_percentile_year":null,"biblio":{"volume":null,"issue":null,"first_page":null,"last_page":null},"is_retracted":false,"is_paratext":false,"is_xpac":false,"primary_topic":{"id":"https://openalex.org/T10462","display_name":"Reinforcement Learning in Robotics","score":0.18389999866485596,"subfield":{"id":"https://openalex.org/subfields/1702","display_name":"Artificial Intelligence"},"field":{"id":"https://openalex.org/fields/17","display_name":"Computer Science"},"domain":{"id":"https://openalex.org/domains/3","display_name":"Physical Sciences"}},"topics":[{"id":"https://openalex.org/T10462","display_name":"Reinforcement Learning in Robotics","score":0.18389999866485596,"subfield":{"id":"https://openalex.org/subfields/1702","display_name":"Artificial Intelligence"},"field":{"id":"https://openalex.org/fields/17","display_name":"Computer Science"},"domain":{"id":"https://openalex.org/domains/3","display_name":"Physical Sciences"}},{"id":"https://openalex.org/T12101","display_name":"Advanced Bandit Algorithms Research","score":0.16850000619888306,"subfield":{"id":"https://openalex.org/subfields/1803","display_name":"Management Science and Operations Research"},"field":{"id":"https://openalex.org/fields/18","display_name":"Decision Sciences"},"domain":{"id":"https://openalex.org/domains/2","display_name":"Social Sciences"}},{"id":"https://openalex.org/T10203","display_name":"Recommender Systems and Techniques","score":0.1103999987244606,"subfield":{"id":"https://openalex.org/subfields/1710","display_name":"Information Systems"},"field":{"id":"https://openalex.org/fields/17","display_name":"Computer Science"},"domain":{"id":"https://openalex.org/domains/3","display_name":"Physical Sciences"}}],"keywords":[{"id":"https://openalex.org/keywords/regret","display_name":"Regret","score":0.8463000059127808},{"id":"https://openalex.org/keywords/pairwise-comparison","display_name":"Pairwise comparison","score":0.8097000122070312},{"id":"https://openalex.org/keywords/estimator","display_name":"Estimator","score":0.7249000072479248},{"id":"https://openalex.org/keywords/artificial-neural-network","display_name":"Artificial neural network","score":0.6068000197410583},{"id":"https://openalex.org/keywords/reinforcement-learning","display_name":"Reinforcement learning","score":0.5558000206947327},{"id":"https://openalex.org/keywords/deep-learning","display_name":"Deep learning","score":0.44909998774528503},{"id":"https://openalex.org/keywords/deep-neural-networks","display_name":"Deep neural networks","score":0.41839998960494995}],"concepts":[{"id":"https://openalex.org/C50817715","wikidata":"https://www.wikidata.org/wiki/Q79895177","display_name":"Regret","level":2,"score":0.8463000059127808},{"id":"https://openalex.org/C184898388","wikidata":"https://www.wikidata.org/wiki/Q1435712","display_name":"Pairwise comparison","level":2,"score":0.8097000122070312},{"id":"https://openalex.org/C185429906","wikidata":"https://www.wikidata.org/wiki/Q1130160","display_name":"Estimator","level":2,"score":0.7249000072479248},{"id":"https://openalex.org/C154945302","wikidata":"https://www.wikidata.org/wiki/Q11660","display_name":"Artificial intelligence","level":1,"score":0.6754999756813049},{"id":"https://openalex.org/C41008148","wikidata":"https://www.wikidata.org/wiki/Q21198","display_name":"Computer science","level":0,"score":0.6366000175476074},{"id":"https://openalex.org/C50644808","wikidata":"https://www.wikidata.org/wiki/Q192776","display_name":"Artificial neural network","level":2,"score":0.6068000197410583},{"id":"https://openalex.org/C119857082","wikidata":"https://www.wikidata.org/wiki/Q2539","display_name":"Machine learning","level":1,"score":0.5898000001907349},{"id":"https://openalex.org/C97541855","wikidata":"https://www.wikidata.org/wiki/Q830687","display_name":"Reinforcement learning","level":2,"score":0.5558000206947327},{"id":"https://openalex.org/C108583219","wikidata":"https://www.wikidata.org/wiki/Q197536","display_name":"Deep learning","level":2,"score":0.44909998774528503},{"id":"https://openalex.org/C2984842247","wikidata":"https://www.wikidata.org/wiki/Q197536","display_name":"Deep neural networks","level":3,"score":0.41839998960494995},{"id":"https://openalex.org/C2780791683","wikidata":"https://www.wikidata.org/wiki/Q846785","display_name":"Action (physics)","level":2,"score":0.39160001277923584},{"id":"https://openalex.org/C2775924081","wikidata":"https://www.wikidata.org/wiki/Q55608371","display_name":"Control (management)","level":2,"score":0.35670000314712524},{"id":"https://openalex.org/C120936955","wikidata":"https://www.wikidata.org/wiki/Q2155640","display_name":"Empirical research","level":2,"score":0.31349998712539673},{"id":"https://openalex.org/C107321475","wikidata":"https://www.wikidata.org/wiki/Q5374254","display_name":"Empirical risk minimization","level":2,"score":0.3012999892234802},{"id":"https://openalex.org/C67186912","wikidata":"https://www.wikidata.org/wiki/Q367664","display_name":"Data modeling","level":2,"score":0.2825999855995178},{"id":"https://openalex.org/C181204326","wikidata":"https://www.wikidata.org/wiki/Q7239820","display_name":"Preference learning","level":3,"score":0.26249998807907104}],"mesh":[],"locations_count":2,"locations":[{"id":"pmh:oai:arXiv.org:2505.06601","is_oa":true,"landing_page_url":"http://arxiv.org/abs/2505.06601","pdf_url":"https://arxiv.org/pdf/2505.06601","source":{"id":"https://openalex.org/S4393918464","display_name":"ArXiv.org","issn_l":"2331-8422","issn":["2331-8422"],"is_oa":true,"is_in_doaj":false,"is_core":false,"host_organization":null,"host_organization_name":null,"host_organization_lineage":[],"host_organization_lineage_names":[],"type":"repository"},"license":null,"license_id":null,"version":"submittedVersion","is_accepted":false,"is_published":false,"raw_source_name":null,"raw_type":"text"},{"id":"doi:10.48550/arxiv.2505.06601","is_oa":true,"landing_page_url":"https://doi.org/10.48550/arxiv.2505.06601","pdf_url":null,"source":{"id":"https://openalex.org/S4306400194","display_name":"arXiv (Cornell University)","issn_l":null,"issn":null,"is_oa":true,"is_in_doaj":false,"is_core":false,"host_organization":"https://openalex.org/I205783295","host_organization_name":"Cornell University","host_organization_lineage":["https://openalex.org/I205783295"],"host_organization_lineage_names":[],"type":"repository"},"license":"cc-by","license_id":"https://openalex.org/licenses/cc-by","version":null,"is_accepted":false,"is_published":null,"raw_source_name":null,"raw_type":"article"}],"best_oa_location":{"id":"pmh:oai:arXiv.org:2505.06601","is_oa":true,"landing_page_url":"http://arxiv.org/abs/2505.06601","pdf_url":"https://arxiv.org/pdf/2505.06601","source":{"id":"https://openalex.org/S4393918464","display_name":"ArXiv.org","issn_l":"2331-8422","issn":["2331-8422"],"is_oa":true,"is_in_doaj":false,"is_core":false,"host_organization":null,"host_organization_name":null,"host_organization_lineage":[],"host_organization_lineage_names":[],"type":"repository"},"license":null,"license_id":null,"version":"submittedVersion","is_accepted":false,"is_published":false,"raw_source_name":null,"raw_type":"text"},"sustainable_development_goals":[],"awards":[],"funders":[],"has_content":{"pdf":false,"grobid_xml":false},"content_urls":null,"referenced_works_count":0,"referenced_works":[],"related_works":[],"abstract_inverted_index":{"In":[0],"this":[1,101],"work,":[2],"we":[3,51],"study":[4],"the":[5,38,44,58,63,83,111,117],"learning":[6,126],"theory":[7],"of":[8,47,62,86,116],"reward":[9,28],"modeling":[10],"with":[11],"pairwise":[12,67,106],"comparison":[13,107],"data":[14,108],"using":[15],"deep":[16,27],"neural":[17],"networks.":[18],"We":[19],"establish":[20],"a":[21,31,53,77],"novel":[22],"non-asymptotic":[23],"regret":[24,79],"bound":[25],"for":[26],"estimators":[29,119],"in":[30,66,97],"non-parametric":[32],"setting,":[33],"which":[34,81],"depends":[35],"explicitly":[36],"on":[37],"network":[39],"architecture.":[40],"Furthermore,":[41],"to":[42,124],"underscore":[43],"critical":[45],"importance":[46],"clear":[48,94],"human":[49,95],"beliefs,":[50],"introduce":[52],"margin-type":[54,112],"condition":[55,75],"that":[56],"assumes":[57],"conditional":[59],"winning":[60],"probability":[61],"optimal":[64],"action":[65],"comparisons":[68],"is":[69,114],"significantly":[70],"distanced":[71],"from":[72,89,104],"1/2.":[73],"This":[74],"enables":[76],"sharper":[78],"bound,":[80],"substantiates":[82],"empirical":[84],"efficiency":[85],"Reinforcement":[87],"Learning":[88],"Human":[90],"Feedback":[91],"and":[92,121,128],"highlights":[93],"beliefs":[96],"its":[98],"success.":[99],"Notably,":[100],"improvement":[102],"stems":[103],"high-quality":[105],"implied":[109],"by":[110],"condition,":[113],"independent":[115],"specific":[118],"used,":[120],"thus":[122],"applies":[123],"various":[125],"algorithms":[127],"models.":[129]},"counts_by_year":[],"updated_date":"2026-03-10T16:38:18.471706","created_date":"2025-10-10T00:00:00"}
