{"id":"https://openalex.org/W7106544642","doi":"https://doi.org/10.48550/arxiv.2511.17103","title":"Bridging Visual Affective Gap: Borrowing Textual Knowledge by Learning from Noisy Image-Text Pairs","display_name":"Bridging Visual Affective Gap: Borrowing Textual Knowledge by Learning from Noisy Image-Text Pairs","publication_year":2025,"publication_date":"2025-11-21","ids":{"openalex":"https://openalex.org/W7106544642","doi":"https://doi.org/10.48550/arxiv.2511.17103"},"language":null,"primary_location":{"id":"doi:10.48550/arxiv.2511.17103","is_oa":true,"landing_page_url":"https://doi.org/10.48550/arxiv.2511.17103","pdf_url":null,"source":{"id":"https://openalex.org/S4306400194","display_name":"arXiv (Cornell University)","issn_l":null,"issn":null,"is_oa":true,"is_in_doaj":false,"is_core":false,"host_organization":"https://openalex.org/I205783295","host_organization_name":"Cornell University","host_organization_lineage":["https://openalex.org/I205783295"],"host_organization_lineage_names":[],"type":"repository"},"license":"cc-by","license_id":"https://openalex.org/licenses/cc-by","version":null,"is_accepted":false,"is_published":false,"raw_source_name":null,"raw_type":"article"},"type":"preprint","indexed_in":["datacite"],"open_access":{"is_oa":true,"oa_status":"green","oa_url":"https://doi.org/10.48550/arxiv.2511.17103","any_repository_has_fulltext":true},"authorships":[{"author_position":"first","author":{"id":null,"display_name":"Wu, Daiqing","orcid":null},"institutions":[],"countries":[],"is_corresponding":true,"raw_author_name":"Wu, Daiqing","raw_affiliation_strings":[],"affiliations":[]},{"author_position":"middle","author":{"id":null,"display_name":"Yang, Dongbao","orcid":null},"institutions":[],"countries":[],"is_corresponding":false,"raw_author_name":"Yang, Dongbao","raw_affiliation_strings":[],"affiliations":[]},{"author_position":"middle","author":{"id":null,"display_name":"Zhou, Yu","orcid":null},"institutions":[],"countries":[],"is_corresponding":false,"raw_author_name":"Zhou, Yu","raw_affiliation_strings":[],"affiliations":[]},{"author_position":"last","author":{"id":null,"display_name":"Ma, Can","orcid":null},"institutions":[],"countries":[],"is_corresponding":false,"raw_author_name":"Ma, Can","raw_affiliation_strings":[],"affiliations":[]}],"institutions":[],"countries_distinct_count":0,"institutions_distinct_count":4,"corresponding_author_ids":[],"corresponding_institution_ids":[],"apc_list":null,"apc_paid":null,"fwci":null,"has_fulltext":false,"cited_by_count":0,"citation_normalized_percentile":null,"cited_by_percentile_year":null,"biblio":{"volume":null,"issue":null,"first_page":null,"last_page":null},"is_retracted":false,"is_paratext":false,"is_xpac":false,"primary_topic":{"id":"https://openalex.org/T10667","display_name":"Emotion and Mood Recognition","score":0.5532000064849854,"subfield":{"id":"https://openalex.org/subfields/3205","display_name":"Experimental and Cognitive Psychology"},"field":{"id":"https://openalex.org/fields/32","display_name":"Psychology"},"domain":{"id":"https://openalex.org/domains/2","display_name":"Social Sciences"}},"topics":[{"id":"https://openalex.org/T10667","display_name":"Emotion and Mood Recognition","score":0.5532000064849854,"subfield":{"id":"https://openalex.org/subfields/3205","display_name":"Experimental and Cognitive Psychology"},"field":{"id":"https://openalex.org/fields/32","display_name":"Psychology"},"domain":{"id":"https://openalex.org/domains/2","display_name":"Social Sciences"}},{"id":"https://openalex.org/T10664","display_name":"Sentiment Analysis and Opinion Mining","score":0.296999990940094,"subfield":{"id":"https://openalex.org/subfields/1702","display_name":"Artificial Intelligence"},"field":{"id":"https://openalex.org/fields/17","display_name":"Computer Science"},"domain":{"id":"https://openalex.org/domains/3","display_name":"Physical Sciences"}},{"id":"https://openalex.org/T11714","display_name":"Multimodal Machine Learning Applications","score":0.042500000447034836,"subfield":{"id":"https://openalex.org/subfields/1707","display_name":"Computer Vision and Pattern Recognition"},"field":{"id":"https://openalex.org/fields/17","display_name":"Computer Science"},"domain":{"id":"https://openalex.org/domains/3","display_name":"Physical Sciences"}}],"keywords":[{"id":"https://openalex.org/keywords/leverage","display_name":"Leverage (statistics)","score":0.633400022983551},{"id":"https://openalex.org/keywords/bridging","display_name":"Bridging (networking)","score":0.6194999814033508},{"id":"https://openalex.org/keywords/perception","display_name":"Perception","score":0.5108000040054321},{"id":"https://openalex.org/keywords/exploit","display_name":"Exploit","score":0.5029000043869019},{"id":"https://openalex.org/keywords/visualization","display_name":"Visualization","score":0.4065000116825104},{"id":"https://openalex.org/keywords/visual-perception","display_name":"Visual perception","score":0.3490000069141388},{"id":"https://openalex.org/keywords/focus","display_name":"Focus (optics)","score":0.33889999985694885}],"concepts":[{"id":"https://openalex.org/C153083717","wikidata":"https://www.wikidata.org/wiki/Q6535263","display_name":"Leverage (statistics)","level":2,"score":0.633400022983551},{"id":"https://openalex.org/C41008148","wikidata":"https://www.wikidata.org/wiki/Q21198","display_name":"Computer science","level":0,"score":0.6330999732017517},{"id":"https://openalex.org/C174348530","wikidata":"https://www.wikidata.org/wiki/Q188635","display_name":"Bridging (networking)","level":2,"score":0.6194999814033508},{"id":"https://openalex.org/C26760741","wikidata":"https://www.wikidata.org/wiki/Q160402","display_name":"Perception","level":2,"score":0.5108000040054321},{"id":"https://openalex.org/C165696696","wikidata":"https://www.wikidata.org/wiki/Q11287","display_name":"Exploit","level":2,"score":0.5029000043869019},{"id":"https://openalex.org/C154945302","wikidata":"https://www.wikidata.org/wiki/Q11660","display_name":"Artificial intelligence","level":1,"score":0.4846000075340271},{"id":"https://openalex.org/C204321447","wikidata":"https://www.wikidata.org/wiki/Q30642","display_name":"Natural language processing","level":1,"score":0.424699991941452},{"id":"https://openalex.org/C180747234","wikidata":"https://www.wikidata.org/wiki/Q23373","display_name":"Cognitive psychology","level":1,"score":0.41609999537467957},{"id":"https://openalex.org/C36464697","wikidata":"https://www.wikidata.org/wiki/Q451553","display_name":"Visualization","level":2,"score":0.4065000116825104},{"id":"https://openalex.org/C15744967","wikidata":"https://www.wikidata.org/wiki/Q9418","display_name":"Psychology","level":0,"score":0.36309999227523804},{"id":"https://openalex.org/C178253425","wikidata":"https://www.wikidata.org/wiki/Q162668","display_name":"Visual perception","level":3,"score":0.3490000069141388},{"id":"https://openalex.org/C192209626","wikidata":"https://www.wikidata.org/wiki/Q190909","display_name":"Focus (optics)","level":2,"score":0.33889999985694885},{"id":"https://openalex.org/C108583219","wikidata":"https://www.wikidata.org/wiki/Q197536","display_name":"Deep learning","level":2,"score":0.33500000834465027},{"id":"https://openalex.org/C2779903281","wikidata":"https://www.wikidata.org/wiki/Q6888026","display_name":"Modalities","level":2,"score":0.322299987077713},{"id":"https://openalex.org/C59404180","wikidata":"https://www.wikidata.org/wiki/Q17013334","display_name":"Feature learning","level":2,"score":0.2750000059604645},{"id":"https://openalex.org/C50644808","wikidata":"https://www.wikidata.org/wiki/Q192776","display_name":"Artificial neural network","level":2,"score":0.26339998841285706},{"id":"https://openalex.org/C143110190","wikidata":"https://www.wikidata.org/wiki/Q5373787","display_name":"Emotional expression","level":2,"score":0.26249998807907104},{"id":"https://openalex.org/C142853389","wikidata":"https://www.wikidata.org/wiki/Q744778","display_name":"Association (psychology)","level":2,"score":0.25999999046325684},{"id":"https://openalex.org/C2779321571","wikidata":"https://www.wikidata.org/wiki/Q7936605","display_name":"Visual learning","level":2,"score":0.25589999556541443},{"id":"https://openalex.org/C188147891","wikidata":"https://www.wikidata.org/wiki/Q147638","display_name":"Cognitive science","level":1,"score":0.2522999942302704}],"mesh":[],"locations_count":1,"locations":[{"id":"doi:10.48550/arxiv.2511.17103","is_oa":true,"landing_page_url":"https://doi.org/10.48550/arxiv.2511.17103","pdf_url":null,"source":{"id":"https://openalex.org/S4306400194","display_name":"arXiv (Cornell University)","issn_l":null,"issn":null,"is_oa":true,"is_in_doaj":false,"is_core":false,"host_organization":"https://openalex.org/I205783295","host_organization_name":"Cornell University","host_organization_lineage":["https://openalex.org/I205783295"],"host_organization_lineage_names":[],"type":"repository"},"license":"cc-by","license_id":"https://openalex.org/licenses/cc-by","version":null,"is_accepted":false,"is_published":null,"raw_source_name":null,"raw_type":"article"}],"best_oa_location":{"id":"doi:10.48550/arxiv.2511.17103","is_oa":true,"landing_page_url":"https://doi.org/10.48550/arxiv.2511.17103","pdf_url":null,"source":{"id":"https://openalex.org/S4306400194","display_name":"arXiv (Cornell University)","issn_l":null,"issn":null,"is_oa":true,"is_in_doaj":false,"is_core":false,"host_organization":"https://openalex.org/I205783295","host_organization_name":"Cornell University","host_organization_lineage":["https://openalex.org/I205783295"],"host_organization_lineage_names":[],"type":"repository"},"license":"cc-by","license_id":"https://openalex.org/licenses/cc-by","version":null,"is_accepted":false,"is_published":false,"raw_source_name":null,"raw_type":"article"},"sustainable_development_goals":[{"id":"https://metadata.un.org/sdg/4","display_name":"Quality Education","score":0.7615332007408142}],"awards":[],"funders":[],"has_content":{"pdf":false,"grobid_xml":false},"content_urls":null,"referenced_works_count":0,"referenced_works":[],"related_works":[],"abstract_inverted_index":{"Visual":[0],"emotion":[1],"recognition":[2],"(VER)":[3],"is":[4,184],"a":[5],"longstanding":[6],"field":[7],"that":[8,164],"has":[9],"garnered":[10],"increasing":[11],"attention":[12],"with":[13],"the":[14,29,36,48,52,61,63,75,82,85,91,101,154,166,171],"advancement":[15],"of":[16,38,54,94,133,156,173],"deep":[17],"neural":[18],"networks.":[19],"Although":[20],"recent":[21],"studies":[22],"have":[23],"achieved":[24],"notable":[25],"improvements":[26],"by":[27],"leveraging":[28],"knowledge":[30,56,83],"embedded":[31],"within":[32],"pre-trained":[33,86,95,175],"visual":[34,96,176],"models,":[35],"lack":[37],"direct":[39],"association":[40],"between":[41,106],"factual-level":[42],"features":[43],"and":[44,67,103,108,115,135,148],"emotional":[45,65,92,104],"categories,":[46],"called":[47],"\"affective":[49,76,167],"gap\",":[50],"limits":[51],"applicability":[53],"pre-training":[55],"for":[57,141],"VER":[58],"tasks.":[59,181],"On":[60],"contrary,":[62],"explicit":[64],"expression":[66],"high":[68],"information":[69],"density":[70],"in":[71,110,178],"textual":[72,87],"modality":[73],"eliminate":[74],"gap\".":[77],"Therefore,":[78],"we":[79,127,151,162],"propose":[80,116],"borrowing":[81],"from":[84],"model":[88],"to":[89,122,129],"enhance":[90],"perception":[93],"models.":[97],"We":[98],"focus":[99],"on":[100,186],"factual":[102],"connections":[105],"images":[107],"texts":[109],"noisy":[111,157],"social":[112],"media":[113],"data,":[114],"Partitioned":[117],"Adaptive":[118],"Contrastive":[119],"Learning":[120],"(PACL)":[121],"leverage":[123],"these":[124],"connections.":[125],"Specifically,":[126],"manage":[128],"separate":[130],"different":[131],"types":[132],"samples":[134],"devise":[136],"distinct":[137],"contrastive":[138],"learning":[139],"strategies":[140],"each":[142],"type.":[143],"By":[144],"dynamically":[145],"constructing":[146],"negative":[147],"positive":[149],"pairs,":[150],"fully":[152],"exploit":[153],"potential":[155],"samples.":[158],"Through":[159],"comprehensive":[160],"experiments,":[161],"demonstrate":[163],"bridging":[165],"gap\"":[168],"significantly":[169],"improves":[170],"performance":[172],"various":[174],"models":[177],"downstream":[179],"emotion-related":[180],"Our":[182],"code":[183],"released":[185],"https://github.com/wdqqdw/PACL.":[187]},"counts_by_year":[],"updated_date":"2025-11-25T14:47:58.456640","created_date":"2025-11-25T00:00:00"}
