{"id":"https://openalex.org/W4386160075","doi":"https://doi.org/10.1109/icarm58088.2023.10218817","title":"A Lightweight Dynamic Gesture Recognition Network with Spatio-Temporal Attention","display_name":"A Lightweight Dynamic Gesture Recognition Network with Spatio-Temporal Attention","publication_year":2023,"publication_date":"2023-07-08","ids":{"openalex":"https://openalex.org/W4386160075","doi":"https://doi.org/10.1109/icarm58088.2023.10218817"},"language":"en","primary_location":{"id":"doi:10.1109/icarm58088.2023.10218817","is_oa":false,"landing_page_url":"https://doi.org/10.1109/icarm58088.2023.10218817","pdf_url":null,"source":null,"license":null,"license_id":null,"version":"publishedVersion","is_accepted":true,"is_published":true,"raw_source_name":"2023 International Conference on Advanced Robotics and Mechatronics (ICARM)","raw_type":"proceedings-article"},"type":"article","indexed_in":["crossref"],"open_access":{"is_oa":false,"oa_status":"closed","oa_url":null,"any_repository_has_fulltext":false},"authorships":[{"author_position":"first","author":{"id":"https://openalex.org/A5083072513","display_name":"Xingyu Li","orcid":"https://orcid.org/0000-0001-6190-3164"},"institutions":[{"id":"https://openalex.org/I38877650","display_name":"Zhengzhou University","ror":"https://ror.org/04ypx8c21","country_code":"CN","type":"education","lineage":["https://openalex.org/I38877650"]}],"countries":["CN"],"is_corresponding":true,"raw_author_name":"Xingyu Li","raw_affiliation_strings":["School of Electrical and Information Engineering, Zhengzhou University,Zhengzhou,Henan Province,China,450001"],"affiliations":[{"raw_affiliation_string":"School of Electrical and Information Engineering, Zhengzhou University,Zhengzhou,Henan Province,China,450001","institution_ids":["https://openalex.org/I38877650"]}]},{"author_position":"middle","author":{"id":"https://openalex.org/A5100698159","display_name":"Lei Yang","orcid":"https://orcid.org/0000-0003-1212-9445"},"institutions":[{"id":"https://openalex.org/I38877650","display_name":"Zhengzhou University","ror":"https://ror.org/04ypx8c21","country_code":"CN","type":"education","lineage":["https://openalex.org/I38877650"]}],"countries":["CN"],"is_corresponding":false,"raw_author_name":"Lei Yang","raw_affiliation_strings":["School of Electrical and Information Engineering, Zhengzhou University,Zhengzhou,Henan Province,China,450001"],"affiliations":[{"raw_affiliation_string":"School of Electrical and Information Engineering, Zhengzhou University,Zhengzhou,Henan Province,China,450001","institution_ids":["https://openalex.org/I38877650"]}]},{"author_position":"last","author":{"id":"https://openalex.org/A5006721216","display_name":"Yanhong Liu","orcid":"https://orcid.org/0000-0002-7349-5871"},"institutions":[{"id":"https://openalex.org/I38877650","display_name":"Zhengzhou University","ror":"https://ror.org/04ypx8c21","country_code":"CN","type":"education","lineage":["https://openalex.org/I38877650"]}],"countries":["CN"],"is_corresponding":false,"raw_author_name":"Yanhong Liu","raw_affiliation_strings":["School of Electrical and Information Engineering, Zhengzhou University,Zhengzhou,Henan Province,China,450001"],"affiliations":[{"raw_affiliation_string":"School of Electrical and Information Engineering, Zhengzhou University,Zhengzhou,Henan Province,China,450001","institution_ids":["https://openalex.org/I38877650"]}]}],"institutions":[],"countries_distinct_count":1,"institutions_distinct_count":3,"corresponding_author_ids":["https://openalex.org/A5083072513"],"corresponding_institution_ids":["https://openalex.org/I38877650"],"apc_list":null,"apc_paid":null,"fwci":0.3453,"has_fulltext":false,"cited_by_count":2,"citation_normalized_percentile":{"value":0.58903424,"is_in_top_1_percent":false,"is_in_top_10_percent":false},"cited_by_percentile_year":{"min":95,"max":96},"biblio":{"volume":null,"issue":null,"first_page":"149","last_page":"154"},"is_retracted":false,"is_paratext":false,"is_xpac":false,"primary_topic":{"id":"https://openalex.org/T11398","display_name":"Hand Gesture Recognition Systems","score":1.0,"subfield":{"id":"https://openalex.org/subfields/1709","display_name":"Human-Computer Interaction"},"field":{"id":"https://openalex.org/fields/17","display_name":"Computer Science"},"domain":{"id":"https://openalex.org/domains/3","display_name":"Physical Sciences"}},"topics":[{"id":"https://openalex.org/T11398","display_name":"Hand Gesture Recognition Systems","score":1.0,"subfield":{"id":"https://openalex.org/subfields/1709","display_name":"Human-Computer Interaction"},"field":{"id":"https://openalex.org/fields/17","display_name":"Computer Science"},"domain":{"id":"https://openalex.org/domains/3","display_name":"Physical Sciences"}},{"id":"https://openalex.org/T10812","display_name":"Human Pose and Action Recognition","score":0.9991000294685364,"subfield":{"id":"https://openalex.org/subfields/1707","display_name":"Computer Vision and Pattern Recognition"},"field":{"id":"https://openalex.org/fields/17","display_name":"Computer Science"},"domain":{"id":"https://openalex.org/domains/3","display_name":"Physical Sciences"}},{"id":"https://openalex.org/T12740","display_name":"Gait Recognition and Analysis","score":0.9951000213623047,"subfield":{"id":"https://openalex.org/subfields/2204","display_name":"Biomedical Engineering"},"field":{"id":"https://openalex.org/fields/22","display_name":"Engineering"},"domain":{"id":"https://openalex.org/domains/3","display_name":"Physical Sciences"}}],"keywords":[{"id":"https://openalex.org/keywords/computer-science","display_name":"Computer science","score":0.83834308385849},{"id":"https://openalex.org/keywords/gesture","display_name":"Gesture","score":0.8239741325378418},{"id":"https://openalex.org/keywords/gesture-recognition","display_name":"Gesture recognition","score":0.7822392582893372},{"id":"https://openalex.org/keywords/artificial-intelligence","display_name":"Artificial intelligence","score":0.6438041925430298},{"id":"https://openalex.org/keywords/discriminative-model","display_name":"Discriminative model","score":0.6409429311752319},{"id":"https://openalex.org/keywords/convolutional-neural-network","display_name":"Convolutional neural network","score":0.6256938576698303},{"id":"https://openalex.org/keywords/feature-extraction","display_name":"Feature extraction","score":0.5698516964912415},{"id":"https://openalex.org/keywords/feature","display_name":"Feature (linguistics)","score":0.4893283247947693},{"id":"https://openalex.org/keywords/computer-vision","display_name":"Computer vision","score":0.4370349645614624},{"id":"https://openalex.org/keywords/pattern-recognition","display_name":"Pattern recognition (psychology)","score":0.43679729104042053},{"id":"https://openalex.org/keywords/speech-recognition","display_name":"Speech recognition","score":0.3736184537410736}],"concepts":[{"id":"https://openalex.org/C41008148","wikidata":"https://www.wikidata.org/wiki/Q21198","display_name":"Computer science","level":0,"score":0.83834308385849},{"id":"https://openalex.org/C207347870","wikidata":"https://www.wikidata.org/wiki/Q371174","display_name":"Gesture","level":2,"score":0.8239741325378418},{"id":"https://openalex.org/C159437735","wikidata":"https://www.wikidata.org/wiki/Q1519524","display_name":"Gesture recognition","level":3,"score":0.7822392582893372},{"id":"https://openalex.org/C154945302","wikidata":"https://www.wikidata.org/wiki/Q11660","display_name":"Artificial intelligence","level":1,"score":0.6438041925430298},{"id":"https://openalex.org/C97931131","wikidata":"https://www.wikidata.org/wiki/Q5282087","display_name":"Discriminative model","level":2,"score":0.6409429311752319},{"id":"https://openalex.org/C81363708","wikidata":"https://www.wikidata.org/wiki/Q17084460","display_name":"Convolutional neural network","level":2,"score":0.6256938576698303},{"id":"https://openalex.org/C52622490","wikidata":"https://www.wikidata.org/wiki/Q1026626","display_name":"Feature extraction","level":2,"score":0.5698516964912415},{"id":"https://openalex.org/C2776401178","wikidata":"https://www.wikidata.org/wiki/Q12050496","display_name":"Feature (linguistics)","level":2,"score":0.4893283247947693},{"id":"https://openalex.org/C31972630","wikidata":"https://www.wikidata.org/wiki/Q844240","display_name":"Computer vision","level":1,"score":0.4370349645614624},{"id":"https://openalex.org/C153180895","wikidata":"https://www.wikidata.org/wiki/Q7148389","display_name":"Pattern recognition (psychology)","level":2,"score":0.43679729104042053},{"id":"https://openalex.org/C28490314","wikidata":"https://www.wikidata.org/wiki/Q189436","display_name":"Speech recognition","level":1,"score":0.3736184537410736},{"id":"https://openalex.org/C138885662","wikidata":"https://www.wikidata.org/wiki/Q5891","display_name":"Philosophy","level":0,"score":0.0},{"id":"https://openalex.org/C41895202","wikidata":"https://www.wikidata.org/wiki/Q8162","display_name":"Linguistics","level":1,"score":0.0}],"mesh":[],"locations_count":1,"locations":[{"id":"doi:10.1109/icarm58088.2023.10218817","is_oa":false,"landing_page_url":"https://doi.org/10.1109/icarm58088.2023.10218817","pdf_url":null,"source":null,"license":null,"license_id":null,"version":"publishedVersion","is_accepted":true,"is_published":true,"raw_source_name":"2023 International Conference on Advanced Robotics and Mechatronics (ICARM)","raw_type":"proceedings-article"}],"best_oa_location":null,"sustainable_development_goals":[{"score":0.6499999761581421,"display_name":"Reduced inequalities","id":"https://metadata.un.org/sdg/10"}],"awards":[{"id":"https://openalex.org/G7196557517","display_name":null,"funder_award_id":"62003309","funder_id":"https://openalex.org/F4320320997","funder_display_name":"Funda\u00e7\u00e3o de Amparo \u00e0 Pesquisa do Estado de S\u00e3o Paulo"}],"funders":[{"id":"https://openalex.org/F4320320997","display_name":"Funda\u00e7\u00e3o de Amparo \u00e0 Pesquisa do Estado de S\u00e3o Paulo","ror":"https://ror.org/02ddkpn78"}],"has_content":{"pdf":false,"grobid_xml":false},"content_urls":null,"referenced_works_count":16,"referenced_works":["https://openalex.org/W581810680","https://openalex.org/W1522734439","https://openalex.org/W1985891935","https://openalex.org/W2004074725","https://openalex.org/W2097117768","https://openalex.org/W2194775991","https://openalex.org/W2219948256","https://openalex.org/W2511561981","https://openalex.org/W2607799870","https://openalex.org/W2770472008","https://openalex.org/W2786423589","https://openalex.org/W2963125010","https://openalex.org/W2963681914","https://openalex.org/W3138118886","https://openalex.org/W4297775537","https://openalex.org/W6737664043"],"related_works":["https://openalex.org/W2902873204","https://openalex.org/W2185750513","https://openalex.org/W2010878661","https://openalex.org/W3147379364","https://openalex.org/W2026258298","https://openalex.org/W3204639664","https://openalex.org/W2970836791","https://openalex.org/W2805039731","https://openalex.org/W2989699735","https://openalex.org/W4322710567"],"abstract_inverted_index":{"As":[0],"one":[1],"of":[2,9,47,60,68,83,148,154,160,200,212],"the":[3,44,58,78,146,152,174,198,201,220,232,247],"simplest":[4],"and":[5,21,25,33,63,156,177,206,243,252],"most":[6],"natural":[7],"ways":[8],"human-computer":[10,49],"interaction,":[11],"gestures":[12,70,123],"can":[13,230],"effectively":[14],"convey":[15],"information":[16,149,179],"through":[17],"silent":[18],"hand":[19],"movements":[20],"facilitate":[22],"our":[23],"production":[24],"life":[26],"in":[27,52,94,106,124,151],"combination":[28],"with":[29,237],"intelligent":[30,48],"devices.":[31],"Accurate":[32],"effective":[34,40,113],"gesture":[35,97,128,142,223,240],"recognition":[36,98,129,224,234,241],"technology":[37],"is":[38,136,169,194,249],"an":[39],"way":[41],"to":[42,57,111,118,138,171,184,196,203],"improve":[43,197],"efficient":[45],"work":[46],"interaction":[50],"devices":[51],"complex":[53,61],"environments.":[54],"However,":[55,100],"due":[56],"influence":[59],"surroundings":[62],"image":[64],"noise,":[65],"accurate":[66],"prediction":[67],"dynamic":[69,96,122,127,141,222],"still":[71,102],"faces":[72],"certain":[73,104],"challenges.":[74],"Recently,":[75],"benefiting":[76],"from":[77],"end-to-end":[79,140],"feature":[80,158,187],"extraction":[81,159],"capabilities":[82],"convolutional":[84],"neural":[85,167],"networks":[86],"(CNN),":[87],"deep":[88],"learning":[89,107],"has":[90,103],"achieved":[91],"good":[92],"results":[93],"different":[95],"tasks.":[99],"it":[101],"shortcomings":[105],"spatio-temporal":[108,133,190,228],"semantic":[109],"features":[110],"identify":[112],"regional":[114],"features.":[115,208],"In":[116,182],"order":[117,183],"accurately":[119],"recognize":[120],"continuous":[121],"video,":[125],"a":[126,163,189],"network":[130,168,202,225,248],"based":[131,226],"on":[132,214,227],"attention":[134,191,229],"mechanism":[135],"proposed":[137,195,221],"realize":[139],"recognition.":[143],"Aimed":[144],"at":[145],"problem":[147],"loss":[150],"process":[153],"spatial":[155,178,207],"temporal":[157,175,205],"video":[161],"frames,":[162],"3D":[164],"residual":[165],"convolution":[166],"used":[170],"simultaneously":[172],"model":[173],"relationship":[176],"between":[180],"frames.":[181],"obtain":[185,231],"discriminative":[186],"information,":[188],"module":[192],"(STA)":[193],"ability":[199],"extract":[204],"A":[209],"large":[210],"number":[211],"experiments":[213],"two":[215],"public":[216],"datasets":[217],"show":[218],"that":[219,246],"best":[233],"performance":[235],"compared":[236],"other":[238],"state-of-the-art":[239],"models,":[242],"further":[244],"prove":[245],"lightweight,":[250],"feasible":[251],"practical":[253],"method.":[254]},"counts_by_year":[{"year":2025,"cited_by_count":2}],"updated_date":"2025-11-06T03:46:38.306776","created_date":"2025-10-10T00:00:00"}
