{"685344":{"#nid":"685344","#data":{"type":"event","title":"PhD Proposal by Anna Pritchard","body":[{"value":"\u003Cp\u003EAnna Pritchard\u003Cbr\u003EBME PhD Proposal Presentation\u003Cbr\u003E\u003Cbr\u003E\u003Cstrong\u003EDate\u003C\/strong\u003E: 2025-10-07\u003Cbr\u003E\u003Cstrong\u003ETime\u003C\/strong\u003E: 4:00 pm\u003Cbr\u003E\u003Cstrong\u003ELocation \/ Meeting Link\u003C\/strong\u003E: HSRBII N600 \u0026amp; Zoom:https:\/\/emory.zoom.us\/j\/97349350711\u003Cbr\u003E\u003Cbr\u003E\u003Cstrong\u003ECommittee Members:\u003C\/strong\u003E\u003Cbr\u003EChethan Pandarinath, PhD (Advisor); Nicholas Au Yong, MD\/PhD; Lena Ting, PhD; Garrett Stanley, PhD; Vikash Gilja, PhD\u0026nbsp;\u003Cbr\u003E\u003Cbr\u003E\u003Cbr\u003E\u003Cstrong\u003ETitle\u003C\/strong\u003E: State decoding for seamless selection between multiple iBCI functions\u003Cbr\u003E\u003Cbr\u003E\u003Cstrong\u003EAbstract:\u003C\/strong\u003E\u003Cbr\u003EIntracortical brain-computer interfaces (iBCIs) have shown remarkable progress in restoring motor and communication functions for people with paralysis by translating brain activity into a user\u2019s intended action such as moving their hand or speaking. Previous demonstrations have included computer cursor use, typing, robotic hand control, and speech decoding, but most studies focused on a single one of these output functions. For iBCIs to be versatile and clinically viable, we must understand how to integrate multiple iBCI functions\u2013 specifically, how to infer which function the user intends to operate at a given time and how to enable seamless switching between them. As part of the BrainGate2 clinical trial (NCT00912041), we have integrated cursor and speech iBCI functions for computer control using constant cursor activation and an on-screen button to enable speech decoding. To scale this approach to a broader repertoire of functions and increase system usability, we are developing methods to gate relevant outputs (e.g. cursor velocities) according to the user\u2019s functional intent. In this work, I will investigate the feasibility of decoding functional intent from motor cortex activity across diverse hand and speech iBCI functions. First, I have implemented real-time intent decoding that allows iBCI users to naturally switch between cursor and speech control for computer use. Next, I will characterize how different hand-related functions (cursor, typing, robotic hand, and handwriting) are represented in the motor cortex, specifically isolating neural features that separate these functions for intent decoding. Finally, I will apply large-scale deep-learning models to real-time functional intent decoding, investigating data-efficient pretraining and fine-tuning strategies to minimize function- and user-specific data collection requirements. Together, this work will establish a foundation for iBCIs that allow natural, flexible control of multiple functions, moving the field beyond single-function restoration towards practical iBCI systems for daily life.\u003C\/p\u003E","summary":"","format":"limited_html"}],"field_subtitle":"","field_summary":[{"value":"\u003Cp\u003EState decoding for seamless selection between multiple iBCI functions\u003Cbr\u003E\u003Cbr\u003E\u0026nbsp;\u003C\/p\u003E","format":"limited_html"}],"field_summary_sentence":[{"value":"State decoding for seamless selection between multiple iBCI functions"}],"uid":"27707","created_gmt":"2025-09-29 14:04:05","changed_gmt":"2025-09-29 14:04:52","author":"Tatianna Richardson","boilerplate_text":"","field_publication":"","field_article_url":"","field_event_time":{"event_time_start":"2025-10-07T16:00:00-04:00","event_time_end":"2025-10-07T18:00:00-04:00","event_time_end_last":"2025-10-07T18:00:00-04:00","gmt_time_start":"2025-10-07 20:00:00","gmt_time_end":"2025-10-07 22:00:00","gmt_time_end_last":"2025-10-07 22:00:00","rrule":null,"timezone":"America\/New_York"},"location":"HSRBII N600 ","extras":[],"groups":[{"id":"221981","name":"Graduate Studies"}],"categories":[],"keywords":[{"id":"102851","name":"Phd proposal"}],"core_research_areas":[],"news_room_topics":[],"event_categories":[{"id":"1788","name":"Other\/Miscellaneous"}],"invited_audience":[{"id":"78771","name":"Public"}],"affiliations":[],"classification":[],"areas_of_expertise":[],"news_and_recent_appearances":[],"phone":[],"contact":[],"email":[],"slides":[],"orientation":[],"userdata":""}}}