This is a follow up to ExoMyo’s previous post
Team ExoMyo has been hard at work co-developing the software for suit (actuator control and pilot intent) as well as the novel hardware “exomyo” actuators, each alone very challenging problems. Since the ExoMyo suit will use the BCI’s EEG signals to control speed and direction (we can’t expect EMG signals below the neck for the pilot), we have been focused on creating a testing platform that can identify a user’s intent. We know that in any individual, we will pick up (identifying is another issue) communication to muscle groups to move. First, we will be looking to identifying simple “commands” like forward, back, left and right. (even though it will be more like “that way” and “this way”) using the EEG data along with machine learning algorithms. We are using a remote control car that can record the timing of action (in this case, moving the joystick which identifies direction) in sync with the EEG data.