Media is too big
VIEW IN TELEGRAM
Introducing motor interface for amputee
That is the first AI model for decoding precise finger movements for people with hand amputation. It uses only 8 surface EMG electrodes.
ALVI Interface can decode different types of moves in virtual reality
🔘 finger flexion
🔘 finger extension
🟣 typing
🟣 some more
💎Full demo: YouTube link
Subscribe and follow the further progress of ALVI Labs:
Twitter: link
Instagram: link
That is the first AI model for decoding precise finger movements for people with hand amputation. It uses only 8 surface EMG electrodes.
ALVI Interface can decode different types of moves in virtual reality
💎Full demo: YouTube link
Subscribe and follow the further progress of ALVI Labs:
Twitter: link
Instagram: link
Please open Telegram to view this post
VIEW IN TELEGRAM
🔥16👍5🦄2👎1
Meet the new ALVI Interface: a breakthrough in intuitive prosthetic control.
This technology offers individuals with hand differences a new movement experience:
✨ Wrist rotation.
🖐 Finger movement.
🕹 Interaction with objects in VR.
Discover how we're turning futuristic dreams into today's reality. Be among the first to step into this new era of possibilities.
Our demo:
https://youtu.be/Dx_6Id2clZ0?si=jF9pX3u7tSiKobM5
Twitter/X
P. S. We are going to build a hand prosthesis powered by ALVI Interface. We need your support to do it.
This technology offers individuals with hand differences a new movement experience:
✨ Wrist rotation.
🖐 Finger movement.
🕹 Interaction with objects in VR.
Discover how we're turning futuristic dreams into today's reality. Be among the first to step into this new era of possibilities.
Our demo:
https://youtu.be/Dx_6Id2clZ0?si=jF9pX3u7tSiKobM5
Twitter/X
P. S. We are going to build a hand prosthesis powered by ALVI Interface. We need your support to do it.
🔥11❤🔥3🤗2👏1
the last neural cell pinned «Meet the new ALVI Interface: a breakthrough in intuitive prosthetic control. This technology offers individuals with hand differences a new movement experience: ✨ Wrist rotation. 🖐 Finger movement. 🕹 Interaction with objects in VR. Discover how we're…»
How to adapt MAE for multimodal data generation?
Background.
We know that MAE is very good approach to learn good features. It can be used for several modalities. For example, it outperforms MAE with trained on pairs text and image via transferable features: M3AE
Also there is MaskGit model which uses MAE for generation tasks. Shortly speaking it unmasks image step by step (aka diffusion but without diffusion).
And deep mind published paper: Muse(3 Jan 2023) in which they adapt mask git for text to image.
Very interesting idea 💡
So what if these approaches merge?
In my opinion using multimodal MAE for EMG and Motions concurrent prediction might be ... well promising.
How?
1) Just use M3AE pipeline and actually get better MAE features for Muscle.
2) Add MaskGit generation pipeline.
Expected results
Universal model which can unmask/generate 2 modalities.
- EMG -> Motion
- Motion -> EMG
- Better EMG features.
What do you think?☺️
P.S. In theory, it might be used for any data: ECoG, Spikes, EEG, fMRI and so on.
Background.
We know that MAE is very good approach to learn good features. It can be used for several modalities. For example, it outperforms MAE with trained on pairs text and image via transferable features: M3AE
Also there is MaskGit model which uses MAE for generation tasks. Shortly speaking it unmasks image step by step (aka diffusion but without diffusion).
And deep mind published paper: Muse(3 Jan 2023) in which they adapt mask git for text to image.
Very interesting idea 💡
Few words about my research)
I'm developing a prosthesis control system at ALVI Labs. My goal is to decode hand motions from muscle activity(EMG).
So what if these approaches merge?
In my opinion using multimodal MAE for EMG and Motions concurrent prediction might be ... well promising.
How?
1) Just use M3AE pipeline and actually get better MAE features for Muscle.
2) Add MaskGit generation pipeline.
Expected results
Universal model which can unmask/generate 2 modalities.
- EMG -> Motion
- Motion -> EMG
- Better EMG features.
What do you think?☺️
P.S. In theory, it might be used for any data: ECoG, Spikes, EEG, fMRI and so on.
arXiv.org
Multimodal Masked Autoencoders Learn Transferable Representations
Building scalable models to learn from diverse, multimodal data remains an open challenge. For vision-language data, the dominant approaches are based on contrastive learning objectives that train...
❤4
How to adapt MAE for multimodal data generation?
Architectures of M3AE and MaskGIT.
Architectures of M3AE and MaskGIT.
❤4
Recent advances in neural population decoding
In this recent paper (NeurIPS) the researchers adapted transformers for flexible decoding of large-scale spiking populations. The model POYO-1 was trained across sessions and participants (non-human primates), thus being effectively a good pre-trained model for whichever decoding paradigm you might want to use. Transfer learning across sessions and participants is enabled in a very "DL engineer" manner: let's just create an embedding of the session and process it along-side with spike embeddings - it seems to work well.
Unfortunately they did not release the code yet, but in the meantime you can check the website of a project.
One of the potential limitations is data was collected from M1, PMd and S1, which was shown to contain the task-relevant information, usually in the population vector coding format. Does the pre-trained transformer generalize to, for example, auditory cortex to decode imagined speech, remains unclear.
For me, seems like a hype paper, cannot wait until the code/weights are released to try it out. Let us know in the comments, which pitfalls you notice - it always feels good to debunk crowd decision making results 🙃
***
If you're rather interested in applying a model to any kind of data to infer low-dimensional trajectories representing the behavior check out CEBRA (paper). They allow you to use any kind of time series data (calcium imaging, spikes, voltages) along side with behavior. They also enable transfer learning and generalization across participants by explicitly encoding trial_ID, session_ID, subject_ID for the model.
👉 Here is the website: https://cebra.ai/
In this recent paper (NeurIPS) the researchers adapted transformers for flexible decoding of large-scale spiking populations. The model POYO-1 was trained across sessions and participants (non-human primates), thus being effectively a good pre-trained model for whichever decoding paradigm you might want to use. Transfer learning across sessions and participants is enabled in a very "DL engineer" manner: let's just create an embedding of the session and process it along-side with spike embeddings - it seems to work well.
Unfortunately they did not release the code yet, but in the meantime you can check the website of a project.
One of the potential limitations is data was collected from M1, PMd and S1, which was shown to contain the task-relevant information, usually in the population vector coding format. Does the pre-trained transformer generalize to, for example, auditory cortex to decode imagined speech, remains unclear.
For me, seems like a hype paper, cannot wait until the code/weights are released to try it out. Let us know in the comments, which pitfalls you notice - it always feels good to debunk crowd decision making results 🙃
***
If you're rather interested in applying a model to any kind of data to infer low-dimensional trajectories representing the behavior check out CEBRA (paper). They allow you to use any kind of time series data (calcium imaging, spikes, voltages) along side with behavior. They also enable transfer learning and generalization across participants by explicitly encoding trial_ID, session_ID, subject_ID for the model.
👉 Here is the website: https://cebra.ai/
arXiv.org
A Unified, Scalable Framework for Neural Population Decoding
Our ability to use deep learning approaches to decipher neural activity would likely benefit from greater scale, in terms of both model size and datasets. However, the integration of many neural...
🔥2❤1👍1
A collegue from Max Planck Institute in Tübingen talks about his foundational model of a cortical neuron, able to capture rich dynamics of real neurons.
https://twitter.com/OpenNeuroMorph/status/1762523010895081807?s=19
https://arxiv.org/abs/2306.16922
https://twitter.com/OpenNeuroMorph/status/1762523010895081807?s=19
https://arxiv.org/abs/2306.16922
X (formerly Twitter)
Open Neuromorphic (@OpenNeuroMorph) on X
The ELM Neuron: An Efficient and Expressive Cortical Neuron Model https://t.co/KQYjKPcZWn
🔥6❤1
Forwarded from Axis of Ordinary
DeepMind introduces SIMA: the first generalist AI agent to follow natural-language instructions in a broad range of 3D virtual environments and video games. 🕹 https://deepmind.google/discover/blog/sima-generalist-ai-agent-for-3d-virtual-environments/
"This research marks the first time an agent has demonstrated it can understand a broad range of gaming worlds, and follow natural-language instructions to carry out tasks within them, as a human might.
SIMA is an AI agent that can perceive and understand a variety of environments, then take actions to achieve an instructed goal. It comprises a model designed for precise image-language mapping and a video model that predicts what will happen next on-screen.
It requires just two inputs: the images on screen, and simple, natural-language instructions provided by the user.
The ability to function in brand new environments highlights SIMA’s ability to generalize beyond its training.
Our results also show that SIMA’s performance relies on language."
"This research marks the first time an agent has demonstrated it can understand a broad range of gaming worlds, and follow natural-language instructions to carry out tasks within them, as a human might.
SIMA is an AI agent that can perceive and understand a variety of environments, then take actions to achieve an instructed goal. It comprises a model designed for precise image-language mapping and a video model that predicts what will happen next on-screen.
It requires just two inputs: the images on screen, and simple, natural-language instructions provided by the user.
The ability to function in brand new environments highlights SIMA’s ability to generalize beyond its training.
Our results also show that SIMA’s performance relies on language."
❤3🔥2👍1
This media is not supported in your browser
VIEW IN TELEGRAM
Figure x OpenAI 2024
This status update just blows my mind. Really amazing progress. Especially the level of hand dexterity.
Figure combines a vision language model from openAI with a control system.
They did it in 13 days after signing a partnership🔥
For more details: X post
This status update just blows my mind. Really amazing progress. Especially the level of hand dexterity.
Figure combines a vision language model from openAI with a control system.
They did it in 13 days after signing a partnership🔥
For more details: X post
❤5🔥2
Grok 1 by X. 314B MoE.
Model Details
- Base model trained on a large amount of text data, not fine-tuned for any particular task.
- 314B parameter Mixture-of-Experts model with 25% of the weights active on a given token.
- Trained from scratch by xAI using a custom training stack on top of JAX and Rust in October 2023.
Link: https://x.ai/blog/grok-os
Model Details
- Base model trained on a large amount of text data, not fine-tuned for any particular task.
- 314B parameter Mixture-of-Experts model with 25% of the weights active on a given token.
- Trained from scratch by xAI using a custom training stack on top of JAX and Rust in October 2023.
Link: https://x.ai/blog/grok-os
❤3
#10 Summary
Humanoid Locomotion as Next Token Prediction
What.
They trained causal decoder to predict next action (and observation)
Data.
Normally, you'd need a bunch of data that shows both what the robot sees (observations) and what it does (actions). But that's tough to get . The authors used videos - some with the actions laid out and some without. This way, the robot can learn even from videos where we don't know what the actions were supposed to be.
In case there’re not action, they replace with [MASK] token. Very simple and straightforward
My thoughts
- I love how this paper makes the robot predict its next move and what it'll see next. It's like it's planning its future steps.
- For the robot to guess what's going to happen next accurately, it needs to have a mini understanding of physics and how the world works. This concept, called a 'world model,' is super intriguing.
- What's next? You can add condition with cross attention and train to understand commands, like VIMA paper.
More examples
Humanoid Locomotion as Next Token Prediction
What.
They trained causal decoder to predict next action (and observation)
Data.
Normally, you'd need a bunch of data that shows both what the robot sees (observations) and what it does (actions).
In case there’re not action, they replace with [MASK] token. Very simple and straightforward
My thoughts
- I love how this paper makes the robot predict its next move and what it'll see next. It's like it's planning its future steps.
- For the robot to guess what's going to happen next accurately, it needs to have a mini understanding of physics and how the world works. This concept, called a 'world model,' is super intriguing.
- What's next? You can add condition with cross attention and train to understand commands, like VIMA paper.
More examples
❤4
Brain-To-Text Competition 2024
This is the most fascinating BCI competition yet, organized by Stanford. Everyone has one month to develop the world's best brain-to-speech decoder!
Task: Predict attempted speech from brain activity.
Deadline: June 2, 2024
Dataset: They've recorded 12,100 sentences from a patient who can no longer speak intelligibly due to amyotrophic lateral sclerosis (ALS).
Just letting you know we're jumping into this challenge!
Together with @Altime, @kovalev_alvi and the team of ALVI Labs, we're going to create something interesting.
Like this post if you want to follow our updates❤️
This is the most fascinating BCI competition yet, organized by Stanford. Everyone has one month to develop the world's best brain-to-speech decoder!
Task: Predict attempted speech from brain activity.
Deadline: June 2, 2024
Dataset: They've recorded 12,100 sentences from a patient who can no longer speak intelligibly due to amyotrophic lateral sclerosis (ALS).
For each sentence, we provide the transcript of what the participant was attempting to say, along with the corresponding time series of neural spiking activity recorded from 256 microelectrodes in speech-related areas of cortex.
Just letting you know we're jumping into this challenge!
Together with @Altime, @kovalev_alvi and the team of ALVI Labs, we're going to create something interesting.
Like this post if you want to follow our updates❤️
👍9❤6🔥4🦄2
Neuralink Update - July 10, 2024
I watched their latest update and prepared some notes about it.
Link for the full
Control of the cursor.
Best cursor control without eye tracking.
How it works: participant imagines hand movements, translated to cursor control. I expect that they predict velocity of the cursor.
They achieved amazing speed and quality of cursor control on Day 133. For example, I couldn't beat this record for the first time. Let's try!
That's link: https://neuralink.com/webgrid/
My result: 7.88 BPS
Device usage up to 70 hours per week. That's almost half their waking life!
Future Plans and Goals:
Main goal is to give people "superpowers", not just restore functionality
Pretty damn cool
- Minimize device size, close gap between implant and brain. So they want to make device feel less external.
- Implant deeper for better signal diversity
- Decode text, enable faster typing (brain-to-text)
Holy Shit
- Develop visual prostheses (currently studying in monkeys)
- Repair paralysis (tested on animals with spinal cord implants)
- Use Optimus (Tesla) arms and legs, controlled by Neuralink implants
Future Implantation Process:
- Aim for fully automated implantation
- Quick procedure (10 minutes, "like cyberpunk")
My thoughts
In my view, the progress in hardware is truly exciting. I think that solving paralysis is possible; there are no restrictions from the laws of physics. However, as Elon Musk said, it's a super hard technical problem. There are plenty of papers on restoring movement and speech in the lab.
For example in brain-to-text competition, we also could decode sentences from brain signals. Really inspired story.
So stay tuned and cyborg might be real very soon.
I watched their latest update and prepared some notes about it.
Link for the full
Control of the cursor.
Best cursor control without eye tracking.
How it works: participant imagines hand movements, translated to cursor control. I expect that they predict velocity of the cursor.
They achieved amazing speed and quality of cursor control on Day 133. For example, I couldn't beat this record for the first time. Let's try!
That's link: https://neuralink.com/webgrid/
My result: 7.88 BPS
Device usage up to 70 hours per week. That's almost half their waking life!
Future Plans and Goals:
Main goal is to give people "superpowers", not just restore functionality
Pretty damn cool
- Minimize device size, close gap between implant and brain. So they want to make device feel less external.
- Implant deeper for better signal diversity
- Decode text, enable faster typing (brain-to-text)
Holy Shit
- Develop visual prostheses (currently studying in monkeys)
- Repair paralysis (tested on animals with spinal cord implants)
- Use Optimus (Tesla) arms and legs, controlled by Neuralink implants
Future Implantation Process:
- Aim for fully automated implantation
- Quick procedure (10 minutes, "like cyberpunk")
My thoughts
In my view, the progress in hardware is truly exciting. I think that solving paralysis is possible; there are no restrictions from the laws of physics. However, as Elon Musk said, it's a super hard technical problem. There are plenty of papers on restoring movement and speech in the lab.
For example in brain-to-text competition, we also could decode sentences from brain signals. Really inspired story.
So stay tuned and cyborg might be real very soon.
❤5🔥1🤔1🐳1
Cursor controlled by a brain implant.
Wireless, compatible with almost all devices using BLE.
You can even use it during a flight!
Wireless, compatible with almost all devices using BLE.
You can even use it during a flight!
🤯4🔥3👍1👻1