TRIBE v2 is a multimodal foundation model developed by Meta AI for predicting human brain activity from naturalistic stimuli such as video, audio, and text. It is designed for in-silico neuroscience, enabling researchers to model how the brain responds to complex real-world inputs. The system integrates state-of-the-art encoders—including LLaMA for text, V-JEPA for video, and Wav2Vec-BERT for audio—into a unified Transformer architecture. This combined representation is mapped onto the cortical surface to predict fMRI responses across thousands of brain regions. TRIBE v2 allows researchers to simulate and analyze brain activity without requiring direct human experiments. Overall, it provides a powerful tool for studying perception, cognition, and multimodal processing in the brain.

Features

  • Multimodal modeling of video, audio, and text for brain response prediction.
  • Transformer-based architecture mapping inputs to fMRI cortical activity.
  • Integration of advanced models like LLaMA, V-JEPA, and Wav2Vec-BERT.
  • Pretrained models available for inference on real-world media inputs.
  • Support for training and experimentation with neuroscience datasets.
  • Visualization tools for analyzing predicted brain activity across regions.

Project Activity

See All Activity >

Categories

AI Models

License

Creative Commons Attribution License

Follow TRIBE v2

TRIBE v2 Web Site

Other Useful Business Software
Gemini 3 and 200+ AI Models on One Platform Icon
Gemini 3 and 200+ AI Models on One Platform

Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

Build generative AI apps with Vertex AI. Switch between models without switching platforms.
Start Free
Rate This Project
Login To Rate This Project

User Reviews

Be the first to post a review of TRIBE v2!

Additional Project Details

Operating Systems

Linux, Mac, Windows

Registered

23 hours ago