The next generation of Meta’s large-scale infrastructure is
being built with AI in mind
, including supporting new generative AI (GenAI) products and services, recommendation systems, and advanced AI research. It’s an investment
we expect will grow
in the years ahead as the compute requirements to support AI models increase alongside the models’ sophistication.
Last year, we unveiled the
Meta Training and Inference Accelerator (MTIA) v1
, our first-generation AI inference accelerator that we designed in-house with Meta’s AI workloads in mind ? specifically our deep learning recommendation models that are improving a variety of experiences across our products.
MTIA is a long-term venture to provide the most efficient architecture for Meta’s unique workloads. As AI workloads become increasingly important to our products and services, this efficiency will improve our ability to provide the best experiences for our users around the world. MTIA v1 was an important step in improving the compute efficiency of our infrastructure and better supporting our software developers as they build AI models that will facilitate new and better user experiences.
Now, we’re sharing details about the next generation of MTIA.