MatX logo

High throughput chips for LLMs

MatX designs hardware tailored for the world’s best AI models: We dedicate every transistor to maximizing performance for large models.

Other products put large models and small models on equal footing; MatX makes no such compromises. For the world’s largest models, we deliver 10× more computing power, enabling AI labs to make models an order of magnitude smarter and more useful.

Our product

We focus on cost efficiency for high-volume pretraining and production inference for large models. This means:

What our product enables

With our hardware: