Positron AI (@positron_ai) 's Twitter Profile
Positron AI

@positron_ai

Developing the next generation of machine learning hardware and software

ID: 1689321438858395649

linkhttps://positron.ai calendar_today09-08-2023 17:03:26

25 Tweet

611 Followers

33 Following

Positron AI (@positron_ai) 's Twitter Profile Photo

When was the last time you saw a sushi chef use a Swiss Army Knife to slice your sashimi? Exactly—because they use knives precision-made for cutting sushi. So why, when you need fast, efficient “inference,” are you using inefficient multi-purpose GPUs built to do everything

Positron AI (@positron_ai) 's Twitter Profile Photo

The demand for inference is skyrocketing. That's why we focused on maximizing inference performance and designing the best memory bandwidth utilization ratio solution. Positron Atlas: the best AI accelerator designed exclusively and unapologetically for inference workloads. It's

Positron AI (@positron_ai) 's Twitter Profile Photo

Even the best-trained model fails if the deployment stack can’t keep up.  Most companies treat inference as an afterthought. They spend millions optimizing for training, then try to retrofit that same infrastructure to run real-world workloads, only to watch performance degrade

Positron AI (@positron_ai) 's Twitter Profile Photo

Deploy your trained models instantly—zero rewrites. GPUs lock you into their ecosystem, forcing complex integrations, endless compiler headaches, and frustrating delays. We built Positron Atlas differently: zero rewrites, zero friction. Upload your Nvidia-trained Hugging Face

Sally Ward-Foxton (@sallywf) 's Twitter Profile Photo

Groq Inc Cerebras I’m sure you know that GPUs often have famously poor memory bandwidth utilization for AI (in the order of 30%); with hardware optimization, it is perfectly possible to utilize more of that bandwidth than GPUs can/do (see: Positron AI) (7/x)

Positron AI (@positron_ai) 's Twitter Profile Photo

We’ve spent enough late nights fighting bloated GPUs to know something had to change. GPUs were a great starting point—but the chronic Nvidia shortages, massive power draw, and memory bottlenecks were killing our ability to deploy transformer models effectively at scale. We got

We’ve spent enough late nights fighting bloated GPUs to know something had to change.

GPUs were a great starting point—but the chronic Nvidia shortages, massive power draw, and memory bottlenecks were killing our ability to deploy transformer models effectively at scale.

We got
Edward Kmett (@kmett) 's Twitter Profile Photo

GPUs made training massive models possible, but inference needs better memory capacity, memory bandwidth utilization, more power efficiency, and an architecture built bottom up with transformers in mind. To that end, I'm excited to share that Positron just raised a $51.6M Series

Positron AI (@positron_ai) 's Twitter Profile Photo

Why we built Positron. Simple: To enable Superintelligence for everyone through the best performance-per-dollar and per-watt systems for generative AI inference. Positron’s first-generation product, Atlas, is shipping now, and we are happy to announce that Positron’s

Mitesh (@mitesh711) 's Twitter Profile Photo

We are excited to announce that Positron AI has raised $51.6M Series A led by Valor Equity Partners, Atreides Management, LP and DFJ Growth with participation from Resilience Reserve, Flume Ventures, 1517 Fund and Unless. We are looking forward to working closely with Gavin Baker ,

Thomas Sohmers (@trsohmers) 's Twitter Profile Photo

My startup, Positron AI just raised $51.6M Series A to rebuild the infrastructure powering AI inference and bring Superintelligence to everyone. Led by Valor Equity Partners (@valorep) , Atreides Management (Atreides Management, LP) , and DFJ Growth (DFJ Growth) —the same teams behind