Hassan Shapourian (@hasan_shap) 's Twitter Profile
Hassan Shapourian

@hasan_shap

AI and Quantum. ex-Cisco, ex-Microsoft. Former postdoc at Harvard, MIT. Former student at Princeton, UIUC, and U.Chicago.

ID: 860783264

calendar_today04-10-2012 03:19:10

196 Tweet

432 Followers

451 Following

Rod Van Meter 🌻 (@rdviii) 's Twitter Profile Photo

Hi Sabine Hossenfelder, let's talk #QuantumInternet. We did this a few years ago, when I generated a tweetstorm about #QuantumComputerArchitecture, and I had been thinking it was about time to do something similar for quantum communications and Quantum Internet.

AK (@_akhaliq) 's Twitter Profile Photo

MMLU-Pro A More Robust and Challenging Multi-Task Language Understanding Benchmark In the age of large-scale language models, benchmarks like the Massive Multitask Language Understanding (MMLU) have been pivotal in pushing the boundaries of what AI can achieve

MMLU-Pro

A More Robust and Challenging Multi-Task Language Understanding Benchmark

In the age of large-scale language models, benchmarks like the Massive Multitask Language Understanding (MMLU) have been pivotal in pushing the boundaries of what AI can achieve
Zlatko Minev (@zlatko_minev) 's Twitter Profile Photo

What do graph theory, many-body physics, the golden ratio, and Fibonacci anyons have in common? In our experiment, arXiv link below, I’m excited how a very fundamental graph problem – buff.ly/3VADwr8 1/...

What do graph theory, many-body physics, the golden ratio, and Fibonacci anyons have in common?

In our experiment, arXiv link below, I’m excited how a very fundamental graph problem – 
buff.ly/3VADwr8
1/...
Pedram Roushan (@pedramroushan) 's Twitter Profile Photo

A major step in error correction Google Quantum AI : Pushing the surface code to the next level Below threshold: Distance-7 logical qubit, 0.0014 error per cycle, >2x better than physical qubits. arxiv.org/abs/2408.13687

A major step in error correction <a href="/GoogleQuantumAI/">Google Quantum AI</a> : Pushing the surface code to the next level 
Below threshold: Distance-7 logical qubit, 0.0014 error per cycle, &gt;2x better than physical qubits.

arxiv.org/abs/2408.13687
Maissam Barkeshli (@mbarkeshli) 's Twitter Profile Photo

The Nobel Committee recognizes profound contributions from Physics to ML / AI. There's a lot more where that came from. We are in an era where an increasing number of physicists are making important contributions to ML / AI, and even more are needed going forward.

Zyphra (@zyphraai) 's Twitter Profile Photo

Did you know that a leading open LLM dataset, DCLM, is ~80% duplicates? We discovered this while making Zyda2. Although performance seems fine on evals, downstream effects are less clear. So here are the 750B tokens of deduped, quality tokens from DCLM: huggingface.co/datasets/Zyphr…

Andrej Karpathy (@karpathy) 's Twitter Profile Photo

Remember exercise pages from textbooks? Large-scale collection of these across all realms of knowledge now moves billions of dollars. Textbooks written primarily for LLMs, compressed to weights, emergent solutions served to humans, or (over time) directly enacted for automation.

Remember exercise pages from textbooks? Large-scale collection of these across all realms of knowledge now moves billions of dollars. Textbooks written primarily for LLMs, compressed to weights, emergent solutions served to humans, or (over time) directly enacted for automation.
Aephraim Steinberg (@quantumaephraim) 's Twitter Profile Photo

Damned impressive, if taken at their word. "Logical computation demonstrated with a neutral atom quantum processor” [claiming up to 28 log. qubits in 256 phys., implementing real error-corrected algorithms] arxiv.org/abs/2411.11822 #Quantum #QuantumComputing #AtomComputing

Zyphra (@zyphraai) 's Twitter Profile Photo

We’ve been hard at work with AMD to optimize training for AMD GPUs. Today, we’re sharing a critical milestone towards this goal: FlashAttention-2 (FA2) and Mamba-2 backward kernels on AMD MI300X that surpass NVIDIA H100. We Zyphra are the first to achieve this.

We’ve been hard at work with <a href="/AMD/">AMD</a> to optimize training for AMD GPUs.

Today, we’re sharing a critical milestone towards this goal: FlashAttention-2 (FA2) and Mamba-2 backward kernels on AMD MI300X that surpass NVIDIA H100.

We <a href="/ZyphraAI/">Zyphra</a> are the first to achieve this.
Anthony Leverrier (@letonyo) 's Twitter Profile Photo

Oh this is great! Who said quantum computers had to work with qubits? Turns out it's possible to design a quantum algorithm for factoring that only requires 3 quantum oscillators and a single qubit! (1/4)

Zyphra (@zyphraai) 's Twitter Profile Photo

We @Zyphra previously described our preliminary RAG system that achieved SOTA performance on the HashHop long-context task. We are now excited to share our paper presenting a newer version of this RAG system that achieves SOTA results across multiple long-context benchmarks.

We @Zyphra previously described our preliminary RAG system that achieved SOTA performance on the HashHop long-context task.

We are now excited to share our paper presenting a newer version of this RAG system that achieves SOTA results across multiple long-context benchmarks.
Zyphra (@zyphraai) 's Twitter Profile Photo

Today, we're excited to announce a beta release of Zonos, a highly expressive TTS model with high fidelity voice cloning. We release both transformer and SSM-hybrid models under an Apache 2.0 license. Zonos performs well vs leading TTS providers in quality and expressiveness.

Condensed Matter Theory Center (@condensed_the) 's Twitter Profile Photo

~ 10K people are working in US quantum tech, this is ~ 5 years worth of US physics PhD production (not all are physics PhDs), estimated total expense/yr ~ $5 billion, income ~ zero How long can this continue? If/when quantum winter comes what would happen to these QC workers?

Zyphra (@zyphraai) 's Twitter Profile Photo

Zyphra is releasing our first reasoning model, ZR1-1.5B. This small but powerful reasoning model excels at both math and code, making it one of the best models in these categories for its size. It also uses 60% less reasoning tokens than comparable models. 🆓Apache 2.0 license.

Zyphra is releasing our first reasoning model, ZR1-1.5B. This small but powerful reasoning model excels at both math and code, making it one of the best models in these categories for its size. It also uses 60% less reasoning tokens than comparable models.

🆓Apache 2.0 license.
Andrej Karpathy (@karpathy) 's Twitter Profile Photo

Products with extensive/rich UIs lots of sliders, switches, menus, with no scripting support, and built on opaque, custom, binary formats are ngmi in the era of heavy human+AI collaboration. If an LLM can't read the underlying representations and manipulate them and all of the

Noam Brown (@polynoamial) 's Twitter Profile Photo

Today, we at OpenAI achieved a milestone that many considered years away: gold medal-level performance on the 2025 IMO with a general reasoning LLM—under the same time limits as humans, without tools. As remarkable as that sounds, it’s even more significant than the headline 🧵

Gabriele Berton (@gabriberton) 's Twitter Profile Photo

A year ago Ross Girschik (the object detection GOAT) gave a talk on "real tasks" (the end goal of an ML system) vs "fake tasks" (the intermediate tasks we created to achieve the real tasks). Most vision tasks, like classification and detection, are fake. (1/5)

A year ago Ross Girschik (the object detection GOAT) gave a talk on "real tasks" (the end goal of an ML system) vs "fake tasks" (the intermediate tasks we created to achieve the real tasks). Most vision tasks, like classification and detection, are fake. (1/5)