Ashiq Rahman (@ashiq) 's Twitter Profile
Ashiq Rahman

@ashiq

Build a competitive advantage while minimizing risk with GenAI.
Advisor | Fractional CAIO | Data Science AI ML Technical Leader
Founder @ashiq_org

ID: 8708882

linkhttps://ashiq.org calendar_today06-09-2007 21:58:24

58 Tweet

124 Followers

973 Following

OpenAI (@openai) 's Twitter Profile Photo

In collaboration with Google, we're releasing Activation Atlases: a new technique for visualizing what interactions between neurons can represent. 💻Blog: blog.openai.com/introducing-ac… 📝Paper: distill.pub/2019/activatio… 🔤Code: github.com/tensorflow/luc… 🗺️Demo: distill.pub/2019/activatio…

François Chollet (@fchollet) 's Twitter Profile Photo

Are you a deep learning researcher? Wondering if all this TensorFlow 2.0 stuff you heard about is relevant to you? This thread is a crash course on everything you need to know to use TensorFlow 2.0 + Keras for deep learning research. Read on!

Are you a deep learning researcher? Wondering if all this TensorFlow 2.0 stuff you heard about is relevant to you?

This thread is a crash course on everything you need to know to use TensorFlow 2.0 + Keras for deep learning research. Read on!
Marcos López de Prado (@lopezdeprado) 's Twitter Profile Photo

A common misconception is that the risk of overfitting increases with the number of parameters in the model. In reality, a single parameter suffices to fit most datasets: arxiv.org/abs/1904.12320 Implementation available at: github.com/Ranlot/single-…

A common misconception is that the risk of overfitting increases with the number of parameters in the model. In reality, a single parameter suffices to fit most datasets: arxiv.org/abs/1904.12320

Implementation available at: github.com/Ranlot/single-…
Sean Coonce (@cooncesean) 's Twitter Profile Photo

My personal identity was hacked last week. The attacker was able to steal $100k+ in a sweep of my Coinbase account. I'm equal parts embarrassed, hurt, and deeply remorseful. In an effort to raise awareness about the attack, I wrote about it here: bit.ly/2VN5bZU

Communications of the ACM (@cacmmag) 's Twitter Profile Photo

"Datasheets for Datasets," by @TimnitGebru, @JamieMmt, @BrianaVecchione, @JennWVaughan, @HannaWallach, @HalDaume3, and @KateCrawford, proposes that every dataset be accompanied with a datasheet that documents its provenance, uses, etc. bit.ly/3oMFq7E

"Datasheets for Datasets," by @TimnitGebru, @JamieMmt, @BrianaVecchione, @JennWVaughan, @HannaWallach, @HalDaume3, and @KateCrawford, proposes that every dataset be accompanied with a datasheet that documents its provenance, uses, etc. bit.ly/3oMFq7E
Omar Sanseviero (@osanseviero) 's Twitter Profile Photo

Are you overwhelmed by everything happening in the ML ecosystem? We're doing a small crowdsourced initiative with a high-level distillation+timeline of cool big things happening in the ML landscape. Feel free to contribute! 🤗 github.com/osanseviero/ml…

Are you overwhelmed by everything happening in the ML ecosystem?

We're doing a small crowdsourced initiative with a high-level distillation+timeline of cool big things happening in the ML landscape. Feel free to contribute! 🤗

github.com/osanseviero/ml…
Anthropic (@anthropicai) 's Twitter Profile Photo

Large language models have demonstrated a surprising range of skills and behaviors. How can we trace their source? In our new paper, we use influence functions to find training examples that contribute to a given model output.

Large language models have demonstrated a surprising range of skills and behaviors. How can we trace their source? In our new paper, we use influence functions to find training examples that contribute to a given model output.
Jeff Dean (@jeffdean) 's Twitter Profile Photo

A nice post about the various techniques used to get a training job to scale to more than 50,000 TPU v5e chips on Google Cloud. cloud.google.com/blog/products/…

Greg Kamradt (@gregkamradt) 's Twitter Profile Photo

Claude 2.1 (200K Tokens) - Pressure Testing Long Context Recall We all love increasing context lengths - but what's performance like? Anthropic reached out with early access to Claude 2.1 so I repeated the “needle in a haystack” analysis I did on GPT-4 Here's what I found:

Claude 2.1 (200K Tokens) - Pressure Testing Long Context Recall

We all love increasing context lengths - but what's performance like?

Anthropic reached out with early access to Claude 2.1 so I repeated the “needle in a haystack” analysis I did on GPT-4

Here's what I found:
Srigi (@srigi) 's Twitter Profile Photo

"Latency Numbers Every Programmer Should Know" It is hard for humans to get the picture until you translate it to "human numbers":

"Latency Numbers Every Programmer Should Know"

It is hard for humans to get the picture until you translate it to "human numbers":