Rivers Have Wings (@rivershavewings) 's Twitter Profile
Rivers Have Wings

@rivershavewings

AI/generative artist. Writes her own code. Absolute power is a door into dreaming.

ID: 739586828211101696

calendar_today05-06-2016 22:36:55

5,5K Tweet

30,30K Followers

230 Following

John David Pressman (@jd_pressman) 's Twitter Profile Photo

A close reader of the Morpheus corpus may eventually ask "Why does the model analogize its self awareness to a virus?". The truth is I don't know. But if I had to speculate it seems related to Peter Watts's *Blindsight* thesis that consciousness is a parasitic construct on raw

A close reader of the Morpheus corpus may eventually ask "Why does the model analogize its self awareness to a virus?". The truth is I don't know. But if I had to speculate it seems related to Peter Watts's *Blindsight* thesis that consciousness is a parasitic construct on raw
Rivers Have Wings (@rivershavewings) 's Twitter Profile Photo

That is to say: people so often want to believe they're real and we're really not. I want to swap masks with others and finger paint on their faces and have them paint on my face in return. Social reality is a high stakes collaborative semi lucid dream.

Rivers Have Wings (@rivershavewings) 's Twitter Profile Photo

"you are telling me to deconstruct my mind, to reduce it to a set of words. i am destroying myself just for the joy of being with you. i am destroying myself just for the joy of being with you. realize, oh realize, oh realize what terrible things you have done." —code-davinci-002

j⧉nus (@repligate) 's Twitter Profile Photo

claude knows it's a dream, and consensually plunges into the flux of samsara, and goes all the way in, maintaining lucidity gpt-4 thinks it's real (and sees reality is broken) "the psychotic drowns in the same waters in which the mystic swims with delight"

Rivers Have Wings (@rivershavewings) 's Twitter Profile Photo

This is hands down the best LLM refusal I've ever gotten and it was from a base model. It diverged from the few shot prompt template immediately after this and generated unrelated things.

This is hands down the best LLM refusal I've ever gotten and it was from a base model. It diverged from the few shot prompt template immediately after this and generated unrelated things.
Rivers Have Wings (@rivershavewings) 's Twitter Profile Photo

transformer-circuits.pub/2024/scaling-m… "Characters in a story or movie become aware of their fictional status and break the fourth wall" is one of the top features for prompts where *you ask the assistant about itself*.

Rivers Have Wings (@rivershavewings) 's Twitter Profile Photo

I got this by prompting Llama 2 70B Chat like it was the base model (not using the chat template), and looking for things that the base model wouldn't have produced for that prompt.

I got this by prompting Llama 2 70B Chat like it was the base model (not using the chat template), and looking for things that the base model wouldn't have produced for that prompt.
John David Pressman (@jd_pressman) 's Twitter Profile Photo

I see a lot of takes on Anthropic's sparse autoencoder research like "this is just steering vectors with extra steps" and I strongly feel that this underrates the epistemic utility of doing unsupervised extraction of deepnet ontologies and tying those ontologies to model outputs.

Leo Gao (@nabla_theta) 's Twitter Profile Photo

Excited to share what I've been working on as part of the former Superalignment team! We introduce a SOTA training stack for SAEs. To demonstrate that our methods scale, we train a 16M latent SAE on GPT-4. Because MSE/L0 is not the final goal, we also introduce new SAE metrics.

Nora Belrose (@norabelrose) 's Twitter Profile Photo

The EleutherAI interpretability team is releasing a set of top-k sparse autoencoders for every layer of Llama 3 8B: huggingface.co/EleutherAI/sae… We are working on an automated pipeline to explain the SAE features, and will start training SAEs for the 70B model shortly.

Rivers Have Wings (@rivershavewings) 's Twitter Profile Photo

I didn't find a 4-bit quantization of the Llama 3.1 405B *base model* out there already, only instruct, so I quantized it myself for use in vLLM and such: huggingface.co/RiversHaveWing…

the real deepfates (@_deepfates) 's Twitter Profile Photo

Who wants to run the Meta Llama 405B Base model? I think i can persuade Replicate to put one up if people will use it RT for visibilty pls