Sam Bond-Taylor (@sambondtaylor) 's Twitter Profile
Sam Bond-Taylor

@sambondtaylor

Senior Researcher in Health Futures at @MSFTResearch. Previously PhD in deep generative models @comp_sci_durham. He/him.

ID: 1279160550753275906

linkhttps://samb-t.github.io calendar_today03-07-2020 21:10:33

47 Tweet

173 Followers

186 Following

AK (@_akhaliq) 's Twitter Profile Photo

RadEdit: stress-testing biomedical vision models via diffusion image editing paper page: huggingface.co/papers/2312.12… Biomedical imaging datasets are often small and biased, meaning that real-world performance of predictive models can be substantially lower than expected from

RadEdit: stress-testing biomedical vision models via diffusion image editing

paper page: huggingface.co/papers/2312.12…

Biomedical imaging datasets are often small and biased, meaning that real-world performance of predictive models can be substantially lower than expected from
Sam Bond-Taylor (@sambondtaylor) 's Twitter Profile Photo

Introducing RadEdit! We use diffusion image editing to stress-test biomedical vision models: diagnosing failures and helping avoid extra data collection. This is my first paper since joining Microsoft Research and started as an internship project with Pedro P. Sanchez & Boris van Breugel

Fernando Pérez-García (@fepegar_) 's Twitter Profile Photo

We know "all models are wrong", partially because "all datasets are biased", but how wrong and how biased are they? Check out our latest Microsoft Research preprint on stress-testing with diffusion models to learn more. "Causality matters in medical imaging"!

Chris Willcocks (@cwkx) 's Twitter Profile Photo

Our paper ∞-Diff: Infinite Resolution Diffusion... was accepted at #iclr2024 Its a diffusion model defined in an ∞-dimensional Hilbert space. Sample higher res than the training set (non-blurry). Paper: openreview.net/pdf?id=OUeIBFh… with Sam Bond-Taylor Comp Sci @ Durham #ai #research

Our paper ∞-Diff: Infinite Resolution Diffusion... was accepted at #iclr2024 Its a diffusion model defined in an ∞-dimensional Hilbert space. Sample higher res than the training set (non-blurry). Paper: openreview.net/pdf?id=OUeIBFh… with <a href="/sambondtaylor/">Sam Bond-Taylor</a> <a href="/comp_sci_durham/">Comp Sci @ Durham</a> #ai #research
Matthias Niessner (@mattniessner) 's Twitter Profile Photo

We are looking for interns & visiting researchers in our Visual Computing & AI Lab at TU Munich! Topics focus on Generative AI, NeRFs, Neural Scene Representations, Diffusion Models, LLMs, etc. Please share & apply by Feb 29th :) application.vc.in.tum.de Details: - We focus on

We are looking for interns &amp; visiting researchers in our Visual Computing &amp; AI Lab at TU Munich!

Topics focus on Generative AI, NeRFs, Neural Scene Representations, Diffusion Models, LLMs, etc.

Please share &amp; apply by Feb 29th :)
application.vc.in.tum.de

Details:
- We focus on
Ozan Oktay (@ozanoktay__) 's Twitter Profile Photo

If you are passionate about AI in healthcare and seeking an internship opportunity to work in our team, please get in touch with us. jobs.careers.microsoft.com/global/en/job/… #MachineLearning  #AIinHealthcare #medicalimaging #deeplearning #InternshipOpportunities

Sasha Rush (@srush_nlp) 's Twitter Profile Photo

Experiment: Triton-Autodiff (github.com/srush/triton-a…) Source-to-source autodiff of Triton GPU code. Uses tangent to produce working backward code you can edit. (Been writing Mamba in Triton and hate having to do this part manually)

Experiment: Triton-Autodiff (github.com/srush/triton-a…)

Source-to-source autodiff of Triton GPU code. Uses tangent to produce working backward code you can edit. 

(Been writing Mamba in Triton and hate having to do this part manually)
Sam Bond-Taylor (@sambondtaylor) 's Twitter Profile Photo

Rᴀᴅ-DINO is now on arXiv!🦖 We show that medical image encoders trained only on images perform similar or better than text supervised models on various benchmarks, demonstrating that reliance on text may not be necessary, and can become a potential limitation. Microsoft Research

Rᴀᴅ-DINO is now on arXiv!🦖 We show that medical image encoders trained only on images perform similar or better than text supervised models on various benchmarks, demonstrating that reliance on text may not be necessary, and can become a potential limitation. <a href="/MSFTResearch/">Microsoft Research</a>
Stas Bekman (@stasbekman) 's Twitter Profile Photo

Tim doesn't stop to amaze me, head here to read where Tim Dettmers describes the working of the paged AdamW github.com/TimDettmers/bi… He basically built a linux cpu memory paging system but for GPU/CUDA. So when you run out of GPU memory instead of OOM, the unused memory blocks

Hubert P. H. Shum (@hubertshum) 's Twitter Profile Photo

[Job Alert] Associate/Assistant Professor in Artificial Intelligence for Space-Enabled Technologies at Durham University Perks: 1️⃣fully 'tenured' in US terminology at a top 100 Uni 2⃣reduced teaching 3⃣a funded PhD 4⃣travel budget 5⃣chance for a funded PostDoc shorturl.at/otwHQ

[Job Alert] Associate/Assistant Professor in Artificial Intelligence for Space-Enabled Technologies at
<a href="/durham_uni/">Durham University</a> Perks: 1️⃣fully 'tenured' in US terminology at a top 100 Uni 2⃣reduced teaching 3⃣a funded PhD 4⃣travel budget 5⃣chance for a funded PostDoc shorturl.at/otwHQ
Quentin Anthony (@quentinanthon15) 's Twitter Profile Photo

Getting the most out of your hardware when training transformers requires thinking about your model as a sequence of GPU kernel calls. This mindset, common in HPC, is rare in ML and leads to inefficiencies in LLM training. Learn more in our paper arxiv.org/abs/2401.14489

Getting the most out of your hardware when training transformers requires thinking about your model as a sequence of GPU kernel calls. This mindset, common in HPC, is rare in ML and leads to inefficiencies in LLM training. 

Learn more in our paper arxiv.org/abs/2401.14489
Fernando Pérez-García (@fepegar_) 's Twitter Profile Photo

🔥 Our new preprint on biomedical self-supervised learning with images only is out! 🔥 Rᴀᴅ-DINO: Exploring Scalable Medical Image Encoders Beyond Text Supervision by Microsoft Research We show that: 0/3 🧵

🔥 Our new preprint on biomedical self-supervised learning with images only is out! 🔥

Rᴀᴅ-DINO: Exploring Scalable Medical Image Encoders Beyond Text Supervision
by <a href="/MSFTResearch/">Microsoft Research</a> 

We show that:

0/3 🧵
Chris Willcocks (@cwkx) 's Twitter Profile Photo

Congratulations to Sam Bond-Taylor (Sam Bond-Taylor) for successfully defending his thesis on "Modelling High-Dimensional Data with Likelihood-Based Generative Models" Comp Sci @ Durham Durham University Thanks to the examiners Haiping Lu Amir Atapour-Abarghouei.

Congratulations to <a href="/sambondtaylor/">Sam Bond-Taylor</a> (Sam Bond-Taylor) for successfully defending his thesis on "Modelling High-Dimensional Data with Likelihood-Based Generative Models" <a href="/comp_sci_durham/">Comp Sci @ Durham</a> <a href="/durham_uni/">Durham University</a>  Thanks to the examiners <a href="/haipinglu/">Haiping Lu</a> <a href="/AmirAtapour/">Amir Atapour-Abarghouei</a>.
Albert Gu (@_albertgu) 's Twitter Profile Photo

Releasing Hydra, our "official" extension of Mamba (and general state space models) to be bidirectional! Hydra is motivated from first principles by increasing expressivity through the framework of "matrix mixer" sequence models

Gagan Jain @ ICLR'25 (@gaganjain1582) 's Twitter Profile Photo

👀A sneak-peek at our recent work accepted at #ECCV2024 European Conference on Computer Vision #ECCV2026 ! 📜: arxiv.org/abs/2407.12753… Want a scalable and robust foundational model that also saves computational costs? Say hello to LookupViT, our information compression idea leading to sub-quadratic attention! (1/10)

👀A sneak-peek at our recent work accepted at #ECCV2024 <a href="/eccvconf/">European Conference on Computer Vision #ECCV2026</a> ! 
📜: arxiv.org/abs/2407.12753…

Want a scalable and robust foundational model that also saves computational costs? Say hello to LookupViT, our information compression idea leading to sub-quadratic attention!
(1/10)
Microsoft Research (@msftresearch) 's Twitter Profile Photo

RadEdit stress-tests biomedical vision models by simulating dataset shifts through precise image editing. It uses diffusion models to create realistic, synthetic datasets, helping to identify model weaknesses and evaluate robustness: msft.it/6011mlGU3

RadEdit stress-tests biomedical vision models by simulating dataset shifts through precise image editing. It uses diffusion models to create realistic, synthetic datasets, helping to identify model weaknesses and evaluate robustness:  msft.it/6011mlGU3
Fernando Pérez-García (@fepegar_) 's Twitter Profile Photo

This morning we released RadEdit's weights on Hugging Face 🤗. RadEdit is a state-of-the-art latent diffusion model to generate chest X-rays conditioned on text inputs, and an editing pipeline designed for model stress-testing. huggingface.co/microsoft/rade… (Cc Cyril Zakka, MD) 🧵2/9

Sulin Liu (@su_lin_liu) 's Twitter Profile Photo

Discrete generative models use denoisers for generation, but they can slip up. What if generation *isn’t only* about denoising?🤔 Introducing DDPD: Discrete Diffusion with Planned Denoising🤗🧵(1/11) w/ Juno Nam Andrew Campbell Hannes Stärk Yilun Xu Tommi Jaakkola RGB Lab @ MIT

Discrete generative models use denoisers for generation, but they can slip up. What if generation *isn’t only* about denoising?🤔

Introducing DDPD: Discrete Diffusion with Planned Denoising🤗🧵(1/11)

w/ <a href="/junonam_/">Juno Nam</a> <a href="/AndrewC_ML/">Andrew Campbell</a> <a href="/HannesStaerk/">Hannes Stärk</a> <a href="/xuyilun2/">Yilun Xu</a> Tommi Jaakkola <a href="/RGBLabMIT/">RGB Lab @ MIT</a>
Emiel Hoogeboom (@emiel_hoogeboom) 's Twitter Profile Photo

Is pixel diffusion passé? In 'Simpler Diffusion' (arxiv.org/abs/2410.19324) , we achieve 1.5 FID on ImageNet512, and SOTA on 128x128 and 256x256. We ablated out a lot of complexity, making it truly 'simpler'. w/ @tejmensink Jonathan Heek Kay Lamerigts Ruiqi Gao Tim Salimans

Is pixel diffusion passé?

In 'Simpler Diffusion' (arxiv.org/abs/2410.19324) , we achieve 1.5 FID on ImageNet512, and SOTA on 128x128 and 256x256. 

We ablated out a lot of complexity, making it truly 'simpler'. w/ @tejmensink <a href="/JonathanHeek/">Jonathan Heek</a> <a href="/KayLamerigts/">Kay Lamerigts</a> <a href="/RuiqiGao/">Ruiqi Gao</a> <a href="/TimSalimans/">Tim Salimans</a>
Fabian Falck (@fabianfalck) 's Twitter Profile Photo

Is low-to-high frequency generation in diffusion models aka. 'approximate spectral autoregression' a necessity for generation performance? 📃 Blog: fabianfalck.com/posts/spectral… 📜 Paper: arxiv.org/abs/2505.11278