Ioannis Kakogeorgiou (@ioanniskakogeo1) 's Twitter Profile
Ioannis Kakogeorgiou

@ioanniskakogeo1

I am a Postdoctoral Researcher at Archimedes AI. My research focuses on deep learning in computer vision and remote sensing.

ID: 1438464792474361856

linkhttps://scholar.google.com/citations?user=B_dKcz4AAAAJ calendar_today16-09-2021 11:28:57

77 Tweet

168 Followers

291 Following

Dimitriadis Nikos @ ICLR (@nikdimitriadis) 's Twitter Profile Photo

Wouldn't it be great if we could merge the knowledge of 20 specialist models into a single one without losing performance? ๐Ÿ’ช๐Ÿป Introducing our new ICML paper "Localizing Task Information for Improved Model Merging and Compression". ๐ŸŽ‰ ๐Ÿ“œ: arxiv.org/pdf/2405.07813 ๐Ÿงต1/9

Wouldn't it be great if we could merge the knowledge of 20 specialist models into a single one without losing performance? ๐Ÿ’ช๐Ÿป

Introducing our new ICML paper "Localizing Task Information for Improved Model Merging and Compression". ๐ŸŽ‰

๐Ÿ“œ: arxiv.org/pdf/2405.07813 

๐Ÿงต1/9
valeo.ai (@valeoai) 's Twitter Profile Photo

SPOT: Self-Training with Patch-Order Permutation for Object-Centric Learning with Autoregressive Transformers by @IoannisKakogeo1@SpyrosGidaris tsiou.karank N. Komodakis tl;dr: improve slot-based autoencoders w/ self-training & patch permutations #CVPR2024 x.com/IoannisKakogeoโ€ฆ

SPOT: Self-Training with Patch-Order Permutation for Object-Centric Learning with Autoregressive Transformers
by @IoannisKakogeo1@SpyrosGidaris <a href="/tsioukarank/">tsiou.karank</a> N. Komodakis
tl;dr: improve slot-based autoencoders w/ self-training &amp; patch permutations #CVPR2024
x.com/IoannisKakogeoโ€ฆ
Ioannis Kakogeorgiou (@ioanniskakogeo1) 's Twitter Profile Photo

Hi #CVPR2024 people! If you are interested in unsupervised object-centric learning and object segmentation, come and chat with us! You'll find us at poster board 315! ๐ŸŽ‰

Hi #CVPR2024 people! If you are interested in unsupervised object-centric learning and 
object segmentation, come and chat with us! You'll find us at poster board 315! ๐ŸŽ‰
IEEE Geoscience and Remote Sensing Society (@ieee_grss) 's Twitter Profile Photo

Day 2 of #IGARSS2024 #Summerschool starts today! From Data to Application ๐Ÿ“Šโžก๏ธ๐Ÿ“ฑ Today we start with an in-depth session on machine learning for Earth Observation by tsiou.karank, Bill Psomas, Ioannis Kakogeorgiou๐ŸŒ๐Ÿค– Let's dive in! ๐Ÿš€๐Ÿ’ป #RemoteSensing #Athens #machinelearning

Day 2 of #IGARSS2024 #Summerschool starts today! 
From Data to Application ๐Ÿ“Šโžก๏ธ๐Ÿ“ฑ
Today we start with an in-depth session on machine learning for Earth Observation by <a href="/tsioukarank/">tsiou.karank</a>,  <a href="/bill_psomas/">Bill Psomas</a>, <a href="/IoannisKakogeo1/">Ioannis Kakogeorgiou</a>๐ŸŒ๐Ÿค–
Let's dive in! ๐Ÿš€๐Ÿ’ป
 #RemoteSensing #Athens #machinelearning
Ioannis Kakogeorgiou (@ioanniskakogeo1) 's Twitter Profile Photo

๐ŸŒŠ Exciting news! Our AI framework for tracking global marine pollution, including debris & oil spills, is featured on #NVIDIA's blog! ๐Ÿš€๐ŸŒ By using deep learning & satellite imagery, we boost ocean cleanup efforts. ๐ŸŒ๐Ÿ’ก Read more: developer.nvidia.com/blog/high-techโ€ฆ #AI NVIDIA AI Developer

Efstathios Karypidis (@k_sta8is) 's Twitter Profile Photo

1/n ๐Ÿš€ Excited to share our latest work: DINO-Foresight, a new framework for predicting the future states of scenes using Vision Foundation Model features! Links to the arXiv and Github ๐Ÿ‘‡

1/n ๐Ÿš€ Excited to share our latest work: DINO-Foresight, a new framework for predicting the future states of scenes using Vision Foundation Model features!
Links to the arXiv and Github ๐Ÿ‘‡
Thodoris Kouzelis (@thkouz) 's Twitter Profile Photo

1/n๐Ÿš€If youโ€™re working on generative image modeling, check out our latest work! We introduce EQ-VAE, a simple yet powerful regularization approach that makes latent representations equivariant to spatial transformations, leading to smoother latents and better generative models.๐Ÿ‘‡

Sander Dieleman (@sedielem) 's Twitter Profile Photo

Here's EQ-VAE as well. Always nice when two papers come to the same conclusion independently, that makes the evidence so much stronger! x.com/ThKouz/status/โ€ฆ

Andrei Bursuc (@abursuc) 's Twitter Profile Photo

EQ-VAE: such a simple & cool trick to regularize multiple kinds of autoencoders: align reconstruction of transformed latents w/ the corresponding transformed inputs. ๐Ÿš€REPA: 4x training speedup ๐Ÿš€MaskGIT: 2x training speedup ๐Ÿš€DiT-XL/2: 7x faster convergence Kudos Thodoris Kouzelis et al.

Dmytro Mishkin ๐Ÿ‡บ๐Ÿ‡ฆ (@ducha_aiki) 's Twitter Profile Photo

ILIAS: Instance-Level Image retrieval At Scale Giorgos Kordopatis-Zilos, Vladan Stojniฤ‡ , Anna Manko, Pavel ล uma, Nikolaos-Antonios Ypsilantis , Nikos Efthymiadis, Zakaria Laskar, Jiล™รญ Matas, Ondล™ej Chum, Giorgos Tolias tl;dr: new retrieval dataset with guaranteed GT. SigLIP rules. arxiv.org/abs/2502.11748 1/

ILIAS: Instance-Level Image retrieval At Scale

<a href="/g_kordo/">Giorgos Kordopatis-Zilos</a>, <a href="/stojnvla/">Vladan Stojniฤ‡</a> , Anna Manko, Pavel ล uma, <a href="/YpsilantisNikos/">Nikolaos-Antonios Ypsilantis</a> , Nikos Efthymiadis, Zakaria Laskar, Jiล™รญ Matas, Ondล™ej Chum, Giorgos Tolias

tl;dr: new retrieval dataset with guaranteed GT. SigLIP rules. 
arxiv.org/abs/2502.11748
1/
valeo.ai (@valeoai) 's Twitter Profile Photo

๐Ÿš— Ever wondered if an AI model could learn to drive just by watching YouTube? ๐ŸŽฅ๐Ÿ‘€ We trained a 1.2B parameter model on 1,800+ hours of raw driving videos. No labels. No maps. Just pure observation. And it works! ๐Ÿคฏ ๐Ÿงต๐Ÿ‘‡ [1/10]

๐Ÿš— Ever wondered if an AI model could learn to drive just by watching YouTube? ๐ŸŽฅ๐Ÿ‘€

We trained a 1.2B parameter model on 1,800+ hours of raw driving videos.

No labels. No maps. Just pure observation.

And it works! ๐Ÿคฏ

๐Ÿงต๐Ÿ‘‡ [1/10]
Efstathios Karypidis (@k_sta8is) 's Twitter Profile Photo

๐Ÿงต Excited to share our latest work: FUTURIST - A unified transformer architecture for multimodal semantic future prediction, is accepted to #CVPR2025 ! Here's how it works (1/n) ๐Ÿ‘‡ Links to the arxiv and github below

Thodoris Kouzelis (@thkouz) 's Twitter Profile Photo

1/n Introducing ReDi (Representation Diffusion): a new generative approach that leverages a diffusion model to jointly capture โ€“ Low-level image details (via VAE latents) โ€“ High-level semantic features (via DINOv2)๐Ÿงต

1/n Introducing ReDi (Representation Diffusion): a new generative approach that leverages a diffusion model to jointly capture
โ€“ Low-level image details (via VAE latents)
โ€“ High-level semantic features (via DINOv2)๐Ÿงต
Sander Dieleman (@sedielem) 's Twitter Profile Photo

One weird trick for better diffusion models: concatenate some DINOv2 features to your latent channels! Combining latents with PCA components extracted from DINOv2 features yields faster training and better samples. Also enables a new guidance strategy. Simple and effective!

DailyPapers (@huggingpapers) 's Twitter Profile Photo

ReDi was released on Hugging Face A new generative image modeling framework that bridges generation and representation learning

ReDi was released on Hugging Face

A new generative image modeling framework that bridges generation and representation learning
#CVPR2025 (@cvpr) 's Twitter Profile Photo

Behind every great conference is a team of dedicated reviewers. Congratulations to this yearโ€™s #CVPR2025 Outstanding Reviewers! cvpr.thecvf.com/Conferences/20โ€ฆ

Sophia Sirko-Galouchenko (@sophia_sirko) 's Twitter Profile Photo

1/n ๐Ÿš€New paper out - accepted at #ICCV2025! Introducing DIP: unsupervised post-training that enhances dense features in pretrained ViTs for dense in-context scene understanding Below: Low-shot in-context semantic segmentation examples. DIP features outperform DINOv2!

1/n ๐Ÿš€New paper out - accepted at <a href="/ICCVConference/">#ICCV2025</a>!

Introducing DIP: unsupervised post-training that enhances dense features in pretrained ViTs for dense in-context scene understanding

Below: Low-shot in-context semantic segmentation examples. DIP features outperform DINOv2!