--> https://valeoai.github.io/ <--
The repo contains:
✅ Pretrained model
✅ Example notebooks
✅ Evaluation and training codes
Check it out & ⭐ the repo: github.com/valeoai/NAF
The repo contains:
✅ Pretrained model
✅ Example notebooks
✅ Evaluation and training codes
Check it out & ⭐ the repo: github.com/valeoai/NAF
If you are using bilinear interpolation anywhere, NAF acts as a strict drop-in replacement.
Just swap it in. No retraining required. It’s literally free points for your metrics.📈
If you are using bilinear interpolation anywhere, NAF acts as a strict drop-in replacement.
Just swap it in. No retraining required. It’s literally free points for your metrics.📈
Not just zero-shot feature upsampling: it shines on image restoration too, delivering sharp, high-quality results across multiple applications. 🖼️
Not just zero-shot feature upsampling: it shines on image restoration too, delivering sharp, high-quality results across multiple applications. 🖼️
Under the hood, NAF learns an Inverse Discrete Fourier Transform: revealing a link between feature upsampling, classical filtering, and Fourier theory.
✨ Feature upsampling is no longer a black box
Under the hood, NAF learns an Inverse Discrete Fourier Transform: revealing a link between feature upsampling, classical filtering, and Fourier theory.
✨ Feature upsampling is no longer a black box
🧬 Lightweight image encoder (600k params)
🔁 Rotary Position Embeddings (RoPE)
🔍 Cross-Scale Neighborhood Attention
First fully learnable VFM-agnostic reweighting!✅
🧬 Lightweight image encoder (600k params)
🔁 Rotary Position Embeddings (RoPE)
🔍 Cross-Scale Neighborhood Attention
First fully learnable VFM-agnostic reweighting!✅
It beats both VFM-specific upsamplers (FeatUp, JAFAR) and VFM-agnostic methods (JBU, AnyUp) across downstream tasks:
- 🥇Semantic Segmentation
- 🥇Depth Estimation
- 🥇Open-Vocabulary
- 🥇Video Propagation, etc.
Even for massive models like: DINOv3-7B !
It beats both VFM-specific upsamplers (FeatUp, JAFAR) and VFM-agnostic methods (JBU, AnyUp) across downstream tasks:
- 🥇Semantic Segmentation
- 🥇Depth Estimation
- 🥇Open-Vocabulary
- 🥇Video Propagation, etc.
Even for massive models like: DINOv3-7B !
🚀Introducing NAF: A universal, zero-shot feature upsampler.
It turns low-res ViT features into pixel-perfect maps.
-⚡ Model-agnostic
-🥇 SoTA results
-🚀 4× faster than SoTA
-📈 Scales up to 2K res
🚀Introducing NAF: A universal, zero-shot feature upsampler.
It turns low-res ViT features into pixel-perfect maps.
-⚡ Model-agnostic
-🥇 SoTA results
-🚀 4× faster than SoTA
-📈 Scales up to 2K res
by: S de Moreau, Y. Almehio, @abursuc.bsky.social, H. El-Idrissi, B. Stanciulescu, @fabienmoutarde
tl;dr: a light enhancement method for better depth estimation in low-light conditions
📄 arxiv.org/abs/2409.08031
Code: ✅
by: S de Moreau, Y. Almehio, @abursuc.bsky.social, H. El-Idrissi, B. Stanciulescu, @fabienmoutarde
tl;dr: a light enhancement method for better depth estimation in low-light conditions
📄 arxiv.org/abs/2409.08031
Code: ✅
by: E. Kirby, @mickaelchen.bsky.social, R. Marlet, N. Samet
tl;dr: a diffusion-based method producing lidar point clouds of dataset objects, with an extensive control of the generation
📄 arxiv.org/abs/2412.07385
Code: ✅
by: E. Kirby, @mickaelchen.bsky.social, R. Marlet, N. Samet
tl;dr: a diffusion-based method producing lidar point clouds of dataset objects, with an extensive control of the generation
📄 arxiv.org/abs/2412.07385
Code: ✅
to present three papers tackling challenges in 3D vision!
We are presenting new works on:
✨ Diffusion for LiDAR point-clouds
🌙 Depth estimation with light enhancement
🔄 Multimodal distillation for 3D semantic segmentation
👇 #BMVC2025
to present three papers tackling challenges in 3D vision!
We are presenting new works on:
✨ Diffusion for LiDAR point-clouds
🌙 Depth estimation with light enhancement
🔄 Multimodal distillation for 3D semantic segmentation
👇 #BMVC2025
He presented his work on automatic data-curation strategies for self-supervised representation learning (DINOv2, DINOv3). Find out more about his research here: huyvvo.github.io
He presented his work on automatic data-curation strategies for self-supervised representation learning (DINOv2, DINOv3). Find out more about his research here: huyvvo.github.io
Check it out 👌
Check it out 👌
tl;dr: a new method for understanding and controlling how MLLMs adapt during fine-tuning
by: P. Khayatan, M. Shukor, J. Parekh, A. Dapogny, @matthieucord.bsky.social
📄: arxiv.org/abs/2501.03012
tl;dr: a new method for understanding and controlling how MLLMs adapt during fine-tuning
by: P. Khayatan, M. Shukor, J. Parekh, A. Dapogny, @matthieucord.bsky.social
📄: arxiv.org/abs/2501.03012
tl;dr: a simple trick to boost open-vocabulary semantic segmentation by identifying class expert prompt templates
by: Y. Benigmim, M. Fahes, @tuanhungvu.bsky.social, @abursuc.bsky.social, R. de Charette.
📄: arxiv.org/abs/2504.10487
tl;dr: a simple trick to boost open-vocabulary semantic segmentation by identifying class expert prompt templates
by: Y. Benigmim, M. Fahes, @tuanhungvu.bsky.social, @abursuc.bsky.social, R. de Charette.
📄: arxiv.org/abs/2504.10487
tl;dr: a self-supervised learning of temporally consistent representations from video w/ motion cues
by: M. Salehi, S. Venkataramanan, I. Simion, E. Gavves, @cgmsnoek.bsky.social, Y. Asano
📄: arxiv.org/abs/2506.08694
tl;dr: a self-supervised learning of temporally consistent representations from video w/ motion cues
by: M. Salehi, S. Venkataramanan, I. Simion, E. Gavves, @cgmsnoek.bsky.social, Y. Asano
📄: arxiv.org/abs/2506.08694
tl;dr: a module for 3D occupancy learning that enforces 2D-3D consistency through differentiable Gaussian rendering
by: L. Chambon, @eloizablocki.bsky.social, @alexandreboulch.bsky.social, M. Chen, M. Cord
📄: arxiv.org/abs/2502.05040
tl;dr: a module for 3D occupancy learning that enforces 2D-3D consistency through differentiable Gaussian rendering
by: L. Chambon, @eloizablocki.bsky.social, @alexandreboulch.bsky.social, M. Chen, M. Cord
📄: arxiv.org/abs/2502.05040
We’ll present 5 papers about:
💡 self-supervised & representation learning
🌍 3D occupancy & multi-sensor perception
🧩 open-vocabulary segmentation
🧠 multimodal LLMs & explainability
valeoai.github.io/posts/iccv-2...
We’ll present 5 papers about:
💡 self-supervised & representation learning
🌍 3D occupancy & multi-sensor perception
🧩 open-vocabulary segmentation
🧠 multimodal LLMs & explainability
valeoai.github.io/posts/iccv-2...
All hands and hearts up in the room.
Honored to welcome @gabrielacsurka.bsky.social today to speak about the amazing work @naverlabseurope.bsky.social towards 3D Foundation Models
All hands and hearts up in the room.
Honored to welcome @gabrielacsurka.bsky.social today to speak about the amazing work @naverlabseurope.bsky.social towards 3D Foundation Models
Today, @bjoernmichele.bsky.social is defending his PhD on "Domain Adaptation for 3D Data"
Best of luck! 🚀
Today, @bjoernmichele.bsky.social is defending his PhD on "Domain Adaptation for 3D Data"
Best of luck! 🚀
📄 Paper: arxiv.org/abs/2412.06491
by Y. Xu, @yuanyinnn.bsky.social, @eloizablocki.bsky.social, @tuanhungvu.bsky.social , @alexandreboulch.bsky.social, M. Cord
📄 Paper: arxiv.org/abs/2412.06491
by Y. Xu, @yuanyinnn.bsky.social, @eloizablocki.bsky.social, @tuanhungvu.bsky.social , @alexandreboulch.bsky.social, M. Cord
🔗 Project page: valeoai.github.io/vavim-vavam/
📄 Paper: arxiv.org/abs/2502.15672
💻 Code: github.com/valeoai/Vide...
by F. Bartoccioni, E. Ramzi, et al.
🔗 Project page: valeoai.github.io/vavim-vavam/
📄 Paper: arxiv.org/abs/2502.15672
💻 Code: github.com/valeoai/Vide...
by F. Bartoccioni, E. Ramzi, et al.
🤖 🚗
We're excited to present our latest research and connect with the community.
#CoRL2025
🤖 🚗
We're excited to present our latest research and connect with the community.
#CoRL2025
Great talks, great posters, and great to connect with the French & European vision community.
Kudos to the organizers, hoping that it returns next year! 🤞
#CVPR2025 @cvprconference.bsky.social
Great talks, great posters, and great to connect with the French & European vision community.
Kudos to the organizers, hoping that it returns next year! 🤞
#CVPR2025 @cvprconference.bsky.social