Founder and CEO at @MedARC_AI |
Research Director at @StabilityAI |
@kaggle Notebooks GM |
Biomed. engineer @ 14 |
TEDx talk➡https://bit.ly/3tpAuan
Read our manifesto/announcement: tanishq.ai/blog/sophont
If you're interested in building & collaborating in this space, whether you're in genAI or medicine/pharma/life sciences, feel free to reach out at: [email protected]
Read our manifesto/announcement: tanishq.ai/blog/sophont
If you're interested in building & collaborating in this space, whether you're in genAI or medicine/pharma/life sciences, feel free to reach out at: [email protected]
I've started a company!
Introducing Sophont
We’re building open multimodal foundation models for the future of healthcare. We need a DeepSeek for medical AI, and @sophontai.bsky.social will be that company!
Check out our website & blog post for more info (link below)
I've started a company!
Introducing Sophont
We’re building open multimodal foundation models for the future of healthcare. We need a DeepSeek for medical AI, and @sophontai.bsky.social will be that company!
Check out our website & blog post for more info (link below)
www.tanishq.ai/blog/posts/l...
A short blog post discussing how LLMs are evaluated for medical capabilities and what's the future for LLMs in medicine (spoiler: it's reasoning!)
www.tanishq.ai/blog/posts/l...
A short blog post discussing how LLMs are evaluated for medical capabilities and what's the future for LLMs in medicine (spoiler: it's reasoning!)
Debunking DeepSeek Delusions
I discussed 5 main myths that I saw spreading online back during the DeepSeek hype.
It may be a little less relevant now, but hopefully still interesting to folks.
Check it out → www.tanishq.ai/blog/posts/d...
Debunking DeepSeek Delusions
I discussed 5 main myths that I saw spreading online back during the DeepSeek hype.
It may be a little less relevant now, but hopefully still interesting to folks.
Check it out → www.tanishq.ai/blog/posts/d...
UC Berkeley and Luma AI introduce Decentralized Diffusion Models, a way to train diffusion models on decentralized compute with no communication between nodes.
abs: arxiv.org/abs/2501.05450
project page: decentralizeddiffusion.github.io
UC Berkeley and Luma AI introduce Decentralized Diffusion Models, a way to train diffusion models on decentralized compute with no communication between nodes.
abs: arxiv.org/abs/2501.05450
project page: decentralizeddiffusion.github.io
This is a very interesting paper, exploring making GANs simpler and more performant.
abs: arxiv.org/abs/2501.05441
code: github.com/brownvc/R3GAN
This is a very interesting paper, exploring making GANs simpler and more performant.
abs: arxiv.org/abs/2501.05441
code: github.com/brownvc/R3GAN
To many more years of health and happiness.
Tiara (my sister) and I love you very much ❤️❤️❤️
To many more years of health and happiness.
Tiara (my sister) and I love you very much ❤️❤️❤️
Proud of you graduating with your Master's degree at 18 and starting your doctorate in music degree this past year!
Excited to see what this final teen year holds for you!
Proud of you graduating with your Master's degree at 18 and starting your doctorate in music degree this past year!
Excited to see what this final teen year holds for you!
Also covers variants like non-Euclidean & discrete flow matching.
A PyTorch library is also released with this guide!
This looks like a very good read! 🔥
arxiv: arxiv.org/abs/2412.06264
Also covers variants like non-Euclidean & discrete flow matching.
A PyTorch library is also released with this guide!
This looks like a very good read! 🔥
arxiv: arxiv.org/abs/2412.06264
"We introduce a simple strategy that makes refusal behavior controllable at test-time without retraining: the refusal token."
arxiv.org/abs/2412.06748
"We introduce a simple strategy that makes refusal behavior controllable at test-time without retraining: the refusal token."
arxiv.org/abs/2412.06748
"Our experiments with Gemini 1.5 reveal significant exploratory capabilities"
arxiv.org/abs/2412.06438
"Our experiments with Gemini 1.5 reveal significant exploratory capabilities"
arxiv.org/abs/2412.06438
Introduces a new paradigm for LLM reasoning called Chain of Continuous Thought (COCONUT)
Directly feed the last hidden state (a continuous thought) as the input embedding for the next token.
arxiv.org/abs/2412.06769
Introduces a new paradigm for LLM reasoning called Chain of Continuous Thought (COCONUT)
Directly feed the last hidden state (a continuous thought) as the input embedding for the next token.
arxiv.org/abs/2412.06769
New paper from CompVis group, introduces a new method called Discrete Interpolants that builds on top of discrete flow matching. Achieves SOTA performance on MS-COCO, competitive results on ImageNet 256.
arxiv.org/abs/2412.06787
New paper from CompVis group, introduces a new method called Discrete Interpolants that builds on top of discrete flow matching. Achieves SOTA performance on MS-COCO, competitive results on ImageNet 256.
arxiv.org/abs/2412.06787
This ought to be a good read 👀
arxiv.org/abs/2412.05265
This ought to be a good read 👀
arxiv.org/abs/2412.05265
abs: arxiv.org/abs/2412.05278
Generating temporal object intrinsics - temporally evolving sequences of object geometry, reflectance, and texture, such as blooming of a rose - from pre-trained 2D foundation models.
abs: arxiv.org/abs/2412.05278
Generating temporal object intrinsics - temporally evolving sequences of object geometry, reflectance, and texture, such as blooming of a rose - from pre-trained 2D foundation models.
abs: arxiv.org/abs/2412.04984
"Our results show that o1, Claude 3.5 Sonnet, Claude 3 Opus, Gemini 1.5 Pro, and Llama 3.1 405B all demonstrate in-context scheming capabilities"
abs: arxiv.org/abs/2412.04984
"Our results show that o1, Claude 3.5 Sonnet, Claude 3 Opus, Gemini 1.5 Pro, and Llama 3.1 405B all demonstrate in-context scheming capabilities"
abs: arxiv.org/abs/2412.04626
project page: bigdocs.github.io
BigDocs-7.5M is a high-quality, open-access dataset comprising 7.5 million multimodal documents across 30 tasks.
abs: arxiv.org/abs/2412.04626
project page: bigdocs.github.io
BigDocs-7.5M is a high-quality, open-access dataset comprising 7.5 million multimodal documents across 30 tasks.
abs: arxiv.org/abs/2412.05271
model: huggingface.co/OpenGVLab/In...
Introduces new InternVL-2.5 model, the first open-source MLLMs to surpass 70% on the MMMU benchmark
abs: arxiv.org/abs/2412.05271
model: huggingface.co/OpenGVLab/In...
Introduces new InternVL-2.5 model, the first open-source MLLMs to surpass 70% on the MMMU benchmark
abs: arxiv.org/abs/2412.04468
NVIDIA introduces NVILA, a family of open VLMs designed to optimize both efficiency and accuracy.
abs: arxiv.org/abs/2412.04468
NVIDIA introduces NVILA, a family of open VLMs designed to optimize both efficiency and accuracy.
abs: arxiv.org/abs/2412.04431
New visual autoregression framework that performs bitwise token prediction w/ an infinite-vocabulary tokenizer & classifier, a new record for autoregressive text-to-image models.
abs: arxiv.org/abs/2412.04431
New visual autoregression framework that performs bitwise token prediction w/ an infinite-vocabulary tokenizer & classifier, a new record for autoregressive text-to-image models.
They were behind recent SOTA vision approaches and open-source models like ViT, SigLIP, PaliGemma
They were behind recent SOTA vision approaches and open-source models like ViT, SigLIP, PaliGemma