Transformer patches don't need to be of uniform size -- choose sizes based on entropy --> faster training/inference. Are scale-spaces gonna make a comeback?
Transformer patches don't need to be of uniform size -- choose sizes based on entropy --> faster training/inference. Are scale-spaces gonna make a comeback?
* math: precision matters
* knowledge: effective param count is more important
* 4B-8bit threshold — for bigger prefer quant, smaller prefer more params
* parallel TTC only works above 4B-8bit
arxiv.org/abs/2510.10964
* math: precision matters
* knowledge: effective param count is more important
* 4B-8bit threshold — for bigger prefer quant, smaller prefer more params
* parallel TTC only works above 4B-8bit
arxiv.org/abs/2510.10964
youtu.be/e5kDHL-nnh4
youtu.be/e5kDHL-nnh4
youtu.be/e5kDHL-nnh4
youtu.be/e5kDHL-nnh4
youtu.be/e5kDHL-nnh4
youtu.be/e5kDHL-nnh4
I wanted to see how the exact same post would perform on both X (Twitter) and Bluesky.
The results were...interesting...
[Thread]
I wanted to see how the exact same post would perform on both X (Twitter) and Bluesky.
The results were...interesting...
[Thread]
CAPI: Cluster and Predict Latents Patches for Improved Masked Image Modeling.
CAPI: Cluster and Predict Latents Patches for Improved Masked Image Modeling.
code: github.com/verlab/accel...
paper: arxiv.org/abs/2404.19174
project: www.verlab.dcc.ufmg.br/descriptors/...
code: github.com/verlab/accel...
paper: arxiv.org/abs/2404.19174
project: www.verlab.dcc.ufmg.br/descriptors/...
fleuret.org/dlc/
And my "Little Book of Deep Learning" is available as a phone-formatted pdf (nearing 700k downloads!)
fleuret.org/lbdl/
fleuret.org/dlc/
And my "Little Book of Deep Learning" is available as a phone-formatted pdf (nearing 700k downloads!)
fleuret.org/lbdl/
Right now, it only has atproto docs but already been useful to me to answer random questions about the project.
github.com/davidgasquez...
Right now, it only has atproto docs but already been useful to me to answer random questions about the project.
github.com/davidgasquez...