Calvin McCarter
calvinmccarter.bsky.social
Calvin McCarter
@calvinmccarter.bsky.social
calvinmccarter.com
i first read that as:

"power washing" but for the feet. idea to fill out later
October 3, 2025 at 1:51 AM
Slightly off-topic, but one of my biggest pet peeves is when people say about baby sadness, "Don't worry: they won't remember this." In the long run, we are all memoryless!
September 29, 2025 at 12:36 AM
yea
August 1, 2025 at 1:22 AM
what are the reasons for this from your perspective?
July 26, 2025 at 7:52 PM
maybe not more expensive to manufacture in an absolute sense. but if chinese internal demand rises, and it no longer needs us exports, then china may no longer see the value in exporting to the us. in which case, the cost of consumer goods *in the us* would rise.
June 27, 2025 at 4:14 PM
also, immigrants will also do less of our low-wage / low-status work for us (declining LatAm TFR). meanwhile, AI will (initially) compete with high-wage / high-status workers. so we're still looking at a combustible political situation.
June 27, 2025 at 3:33 PM
tbf, while AI is going to start doing more of our work for us, East Asia will start doing less of our work for us (their trade surpluses will decline due to their aging demographics and the end of the USD as reserve currency)
June 27, 2025 at 3:31 PM
In principle I agree, but federal lands are relatively poorly managed wrt fire prevention and management. Obviously it would be preferable to just fix that, but I'm not sure whether that's realistic.
June 20, 2025 at 12:57 AM
Related to your earlier assessment of the probability of the US striking Iran, how would you assess the probability of an Israeli strike? Is it harder to predict, because Israel doesn't need to bring assets into the region?
June 12, 2025 at 2:36 AM
"autoregression for training, diffusion for inference"
May 21, 2025 at 8:38 PM
alphaxiv does this for all arxiv papers -- just s/ar/alpha/ in the url -- and i've been told that it's coming soon to {bio,chem,med}rxiv as well.
March 18, 2025 at 12:44 AM
I don't disagree with you exactly, but if an institution's natural defenders are AWOL (or even siding with its enemies) due to a litany of grievances, then it's probably already too late to save that institution.
March 16, 2025 at 9:00 PM
(not that there's anything wrong with that)
March 16, 2025 at 5:26 PM
I just delete my old repo, then give myself a new username...
March 11, 2025 at 6:41 PM
it's definitely a Michigan thing
February 12, 2025 at 4:17 AM
I am slightly cynical about the Clean Label Project, given that it seems to be "pay to play". Also, afaict, it was started by the founder of ByHeart formula, and theirs was like the first thing to get certification. Not that this is necessarily bad -- ByHeart and Kendamil are the best formula IMO.
January 12, 2025 at 11:34 PM
the website has a list: cleanlabelproject.org/product-cate... which is specific and helpful, though oddly it slightly differs from the brands in the report:
January 12, 2025 at 11:31 PM
i cooked him
December 23, 2024 at 3:26 AM
There are ways around this, eg arxiv.org/abs/1902.04094, but I think uniform rate MLMs are the cleaner approach. Plus, as shown in Eq 5 of the Factorization Curse paper, it's what you get after marginalizing over all XLNet-style autoregressive permutations.
BERT has a Mouth, and It Must Speak: BERT as a Markov Random Field Language Model
We show that BERT (Devlin et al., 2018) is a Markov random field language model. This formulation gives way to a natural procedure to sample sentences from BERT. We generate from BERT and find that it...
arxiv.org
December 20, 2024 at 7:52 PM
This creates a train-inference gap if one is training at any fixed masking rate. Of course, with MLMs people don't even try that, and instead use pseudo-likelihood as an approximation for likelihood. Besides being approximate, the problem with this is that this takes L separate forward-passes.
December 20, 2024 at 7:52 PM
When one evaluates log-likelihood of a sequence of length L via the chain rule of probability, the first term has missingness fraction of 1, the second has missingness of (L-1)/L, etc. So the inference-time masking rate is ~ Uniform[0, 1].
December 20, 2024 at 7:52 PM