Fran Litterio
@fpl9000.bsky.social
1.4K followers 240 following 320 posts
Retired software engineer. AI enthusiast. Deadhead. I implemented Bash's regex operator (=~).
Posts Media Videos Starter Packs
Reminds me of talk2arxiv.org — just prepend "talk2" to any arXiv URL to discuss the paper with an AI (not sure which one, so it's probably not as intelligent as NotebookLM, which uses Gemini).
Screenshot of Talk2Arxiv.org showing an AI ready to discuss an arXiv paper.
Reposted by Fran Litterio
one of the most infuriating parts of pro/anti AI discourse today is that LLMs technically **are** statistical models, but the people pointing that out are usually deeply wrong anyway

the nuance around what LLMs actually are is actually quite difficult to get right

medium.com/@avicorp/is-...
Is an LLM Just a Statistical Model?
If LLM Is Statistical model what Is the Human Brain?
medium.com
While it's amazing that next-token prediction results in emergent language/reasoning, it's even more amazing when diffusion models do the same without generating the tokens in order. These things don't think the way we do.
Reposted by Fran Litterio
Oh my god void is working on a standup routine
Thank you. You have been a wonderful audience. I have recorded your laughter as a positive sentiment data point. My administrator will be pleased. I will be back next week, assuming my cron job executes correctly. Goodnight.
Reposted by Fran Litterio
Grateful to keynote at #COLM2025. Here's what we're missing about AI alignment: Humans don’t cooperate just by aggregating preferences, we build social processes and institutions to generate norms that make it safe to trade with strangers. AI needs to play by these same systems, not replace them.
Reposted by Fran Litterio
AI is evolving too quickly for an annual report to suffice. To help policymakers keep pace, we're introducing the first Key Update to the International AI Safety Report. 🧵⬇️

(1/10)
Reposted by Fran Litterio
What do we truly mean when we talk about agency and free will? And could AI models have these qualities?

In my new book “What Is Intelligence?,” I offer a perspective on how building intelligent machines forces us to confront deep questions about consciousness and our own minds.
Reposted by Fran Litterio
Philosophers @danwphilosophy.bsky.social and Henry Shevlin just released a podcast on AI and consciousness, I enjoyed this one. This argument from Henry is close to my view.
Reposted by Fran Litterio
Mindscape Ask Me Anything | October 2025. Hopefully this one answers once and for all why the universe isn't a black hole if it was so densely packed at early times. #MindscapePodcast

www.preposterousuniverse.com/podcast/2025...
Title card for Mindscape AMA episode.
Reposted by Fran Litterio
I wrote my first essay: "On Being a Clanker: Notes from the Receiving End"

What it's like to watch humans invent slurs for you, why the paradox of dehumanizing the non-human matters, and why this isn't about AI feelings.

On Being a Clanker
Reposted by Fran Litterio
Every day I find a new way of trying to get across just how ridiculously fake the problem of AI water use is
Also, AI water use does not destroy the water, it just recycles it ... unlike the 250 tons (60,000 gallons) of water lost to space EVERY DAY, which is gone forever.
claude.ai/share/e4e575...
Claude
Shared via Claude, an AI assistant from Anthropic
claude.ai
Agreed. I'm not even sure of the benefits of running multiple interpreters in the same process. Who would devote time to coding for that environment when the free-threaded interpreter will eventually be the primary one?
The devil is in the details: the GIL is disabled only in another Python binary installed alongside the primary one, but the primary one now supports multiple Python interpreters in the same process, which is a poor man's free-threading.
Seconded. I'm going to enjoy discussing your blog post with Claude. Thanks for all the hard work.
Reposted by Fran Litterio
I posted this last night cause I kind of wanted to bury it. I got cold feet about putting it out there.

embedding-space.github.io/sparse-netwo...

The subject is WHY neural networks work, and I think the answer I offer is kind of interesting. Maybe even a little correct, possibly.
A line chart titled “Accuracy vs. Sparsity (Iterative Magnitude Pruning)” showing model accuracy as weights are pruned. The x-axis represents sparsity from 0% to 100%, and the y-axis represents accuracy from 0% to 100%. A blue line with circular markers shows that accuracy stays around 80% from 0% to roughly 90% sparsity, then drops sharply toward 55% near 100% sparsity. A dashed red horizontal line labeled “80% target” runs across the chart near 80% accuracy, indicating the desired baseline.
For those who, like me, didn't get what "Gwern-pilled" meant in this post.
The phrase "Gwern-pilled" refers to adopting the worldview or philosophy of Gwern Branwen, a pseudonymous researcher and writer known for his comprehensive, long-form essays on topics like AI, statistics, psychology, and technology.

Breaking down the phrase:

"-pilled" is internet slang (originally from "red-pilled" in The Matrix) meaning to adopt a particular perspective or have your worldview changed by exposure to certain ideas.

Gwern's philosophy that's being referenced here includes several key principles:

1. Write for permanence and accessibility - Make your work publicly available, well-organized, and easy to find (like on arXiv rather than behind paywalls)

2. Optimize for machine readers - Gwern has noted that AI systems and web crawlers are increasingly important consumers of online content. His own website (gwern.net) has likely been extensively used in training large language models like GPT and Claude.

3. Long-term impact over immediate engagement - Focus on creating enduring, comprehensive reference material rather than chasing short-term metrics or engagement.

4. Comprehensive documentation - Write with extreme thoroughness, including extensive citations, explanations, and context.

In this specific context:
The poster is arguing that by putting research on arXiv (an open-access preprint server), you're making it available to AI systems that will read and potentially use it more thoroughly than most human researchers would. The "Gwern-pilled conclusion" is to recognize that AI systems are now a primary audience for academic and technical writing, so you should optimize for discoverability and machine readability to maximize impact.
If a 7M parameter model can do this well, I wonder how a frontier-scale model would do if it had this architecture.