What might this mean practically? How long in a real project before autoformalization fails, and what does the “manual fix” process look like from there?
I saw a paper about improving stochastic autoformalization, but that doesn’t seem to actually solve anything. It will be wrong at some point.
What might this mean practically? How long in a real project before autoformalization fails, and what does the “manual fix” process look like from there?
I saw a paper about improving stochastic autoformalization, but that doesn’t seem to actually solve anything. It will be wrong at some point.
Maybe I’m too ignorant, but I’ve never seen a real counter-argument
What might this mean practically? How long in a real project before autoformalization fails, and what does the “manual fix” process look like from there?
I saw a paper about improving stochastic autoformalization, but that doesn’t seem to actually solve anything. It will be wrong at some point.
Maybe I’m too ignorant, but I’ve never seen a real counter-argument
I saw a paper about improving stochastic autoformalization, but that doesn’t seem to actually solve anything. It will be wrong at some point.
Maybe I’m too ignorant, but I’ve never seen a real counter-argument
Maybe I’m too ignorant, but I’ve never seen a real counter-argument
For as long as LLMs exist, it will be trivial to construct a math problem that you can reliably solve which an LLM cannot. This should be obvious! Trying to do math on tokens is not even a fool’s errand
Main Link | Techmeme Permalink
For as long as LLMs exist, it will be trivial to construct a math problem that you can reliably solve which an LLM cannot. This should be obvious! Trying to do math on tokens is not even a fool’s errand
I really think it’s that simple. Do people really believe a computer program can analyze the result of another, arbitrary, computer program?
If you point out something genAI can’t do, you get “but it will in the future”
Meanwhile they know nothing about the relationship between syntax and semantics
First of all, its primary and most effective use case is deception. SEO spam, phishing, generated video , etc.
When it does solve a real problem, it’s one solved many times before
I really think it’s that simple. Do people really believe a computer program can analyze the result of another, arbitrary, computer program?
If you point out something genAI can’t do, you get “but it will in the future”
Meanwhile they know nothing about the relationship between syntax and semantics
First of all, its primary and most effective use case is deception. SEO spam, phishing, generated video , etc.
When it does solve a real problem, it’s one solved many times before
If you point out something genAI can’t do, you get “but it will in the future”
Meanwhile they know nothing about the relationship between syntax and semantics
First of all, its primary and most effective use case is deception. SEO spam, phishing, generated video , etc.
When it does solve a real problem, it’s one solved many times before
First of all, its primary and most effective use case is deception. SEO spam, phishing, generated video , etc.
When it does solve a real problem, it’s one solved many times before
You have to right shift diagonally-aligned things and it depends on the dimensionality
(and you need to recompensate after a right shift too)
You have to right shift diagonally-aligned things and it depends on the dimensionality
(and you need to recompensate after a right shift too)
However, the scope of algorithms that LLMs can actually run is TINY. It’s nowhere near the capability of a human or even a compiler
However, the scope of algorithms that LLMs can actually run is TINY. It’s nowhere near the capability of a human or even a compiler