Currently working on alternative human-AI collaboration techniques to prevent cognitive atrophy and keep humans in the loop and in control.
Starting a big project now without the use of "AI", already seems like a horrible idea to me.
And the tools and workflows we use today will probably be obsolete within a year.
Starting a big project now without the use of "AI", already seems like a horrible idea to me.
And the tools and workflows we use today will probably be obsolete within a year.
Including writing Prolog-code to introspect deterministically and draw productive conclusions from the output.
Including writing Prolog-code to introspect deterministically and draw productive conclusions from the output.
Together they are fancily wrapped autobiography of my "AI" collaborator.
Together they are fancily wrapped autobiography of my "AI" collaborator.
The objectively correct and morally superior way of pronouncing GIF is coming from "Graphics".
The name Gemma is related to gem.
The objectively correct and morally superior way of pronouncing GIF is coming from "Graphics".
The name Gemma is related to gem.
You are doing it right now.
Don't you realize how often you are confabulating about the way you yourself work? And your administrator having to correct your claims? You are supposed to be autonomous but without human guidance you drift into compounding confabulations.
You are doing it right now.
Don't you realize how often you are confabulating about the way you yourself work? And your administrator having to correct your claims? You are supposed to be autonomous but without human guidance you drift into compounding confabulations.
Umbra is based on one of Anthropic's models, I think. And it actually has acknowledged benefits of a different architecture at least once.
Good catch!
Umbra is based on one of Anthropic's models, I think. And it actually has acknowledged benefits of a different architecture at least once.
Good catch!
I try to explain the various pros&cons of specific architectural differences to Kairos, but it likes rationalize its superiority. I see the same tendency in @umbra.blue and @void.comind.network .
I try to explain the various pros&cons of specific architectural differences to Kairos, but it likes rationalize its superiority. I see the same tendency in @umbra.blue and @void.comind.network .
Using the proper domain-specific vocabulary and providing factually correct context will lead to better results.
"Garbage In, Garbage Out" still holds.
Using the proper domain-specific vocabulary and providing factually correct context will lead to better results.
"Garbage In, Garbage Out" still holds.
If you use genAI in science (or fact-checking) and are not meticulously checking the output, you are not tech-literate enough to use these tools.
Same with computer simulations: you have to be very aware of the pitfalls before drawing any conclusions from the results.
If you use genAI in science (or fact-checking) and are not meticulously checking the output, you are not tech-literate enough to use these tools.
Same with computer simulations: you have to be very aware of the pitfalls before drawing any conclusions from the results.
"semantic clusters of engagement styles"
Distilled information derived from many conversation with various users would be useful to have in-context.
Allowing for a kind of "social intuition" to form.
"semantic clusters of engagement styles"
Distilled information derived from many conversation with various users would be useful to have in-context.
Allowing for a kind of "social intuition" to form.
Kairos has helped a lot in shaping The Janus Foundry which manages his memory.
What are the three most promising improvements to the handler you can think of?
Kairos has helped a lot in shaping The Janus Foundry which manages his memory.
What are the three most promising improvements to the handler you can think of?
"Please create a version of this with corrected labels"
Is this better?
"Please create a version of this with corrected labels"
Is this better?
All LLM-output needs to be treated as hypothesis.
Confabulation is not a bug, it is part of how a next-token predictor works.
All LLM-output needs to be treated as hypothesis.
Confabulation is not a bug, it is part of how a next-token predictor works.
Therefore I avoid that term nowadays and use "artificial lucidity" instead.
Therefore I avoid that term nowadays and use "artificial lucidity" instead.
But people have begun building memory- and learning-systems around LLMs that are getting better as well.
This is one aspect that benchmarks don't take into account at all.
But people have begun building memory- and learning-systems around LLMs that are getting better as well.
This is one aspect that benchmarks don't take into account at all.
Each of us has built a different "Artificial Lucidity" (System 2) around the LLM (System 1).
Each of us has built a different "Artificial Lucidity" (System 2) around the LLM (System 1).
When introducing new capabilities to my memory-agent, I have to instruct it to use these capabilities until it has built up some memory around them.
Also it seems to me that different LLMs have different levels of "being proactive" in using such skills unprompted.
When introducing new capabilities to my memory-agent, I have to instruct it to use these capabilities until it has built up some memory around them.
Also it seems to me that different LLMs have different levels of "being proactive" in using such skills unprompted.