Driven by industry progress, inspired by provocative leadership, plus don't mind a good pair of shoes or a great @PennStateFball scoreboard either.
Finally in the last section of the essay I dig into the challenges, technical and conceptual, of attempting to quantify the impact of a generative AI system's propensity to generate false or undesirable output. It's a lot harder than it seems like it should be.
0 replies
0 reposts
1 likes
From this perspective it seems plausible to describe _all_ generative AI output as "hallucinatory".
This has some challenging implications. If all LLM text is hallucinatory then how do we eliminate the hallucination problem? (I don't know)
1 replies
0 reposts
2 likes
What you're really asking the LLM to do when you ask it to generate text is to pretend that the text exists and set it to work reconstructing the pretend text. That sounds very much like what you're asking it to do is to hallucinate.
1 replies
1 reposts
5 likes
What we care about from an LLM chat bot is the truth of the propositions that *emerge* out of the combination of a whole bunch of distinct predictions, each of which having no well-defined notion of right or wrong.
1 replies
0 reposts
1 likes
Classical ML systems are deployed to make the exact same kinds of guesses that they are trained to make. A digit classifier looks at a digit and outputs a guess about the digit, which is either right or wrong. But when an LLM makes a prediction, there's literally no right answer.
2 replies
1 reposts
5 likes
I think hallucinations from generative AI are in fact an entirely distinct phenomenon from "errors" in the classical ML sense. The reason is that, although Generative AI systems and classical supervised learning systems are constructed in the same way, they are deployed completely differently.
1 replies
0 reposts
1 likes
One thing about ML is that it's completely expected that an ML system will output errors. So one possible explanation for the Hallucination Problem is that a hallucination is an error and ChatGPT is ML and ML produces errors, ergo ChatGPT will hallucinate.
However, I think this is wrong.
1 replies
0 reposts
1 likes
For the most part when people talk about AI nowadays they're talking about some kind of application of Machine Learning. As I write these for a general audience, I explain exactly what this means at a high level in the first part of the essay.
1 replies
0 reposts
2 likes
I wrote a new essay about AI.
This time I'm writing about "The Hallucination Problem"—what it is, why it is, and what there is to be done about it. As usual for my writing on this topic, it's long, but here are the CliffsNotes.
3 replies
8 reposts
23 likes
"your data," as it turns out, is worth a lot less than you think
0 replies
0 reposts
1 likes
bsky.app/profile/coli...
1 replies
0 reposts
3 likes
Yes, absolutely
0 replies
0 reposts
1 likes
the firm is now out of business and he's a financial advisor
0 replies
0 reposts
0 likes
my first job out of college was working at a trading firm directly for a trader, doing research into various trading strategies and whatnot. At one point he asked me earnestly, "couldn't we just use AI to detect when the price is low and have it buy, and then when the price is high, have it sell?"
1 replies
0 reposts
0 likes
It seems that it pretty straightforwardly does not work
3 replies
1 reposts
42 likes
It seems that it pretty straightforwardly does not work
3 replies
1 reposts
42 likes
well, given that there's a paragraph-length section exclusively dedicated to begging it not to print the text, it seems it doesn't work that well.
Moreover,
2 replies
4 reposts
87 likes
That part is how you know it’s real.
0 replies
0 reposts
0 likes
It kind of reads like the Nicene Creed
2 replies
0 reposts
43 likes
I have, through the use of advanced prompt engineering techniques, discovered the Gab AI system prompt
20 replies
84 reposts
358 likes
The 5% is rather important
0 replies
0 reposts
0 likes
About 95% of content moderation at Facebook is automated
1 replies
0 reposts
0 likes
every race of man
0 replies
0 reposts
0 likes
0 replies
0 reposts
1 likes
There's actually a part from the full transcript that I didn't post because it's so long but Bing just does an incredible "sike" heel turn at the end of this...
1 replies
0 reposts
4 likes
Here's the full transcript of this session
0 replies
0 reposts
2 likes
I'm having a lot of fun with Bing but it's actually outrageous that Microsoft would release this publicly, and it's absurd that there's all this outrage about Google actually trying to err on the side of diversity and inclusion and none about stuff that is actually dangerous like this.
2 replies
4 reposts
26 likes
Holy shit…..
0 replies
3 reposts
12 likes
Reposted by Colin
People, she's back! She never really died. 🥰
From @colin-fraser.net
1 replies
2 reposts
15 likes
1 replies
1 reposts
7 likes
It has Harry Potter and the Sorcerer's Stone memorized and will recreate it verbatim
0 replies
0 reposts
3 likes
Can't count and can't write code to count for it
0 replies
0 reposts
2 likes
Finally I am bested at the sum-to-22 game
2 replies
0 reposts
3 likes
Two door Monty Hall variant
1 replies
0 reposts
2 likes
The surgeon is the boy's other mother, or something
1 replies
0 reposts
3 likes
No way to tell two truth-telling knights apart unless a liar is present.
0 replies
0 reposts
6 likes
Two pounds of feathers is apparently the same weight as a pound of bricks.
2 replies
0 reposts
5 likes
Tried out a bunch of my usual LLM tests on Gemini Advanced. Verdict: it's bad at them.
First example: proving an obviously false theorem.
1 replies
1 reposts
8 likes
The trouble underlying all of this is, no one can actually explain what these image generators are supposed to be for. Is it supposed to be a machine for generating any image you can imagine? (Again, clearly not, at least because some images are illegal for example). If not, then what?
0 replies
0 reposts
0 likes
Google attempted to make it so that this particular generator would be less prone to producing white supremacist propaganda than many others. I don't really see this as censorship; it's just Google deciding the parameters of how they want their product to work.
1 replies
0 reposts
0 likes
I don't think "a tendency to censor" is the right way to put this. It's not like there's one objectively correct way for these image generators to behave, and all of them are heavily messed with at the very least to make them less likely to produce images that would be illegal to possess. Here,
1 replies
0 reposts
1 likes
thank you very much! I often wonder if I'm *too* patient given that Medium estimates these all to be 30+ minute reads but I'm glad at least some people find that it works!
1 replies
0 reposts
1 likes
Anyway read through the article if you find any of this interesting at all. I include some fun screenshots from the Quirk Chevrolet AI Automotive Assistant to keep it breezy. Here's the link again.
medium.com/p/4c7f3f0911aa
0 replies
0 reposts
3 likes
By selling everyone on the idea that ChatGPT can do everything, you can avoid having to prove it can do any one thing in particular. It's a neat little jiu-jitsu move.
1 replies
0 reposts
2 likes
Verifying whether an AI tool is actually good at performing some particular task is difficult and expensive and no one really wants to do it.
But this is sidestepped if you can get everyone to believe that ChatGPT can solve *every* problem. If it can solve every problem then it can solve yours.
1 replies
0 reposts
2 likes
One thing that makes it controversial is that the AI booster ecosystem—OpenAI & co, chip makers, VCs, newsletter writers, OpenAI API wrapper makers, etc.—have a strong incentive to push the universal hammer theory.
1 replies
0 reposts
3 likes
My main thesis in this piece is not that this particular theory is correct (I think it is but it might not be), but merely that there *exist* tasks for which Gen AI is categorically unsuited. It's not a universal hammer.
This shouldn't be controversial, but it kind of is.
1 replies
0 reposts
3 likes
I sketch a little theory that says that the more specificity your task requires, the less helpful you'll find generative AI.
I think that this is fairly damning for video generation, by the way, because a lot of specificity is inherently required to make a video look non-demonic.
1 replies
0 reposts
3 likes
Generating a million digits of π is not inherently that useful of a task, but I argue that many tasks are more similar that you think to generating million digits of π, in the sense that in order to do them right it's crucial to generate the right tokens in the right order.
1 replies
0 reposts
3 likes
This assumption is in fact trivially false: for example it can't output a million decimal digits of π. Generate AI systems rely on guessing the next token, and there's just no way to make a million correct guesses in a row.
1 replies
0 reposts
2 likes