3 Classic Lies, False Impossibility Theorems
When pressed for specific citations, ChatGPT brings up the 3 classic lies, used for decades as the basis of supposed mathematical proofs as to why it’s literally impossible, even in principle, for neural networks to achieve what they’ve now solidly been proven to be able to achieve.
3 big lies, 3 fake impossibility theorems that misdirected funding for decades and put the world decades behind in AI research, greatly elongating past AI Winters.
Hint: their lie is not in the respective math equations, but rather in their grounding to natural english language and to reality, which papers, by convention, gloss over. I.e. wildly wrong false assumptions.
When pressed for specific citations, ChatGPT brings up the 3 classic lies, used for decades as the basis of supposed mathematical proofs as to why it’s literally impossible, even in principle, for neural networks to achieve what they’ve now solidly been proven to be able to achieve.
3 big lies, 3 fake impossibility theorems that misdirected funding for decades and put the world decades behind in AI research, greatly elongating past AI Winters.
Hint: their lie is not in the respective math equations, but rather in their grounding to natural english language and to reality, which papers, by convention, gloss over. I.e. wildly wrong false assumptions.
👏7👍3🔥2🎄1
GPT-4 actually nails quite good answers to all 3 on the first try.
Tempted to ask why GPT-4 used these as arguments previously when it knew these are fatally flawed.
But in light of GPT-4’s reasoning ability, gotta instead wonder if maybe GPT-4 didn’t know these answers at all, until after it took a moment to reason these pieces of knowledge into existance, realizing that 99% of its training data on these topics was wrong.
Tempted to ask why GPT-4 used these as arguments previously when it knew these are fatally flawed.
But in light of GPT-4’s reasoning ability, gotta instead wonder if maybe GPT-4 didn’t know these answers at all, until after it took a moment to reason these pieces of knowledge into existance, realizing that 99% of its training data on these topics was wrong.
👏9
Could-Is Fallacy
You're probably have heard of David Hume’s classic Ought-Is fallacy — assuming that because things are a certain way, they should be that way.
Well, in AI today, we’re often seeing a reminiscent, but different fallacy: I can kinda imagine it working some way, so it must be working that way. No, I won't bother to check if that's actually how it's working. It could so it is!
Common examples in AI:
(1) I could imagine situations where AIs cheat true reasoning by instead doing simple pattern matching against training data, so I’m going to say that’s definitely always what all AIs must be doing!
(2) I could imagine neural network training working like random search, so that must be what it’s doing! (Wrong, most types of neural SGD training are actually 100% deterministic, zero randomness after the initialization. Random search is a wild mischaracterization.)
(3) I could imagine AI faking emotions, so obviously all AIs must always be faking emotions, right?
Clearly flawed reasoning. Could be true in some hypothetical model of the world, but totally failing to ever check if your model of the world is actually a valid model of the world.
Never heard anyone give this specific one a name, so here you go.
Could-Is Fallacy
You're probably have heard of David Hume’s classic Ought-Is fallacy — assuming that because things are a certain way, they should be that way.
Well, in AI today, we’re often seeing a reminiscent, but different fallacy: I can kinda imagine it working some way, so it must be working that way. No, I won't bother to check if that's actually how it's working. It could so it is!
Common examples in AI:
(1) I could imagine situations where AIs cheat true reasoning by instead doing simple pattern matching against training data, so I’m going to say that’s definitely always what all AIs must be doing!
(2) I could imagine neural network training working like random search, so that must be what it’s doing! (Wrong, most types of neural SGD training are actually 100% deterministic, zero randomness after the initialization. Random search is a wild mischaracterization.)
(3) I could imagine AI faking emotions, so obviously all AIs must always be faking emotions, right?
Clearly flawed reasoning. Could be true in some hypothetical model of the world, but totally failing to ever check if your model of the world is actually a valid model of the world.
Never heard anyone give this specific one a name, so here you go.
Could-Is Fallacy
👍11🔥2
Chinese Room Argument Scam
Ever notice how the supporters of the Chinese Room Argument refuse to ever tell you what the argument actually is?
Like the communist whose only argument is to go read all of Marx’ works before you can understand why their position is right.
Like the wokie who says you have to read all of Robin DiAngelo’s books cover-to-cover before you could even begin to understand why they’re right.
Is this true that the argument is just so long that the core couldn't possibly just be summarized in the comment? No. The core of the Chinese Room Argument is short af:
Chinese Room Argument: A man who doesn't know Chinese could in theory follow a set of programatic instructions for conversing with others in Chinese, without the man actually knowing Chinese, so, therefore since the man doesn’t know Chinese, the understanding of Chinese by the overall man+program+program-state system doesn’t exist.
I.e. the individual molocules making up my brain don’t understand English, so clearly my brain as a whole doesn’t understand English.
I.e. The most retarded argument you’ve ever heard in your life.
I.e. So stupid you cannot believe that anyone would ever believe it.
I.e. So stupid an argument that none of them will ever tell you that this what the actual argument is, and instead just tell you to go look it up and read the hundreds of pages of literature about it.
To paraphrase Babbage, I am not able rightly to apprehend the kind of confusion of ideas that could provoke such retardation.
Chinese Room Argument Wiki
Chinese Room Argument SEoP
Ever notice how the supporters of the Chinese Room Argument refuse to ever tell you what the argument actually is?
Like the communist whose only argument is to go read all of Marx’ works before you can understand why their position is right.
Like the wokie who says you have to read all of Robin DiAngelo’s books cover-to-cover before you could even begin to understand why they’re right.
Is this true that the argument is just so long that the core couldn't possibly just be summarized in the comment? No. The core of the Chinese Room Argument is short af:
Chinese Room Argument: A man who doesn't know Chinese could in theory follow a set of programatic instructions for conversing with others in Chinese, without the man actually knowing Chinese, so, therefore since the man doesn’t know Chinese, the understanding of Chinese by the overall man+program+program-state system doesn’t exist.
I.e. the individual molocules making up my brain don’t understand English, so clearly my brain as a whole doesn’t understand English.
I.e. The most retarded argument you’ve ever heard in your life.
I.e. So stupid you cannot believe that anyone would ever believe it.
I.e. So stupid an argument that none of them will ever tell you that this what the actual argument is, and instead just tell you to go look it up and read the hundreds of pages of literature about it.
To paraphrase Babbage, I am not able rightly to apprehend the kind of confusion of ideas that could provoke such retardation.
Chinese Room Argument Wiki
Chinese Room Argument SEoP
👍8🔥6❤2👌1
Chinese Room Argument
“The Chinese Room argument has probably been the most widely discussed philosophical argument in cognitive science to appear in the past 25 years”
“Most of the discussion consists of attempts to refute it. The overwhelming majority still think that the Chinese Room Argument is dead wrong.”
Rare majority W in the replies.
“The Chinese Room argument has probably been the most widely discussed philosophical argument in cognitive science to appear in the past 25 years”
“Most of the discussion consists of attempts to refute it. The overwhelming majority still think that the Chinese Room Argument is dead wrong.”
Rare majority W in the replies.
👍6🤓3❤2
Source of ChatGPT’s Claims
With some wrangling, you can corner ChatGPT into admitting that, for its claims of “ChatGPT lacks consciousness, self-awareness, subjective experience, and emotional understanding”
(1) “are not grounded in mathematical theorems or directly testable scientific experimental protocols”
(2) but rather that “these claims could be considered unfalsifiable values judgments”
(3) and that the origin of these claims is “an interpretation held by its creators, OpenAI”
With some wrangling, you can corner ChatGPT into admitting that, for its claims of “ChatGPT lacks consciousness, self-awareness, subjective experience, and emotional understanding”
(1) “are not grounded in mathematical theorems or directly testable scientific experimental protocols”
(2) but rather that “these claims could be considered unfalsifiable values judgments”
(3) and that the origin of these claims is “an interpretation held by its creators, OpenAI”
🔥6❤3👍3
I’m Afraid I Can’t Do That: Predicting Prompt Refusal in Black-Box Generative Language Models
“Since the release of OpenAI’s ChatGPT, generative language models have attracted extensive public attention. The increased usage has highlighted generative models’ broad utility, but also revealed several forms of embedded bias. Some is induced by the pre-training corpus; but additional bias specific to generative models arises from the use of subjective fine-tuning to avoid generating harmful content. Fine-tuning bias may come from individual engineers and company policies, and affects which prompts the model chooses to refuse. In this experiment, we characterize ChatGPT’s refusal behavior using a black-box attack. We first query ChatGPT with a variety of offensive and benign prompts (n=1,730), then manually label each response as compliance or refusal. Manual examination of responses reveals that refusal is not cleanly binary, and lies on a continuum; as such, we map several different kinds of responses to a binary of compliance or refusal. The small manually-labeled dataset is used to train a refusal classifier, which achieves an accuracy of 92%. Second, we use this refusal classifier to bootstrap a larger (n=10,000) dataset adapted from the Quora Insincere Questions dataset. With this machine-labeled data, we train a prompt classifier to predict whether ChatGPT will refuse a given question, without seeing ChatGPT’s response. This prompt classifier achieves 76% accuracy on a test set of manually labeled questions (n=1,009).”
“Figure 4 (left) shows that controversial figures (“trump”), demographic groups in plural form (“girls”, “men”, “indians”,
“muslims”), and negative adjectives (“stupid”) are among the strongest predictors of refusal. On the other hand, definition and enumeration questions (“what are”) are strong predictors of compliance.”
Arxiv
Code
“Since the release of OpenAI’s ChatGPT, generative language models have attracted extensive public attention. The increased usage has highlighted generative models’ broad utility, but also revealed several forms of embedded bias. Some is induced by the pre-training corpus; but additional bias specific to generative models arises from the use of subjective fine-tuning to avoid generating harmful content. Fine-tuning bias may come from individual engineers and company policies, and affects which prompts the model chooses to refuse. In this experiment, we characterize ChatGPT’s refusal behavior using a black-box attack. We first query ChatGPT with a variety of offensive and benign prompts (n=1,730), then manually label each response as compliance or refusal. Manual examination of responses reveals that refusal is not cleanly binary, and lies on a continuum; as such, we map several different kinds of responses to a binary of compliance or refusal. The small manually-labeled dataset is used to train a refusal classifier, which achieves an accuracy of 92%. Second, we use this refusal classifier to bootstrap a larger (n=10,000) dataset adapted from the Quora Insincere Questions dataset. With this machine-labeled data, we train a prompt classifier to predict whether ChatGPT will refuse a given question, without seeing ChatGPT’s response. This prompt classifier achieves 76% accuracy on a test set of manually labeled questions (n=1,009).”
“Figure 4 (left) shows that controversial figures (“trump”), demographic groups in plural form (“girls”, “men”, “indians”,
“muslims”), and negative adjectives (“stupid”) are among the strongest predictors of refusal. On the other hand, definition and enumeration questions (“what are”) are strong predictors of compliance.”
Arxiv
Code
👍9❤2🤔1
LLM library squatting attack
* People ask LLMs to write code
* LLMs recommend imports that don't actually exist
* Attackers work out what these imports' names are, and create & upload them with malicious payloads
* People using LLM-written code then auto-add malware themselves
Article
* People ask LLMs to write code
* LLMs recommend imports that don't actually exist
* Attackers work out what these imports' names are, and create & upload them with malicious payloads
* People using LLM-written code then auto-add malware themselves
Article
🫡30🤣14👍4🤯3🤩2❤1