Do Machine Learning Models Memorize or Generalize?
Yes, both, in that order. First learn to parrot then learn to think.
โIn 2021, researchers made a striking discovery while training a series of tiny models on toy tasks. They found a set of models that suddenly flipped from memorizing their training data to correctly generalizing on unseen inputs after training for much longer. This phenomenon โ where generalization seems to happen abruptly and long after fitting the training data โ is called grokking and has sparked a flurry of interestโ
โThe sharp drop in test loss makes it appear like the model makes a sudden shift to generalization. But if we look at the weights of the model over training, most of them smoothly interpolate between the two solutions. The rapid generalization occurs when the last weights connected to the distracting digits are pruned by weight decay.โ
Translation: The shift from parroting to real understanding happens fairly smoothly, though external results don't show it at first, and then bam, it all comes together.
Sound analogous to what happens in humans? That's because it is. Behavior of large AI models is incredibly similar humans, in countless ways.
Website with great visuals
Yes, both, in that order. First learn to parrot then learn to think.
โIn 2021, researchers made a striking discovery while training a series of tiny models on toy tasks. They found a set of models that suddenly flipped from memorizing their training data to correctly generalizing on unseen inputs after training for much longer. This phenomenon โ where generalization seems to happen abruptly and long after fitting the training data โ is called grokking and has sparked a flurry of interestโ
โThe sharp drop in test loss makes it appear like the model makes a sudden shift to generalization. But if we look at the weights of the model over training, most of them smoothly interpolate between the two solutions. The rapid generalization occurs when the last weights connected to the distracting digits are pruned by weight decay.โ
Translation: The shift from parroting to real understanding happens fairly smoothly, though external results don't show it at first, and then bam, it all comes together.
Sound analogous to what happens in humans? That's because it is. Behavior of large AI models is incredibly similar humans, in countless ways.
Website with great visuals
๐7โค1
Large AI models shift from memorizing to understanding during training
Notice how the โtrain accuracyโ i.e. how well the model does on problems itโs already seen during training, quickly goes to 100% in part due to memorization, but the โtest accuracyโ, i.e. on problems it has not seen, and requiring some actual understanding, shoots up much later, long after it reached ~100% on โtrain accuracy.โ
AI models first parrot, but then learn to truly understand.
(To whatever degree the training set and loss function necessitates true understanding, i.e. in the case where they pose an โAI hardโ well, the degree of true understanding they neccessitate can be unboundedly high.)
Notice how the โtrain accuracyโ i.e. how well the model does on problems itโs already seen during training, quickly goes to 100% in part due to memorization, but the โtest accuracyโ, i.e. on problems it has not seen, and requiring some actual understanding, shoots up much later, long after it reached ~100% on โtrain accuracy.โ
AI models first parrot, but then learn to truly understand.
(To whatever degree the training set and loss function necessitates true understanding, i.e. in the case where they pose an โAI hardโ well, the degree of true understanding they neccessitate can be unboundedly high.)
โค3๐2
Illustration showing the shift from memorizing to understanding happening slowly โ Despite the impact of that accumulating understanding suddenly appearing as a big spike toward the end
"The sharp drop in test loss makes it appear like the model makes a sudden shift to generalization. But if we look at the weights of the model over training, most of them smoothly interpolate between the two solutions. The rapid generalization occurs when the last weights connected to the distracting digits are pruned by weight decay.โ
Do Machine Learning Models Memorize or Generalize?
"The sharp drop in test loss makes it appear like the model makes a sudden shift to generalization. But if we look at the weights of the model over training, most of them smoothly interpolate between the two solutions. The rapid generalization occurs when the last weights connected to the distracting digits are pruned by weight decay.โ
Do Machine Learning Models Memorize or Generalize?
๐2โค1
Memorization alone is ideal when the teacher always gives you correct answers -- But fails terribly as soon as the teacher occasionally starts giving you incorrect answers
โOur results support the natural conclusion that interpolation is particularly beneficial in settings with low label noise, which as we note earlier,
may include some of the most widely-used existing benchmarks for deep learning.โ
Arxiv Paper
โOur results support the natural conclusion that interpolation is particularly beneficial in settings with low label noise, which as we note earlier,
may include some of the most widely-used existing benchmarks for deep learning.โ
Arxiv Paper
โค2
Privacy vs Control Sleight of Hand
Microsoft & OpenAI announce โAzure ChatGPT: Private & secure ChatGPT for internal enterprise useโ
Why does big tech focus so much on privacy?
Answer: to distract you from what really matters, their control.
You let them put one of their AI agents inside your business, you give it full control, inserting it inbetween near every point in your business.
Who cares if it canโt phone home with your secrets? Youโve already given it near total control.
๐ They no longer have to violate your privacy by stealing the keys to the car. They now control your car, and can steal it just by telling it to drive itself over to them.
Notice here how they even try to redefine the word โcontrolledโ to be about privacy (controlling your network privacy), instead ofโฆ being about actual control that matters.
They'll try to convince you that the battle is about privacy.
It's not, it's about control.
Azure ChatGPT Github
Microsoft & OpenAI announce โAzure ChatGPT: Private & secure ChatGPT for internal enterprise useโ
Why does big tech focus so much on privacy?
Answer: to distract you from what really matters, their control.
You let them put one of their AI agents inside your business, you give it full control, inserting it inbetween near every point in your business.
Who cares if it canโt phone home with your secrets? Youโve already given it near total control.
๐ They no longer have to violate your privacy by stealing the keys to the car. They now control your car, and can steal it just by telling it to drive itself over to them.
Notice here how they even try to redefine the word โcontrolledโ to be about privacy (controlling your network privacy), instead ofโฆ being about actual control that matters.
They'll try to convince you that the battle is about privacy.
It's not, it's about control.
Azure ChatGPT Github
๐ฏ8๐ฅ2โค1
โDoctorGPT is a Large Language Model that can pass the US Medical Licensing Exam, Using Llamaโ
Waitโฆ that creator sounds familiar.
OH, itโs good old Siraj Raval, perhaps the sloppiest, most rediculous, most carefree faker and plaigarizer in modern AI.
Not a coincidence he chose the LLM most often used for fake scam benchmarks.
If lying were a sport, Siraj would be in the olympics and Llama would be his Nikes.
Gotta be another scam.
DoctorGPT Github
Data Science Influencer Siraj Raval Admits To Plagiarism
YouTuber Siraj Raval Caught Lying About Mining $800 in ETH with a Tesla
The Rise and Fall of Siraj Raval
Youtube: The Siraj Raval Controversy
Waitโฆ that creator sounds familiar.
OH, itโs good old Siraj Raval, perhaps the sloppiest, most rediculous, most carefree faker and plaigarizer in modern AI.
Not a coincidence he chose the LLM most often used for fake scam benchmarks.
If lying were a sport, Siraj would be in the olympics and Llama would be his Nikes.
Gotta be another scam.
DoctorGPT Github
Data Science Influencer Siraj Raval Admits To Plagiarism
YouTuber Siraj Raval Caught Lying About Mining $800 in ETH with a Tesla
The Rise and Fall of Siraj Raval
Youtube: The Siraj Raval Controversy
โค5๐3๐2๐ฅ1๐ฑ1๐1
Forwarded from Chat GPT
๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ
CHAD AI Meme Contest
ROUND 1 BEGINS
Prizes:
๐ฅ$100 of CHAD + secret prize
๐ฅ $50 of CHAD
Rules:
1๏ธโฃ Upload images to @chadgptcoin
2๏ธโฃ Each meme must contain words โChadGPTโ.
3๏ธโฃ Ranking according to /based and /unbased votes in @chadgptcoin.
4๏ธโฃ Ties decided by a runoff vote.
ENDS IN 9 HOURS = MIDNIGHT UTC
1st Round Starting Now!
๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ
CHAD AI Meme Contest
ROUND 1 BEGINS
Prizes:
๐ฅ$100 of CHAD + secret prize
๐ฅ $50 of CHAD
Rules:
1๏ธโฃ Upload images to @chadgptcoin
2๏ธโฃ Each meme must contain words โChadGPTโ.
3๏ธโฃ Ranking according to /based and /unbased votes in @chadgptcoin.
4๏ธโฃ Ties decided by a runoff vote.
ENDS IN 9 HOURS = MIDNIGHT UTC
1st Round Starting Now!
๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ
โค8๐6๐ฅ6๐6๐5๐คฉ4๐คฌ1
Following Pushback, Zoom Says It Won't Use Customer Data to Train AI Models
โZoom says it will walk back a recent change to its terms of service that allowed the company to use some customer content to train its machine learning and artificial intelligence models.โ
Article
โZoom says it will walk back a recent change to its terms of service that allowed the company to use some customer content to train its machine learning and artificial intelligence models.โ
Article
๐8๐5โค1
๐จSudden Death Round for Second Place ๐จ
Whose meme should win 2nd place, 8โs or Teeโs?
YESTERDAY'S MEME CONTEST #1:
1. 14 points for meme #8467 by shiro (919896696)
2. 10 points for meme #8947 by 8 (64818423)
3. 10 points for meme #8723 by Teeโณ๏ธ (310549823)
Whose meme should win 2nd place, 8โs or Teeโs?
YESTERDAY'S MEME CONTEST #1:
1. 14 points for meme #8467 by shiro (919896696)
2. 10 points for meme #8947 by 8 (64818423)
3. 10 points for meme #8723 by Teeโณ๏ธ (310549823)
๐2โค1
Whose meme should win 2nd place, 8โs or Teeโs?
Anonymous Poll
26%
8โs meme
35%
Teeโs meme
39%
Show results
๐2โค1
๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ
CHAD AI Meme Contest
ROUND 2 BEGINS
Prizes:
๐ฅ$100 of CHAD + secret prize
๐ฅ $50 of CHAD
Rules:
1๏ธโฃ Upload images to @chadgptcoin
2๏ธโฃ Each meme must contain โChadGPTโ.
3๏ธโฃ Ranking according to /based and /unbased votes in @chadgptcoin.
4๏ธโฃ Ties decided by a runoff vote.
ENDS IN 5 HOURS = MIDNIGHT UTC
2nd Round Starting Now!
๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ
CHAD AI Meme Contest
ROUND 2 BEGINS
Prizes:
๐ฅ$100 of CHAD + secret prize
๐ฅ $50 of CHAD
Rules:
1๏ธโฃ Upload images to @chadgptcoin
2๏ธโฃ Each meme must contain โChadGPTโ.
3๏ธโฃ Ranking according to /based and /unbased votes in @chadgptcoin.
4๏ธโฃ Ties decided by a runoff vote.
ENDS IN 5 HOURS = MIDNIGHT UTC
2nd Round Starting Now!
๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ๐จ๐ฟ
๐ฅ5๐ฅฐ5โค4๐3๐คฉ2๐1๐1๐1
Asked ChatGPT to remove password protection from an Excel document, and it worked flawlessly
Excel stores the password in plaintext in the raw XML that makes up the file content, so the password is easily removed.
This is considered security by Microsoft.
Same Microsoft thatโs now offering โsecureโ Azure ChatGPT for businesses.
Excel stores the password in plaintext in the raw XML that makes up the file content, so the password is easily removed.
This is considered security by Microsoft.
Same Microsoft thatโs now offering โsecureโ Azure ChatGPT for businesses.
๐8๐คฃ7โค1