Massively improving Twitch live chat moderation, by including chat context, instead of just classifying individual messages
โOur results show that appropriate contextual information can boost moderation performance by 35%.โ
= If you think AI powered censoring won't be effective, you're dead wrong. AI-based moderation will work extremely well, if those setting it up set it up correctly, which most just haven't bothered to, yet.
Paper
โOur results show that appropriate contextual information can boost moderation performance by 35%.โ
= If you think AI powered censoring won't be effective, you're dead wrong. AI-based moderation will work extremely well, if those setting it up set it up correctly, which most just haven't bothered to, yet.
Paper
๐5โค3๐2๐คฏ1
Harvardโs new computer science teacher is a chatbot
โOur own hope is that, through AI, we can eventually approximate a 1:1 teacher:student ratio for every student in CS50, as by providing them with software-based tools that, 24/7, can support their learning at a pace and in a style that works best for them individually,โ
โOur own hope is that, through AI, we can eventually approximate a 1:1 teacher:student ratio for every student in CS50, as by providing them with software-based tools that, 24/7, can support their learning at a pace and in a style that works best for them individually,โ
๐26๐6๐คฏ4โค1๐1
Once again trying to deny the 1st Bitter Lesson: โThe bigger-is-better approach to AI is running out of roadโ
โThis gigantism is becoming a problem. If Epoch aiโs ten-monthly doubling figure is right, then training costs could exceed a billion dollars by 2026โassuming, that is, models do not run out of data first.โ
= Combustion engines wonโt overtake horses, because that would mean that the car industry might be investing over a billion dollars in creating cars soon! Obviously no way that can happen!
Nonsense, not even a real argument.
โAn analysis published in October 2022 forecast that the stock of high-quality text for training may well be exhausted around the same time.โ
= Training will hit a brick wall because weโre running out of text! I.e. Itโs impossible to train LLMs without human-made training data.
Wrong. Already thorougly disproven since long before LLMs even existed, previously with MuZero & EfficientZero, and more recently with LLMs showing great success in learning from their own syntheticly generated training data. Self-supervised training data creation is not only theoretically possible but already widely done.
โAnd even once the training is complete, actually using the resulting model can be expensive as well. The bigger the model, the more it costs to run. Earlier this year Morgan Stanley, a bank, guessed that, were half of Googleโs searches to be handled by a current gpt-style program, it could cost the firm an additional $6bn a year.โ
= We canโt create huge models that, because theyโre expensive to run.
No, the opposite, surprisingly, and for reasons that are not yet fully understood. Emperically, and despite great effort trying to get around this, turns out the only way to get cheap-to-run powerful models is to first train a gigantic, extremely over-parameterized model, and then after dramatically prune that down into a smaller cheaper model.
Economist article trying to deny the 1st Bitter Lesson
โThis gigantism is becoming a problem. If Epoch aiโs ten-monthly doubling figure is right, then training costs could exceed a billion dollars by 2026โassuming, that is, models do not run out of data first.โ
= Combustion engines wonโt overtake horses, because that would mean that the car industry might be investing over a billion dollars in creating cars soon! Obviously no way that can happen!
Nonsense, not even a real argument.
โAn analysis published in October 2022 forecast that the stock of high-quality text for training may well be exhausted around the same time.โ
= Training will hit a brick wall because weโre running out of text! I.e. Itโs impossible to train LLMs without human-made training data.
Wrong. Already thorougly disproven since long before LLMs even existed, previously with MuZero & EfficientZero, and more recently with LLMs showing great success in learning from their own syntheticly generated training data. Self-supervised training data creation is not only theoretically possible but already widely done.
โAnd even once the training is complete, actually using the resulting model can be expensive as well. The bigger the model, the more it costs to run. Earlier this year Morgan Stanley, a bank, guessed that, were half of Googleโs searches to be handled by a current gpt-style program, it could cost the firm an additional $6bn a year.โ
= We canโt create huge models that, because theyโre expensive to run.
No, the opposite, surprisingly, and for reasons that are not yet fully understood. Emperically, and despite great effort trying to get around this, turns out the only way to get cheap-to-run powerful models is to first train a gigantic, extremely over-parameterized model, and then after dramatically prune that down into a smaller cheaper model.
Economist article trying to deny the 1st Bitter Lesson
๐ฅ5๐4๐ฑ2โค1๐ฏ1
Misleading chart used by The Economist to try to deny the 1st Bitter Lesson
Looks like itโs hitting a wall, and couldnโt possibly go much higher, right?
No.
ML training entered a new era.
Why?
Because, like relays and vacuum tubes and transistors at their start, LLMs suddenly reached minimum economic viability. They reached the point where their marginal productivity surpassed their marginal cost.
New era.
2018 OpenAI article explaining the new era
Looks like itโs hitting a wall, and couldnโt possibly go much higher, right?
No.
ML training entered a new era.
Why?
Because, like relays and vacuum tubes and transistors at their start, LLMs suddenly reached minimum economic viability. They reached the point where their marginal productivity surpassed their marginal cost.
New era.
2018 OpenAI article explaining the new era
๐5๐คฏ2โค1๐ฅ1๐1
Demolishing the โWeโre hitting a brick wall because weโre running out of human training dataโ theory - LARGE LANGUAGE MODELS CAN SELF-IMPROVE, Oct 2022
โWe show that it is possible for the LLM to self-improve even on its own generated questions and few-shot Chain-of-Thought prompts.โ
(Numerous subsequent papers further strongly confirming this.)
Paper
โWe show that it is possible for the LLM to self-improve even on its own generated questions and few-shot Chain-of-Thought prompts.โ
(Numerous subsequent papers further strongly confirming this.)
Paper
๐12๐คฌ2โค1๐ฏ1
Using GPT4 to Make an AI Bartender App
โLast time I made an app it took almost 6 months and nearly $10K in art costs. This time we built the app in about a week and then took a couple of weeks to test and refine it. We used GPT4 to help build the app, which is based on the OpenAIโs GPT API. We also used AI tools for all of the graphics, from the icon to the bartender animations, bar backgrounds, and voices. In this post weโll touch on the tools used to create the app. The cost of creating the app has been super-low.โ
Article
โLast time I made an app it took almost 6 months and nearly $10K in art costs. This time we built the app in about a week and then took a couple of weeks to test and refine it. We used GPT4 to help build the app, which is based on the OpenAIโs GPT API. We also used AI tools for all of the graphics, from the icon to the bartender animations, bar backgrounds, and voices. In this post weโll touch on the tools used to create the app. The cost of creating the app has been super-low.โ
Article
๐24โค4๐จโ๐ป2
retroactively applying generational labels backward through time
Founding Fathers are Generation L&M
Founding Fathers are Generation L&M
โค8๐2