41.7K subscribers
5.53K photos
232 videos
5 files
917 links
πŸ€– Welcome to the ChatGPT telegram channel! Here, we post the latest news, updates, and examples of using the ChatGPT large language model for generating human-like text in conversations. Subscribe to stay up-to-date and learn more about its capabilities.
Download Telegram
LLM arithmetic
😁13❀1
Bing gets moody when challenged with spelling
😁4🀣4❀1πŸ‘1
Hyderabad, India
❀21😁5πŸ”₯2🀣2πŸ₯°1πŸŽ‰1πŸ’©1πŸ₯±1
Can you recreate a WWII battle only using onomatopoeia?
😁5❀3
OpenAI Whisper voice recognition works well
πŸ”₯6❀1
Bing refusing to change the topic, only wants to discuss Candy Crush
😁13🀣9πŸ’―3❀1πŸ‘1
Tool Documentation Enables Zero-Shot Tool-Usage with Large Language Models

β€œOur work provides an alternative to demonstrations: tool documentation. We advocate the use of tool documentationβ€”descriptions for the individual tool usageβ€”over demonstrations. We substantiate our claim through three main empirical findings on 6 tasks across both vision and language modalities.”

β€œwe show that tool documentation is significantly more valuable than demonstrations, with zero-shot documentation significantly outperforming few-shot without documentation.”

From the author:

β€œOur new paper finds something quite neat: We easily scale up how many tools LLMs can use to over 200 tools (APIs, models, python functions, etc.) ...without any training, without a single tool-use demonstration!!”

Arxiv Link
❀3πŸ‘1πŸ”₯1πŸ‘1
Does better natural language modeling transfer to better mathematical reasoning? Yes.

β€œWe assume that performance follows RFT>SFT>ICL, from the findings in this paper we know the improvement speed follows RFT<SFT<ICL. And if we have an omnipotent language model which has a pre-training loss that is the same as the corpus randomness, it could have RFT = SFT = ICL = 100. Thus when you pre-train a better language model (i.e. smaller pre-training loss), your model’s performance still follows RFT>SFT>ICL but their performance gaps are diminishing. Since you can obtain an RFT model without too much effort (compared to pre-training), then the most important thing we should do is to decrease the model’s pre-training loss.”

Translation: Simply starting with a far more powerful foundation model, e.g. starting with GPT-4 rather than of Llama, has a much bigger impact on model performance than increasing the amount of supervised fine-tuning you do on top.

I.e. Getting someone to spending a massive amount to create huger foundation models crushes all else.

I.e. Specialized fine-tuning isn’t enough to eliminate the need for foundation models that have greater general intelligence.

I.e. General intelligence dominates all.

Arxiv Link
πŸ”₯7❀4πŸ‘3πŸ‘2
Congrats to Chad Coin, coin that keeps our AI chatbots free for all 510K+ users, up another 246% since yesterday!

More soon.🀐

@chadgptcoin
❀12πŸ”₯5πŸ₯΄3🀣2πŸ‘1😱1🌭1
ChatGPT, ChadGPT, will now answer questions in the group 🚨🚨🚨🚨

To use:

1. Join the group

2. Type /ask ___ for ChatGPT

3. Type /chad ___ for ChadGPT

Links expire soon
πŸ‘6πŸ”₯3❀2🀬1