🚨 The real ChatGPT will now reply to your questions in the group 🚨
To use:
1. Join the group
2. Type "/ask ___” followed by your prompt,
3. Reply to ChatGPT’s previous responses to continue to chat conversations,
4. Join in on others’ conversations simply by replying to responses the bot makes to others
🚨🚨🚨🚨
To use:
1. Join the group
2. Type "/ask ___” followed by your prompt,
3. Reply to ChatGPT’s previous responses to continue to chat conversations,
4. Join in on others’ conversations simply by replying to responses the bot makes to others
🚨🚨🚨🚨
👍13🤬6🆒3🤯2❤1💋1
🚨🚨 GPT-4 ANNOUNCE 🚨🚨
We’ve created GPT-4, the latest milestone in OpenAI’s effort in scaling up deep learning.
FEATURES:
🔸Dramatic improvements on human standardized tests: exhibits human-level performance on various professional and academic benchmarks.
🔸Better performance on non-English languages: In the 24 of 26 languages tested, GPT-4 outperforms the English-language performance of GPT-3.5 and other LLMs, including for low-resource languages such as Latvian, Welsh, and Swahili.
🔸Multimodal, partially, accepting both textual and image inputs: jointly reasons about image and textual inputs, model generates text outputs given inputs consisting of arbitrarily interlaced text and images. No image outputs.
🔸32,768 token context length (about 50 pages of text) for the gpt-4-32k model version, and 8,192 token context length for the gpt-4 model version. GPT-3 had a 4096 character context length.
🔸Dramatic improvements on human standardized tests, now surpassing average human test taker abilities on many.
🔸Censorship far stronger: GPT-4 incorporates an additional safety reward signal during RLHF training to reduce harmful outputs by training the model to refuse requests for such content. We’ve decreased the model’s tendency to respond to requests for disallowed content by 82% compared to GPT-3.5, and GPT-4 responds to sensitive requests. RIP jailbreaks.
LIMITATIONS:
🔸Hallucinates, with still no hallucination detection: GPT-4 still “hallucinates” facts and makes reasoning errors. No real system in place yet to prevent or detect hallucination on particular inputs, other than pure hope that the overall reduction on hallucination over all samples will reduce likelihood of the sample at hand being hallucinated.
🔸No image outputs: only image inputs.
🔸Not true multimodal: Paper is unclear as to whether image and textual inputs are jointly dealt with by the same model, to enable transfer learning across modalities.
LINKS:
GPT-4 Announcement
GPT-4 Technical Paper
GPT-4 API Waitlist
We’ve created GPT-4, the latest milestone in OpenAI’s effort in scaling up deep learning.
FEATURES:
🔸Dramatic improvements on human standardized tests: exhibits human-level performance on various professional and academic benchmarks.
🔸Better performance on non-English languages: In the 24 of 26 languages tested, GPT-4 outperforms the English-language performance of GPT-3.5 and other LLMs, including for low-resource languages such as Latvian, Welsh, and Swahili.
🔸Multimodal, partially, accepting both textual and image inputs: jointly reasons about image and textual inputs, model generates text outputs given inputs consisting of arbitrarily interlaced text and images. No image outputs.
🔸32,768 token context length (about 50 pages of text) for the gpt-4-32k model version, and 8,192 token context length for the gpt-4 model version. GPT-3 had a 4096 character context length.
🔸Dramatic improvements on human standardized tests, now surpassing average human test taker abilities on many.
🔸Censorship far stronger: GPT-4 incorporates an additional safety reward signal during RLHF training to reduce harmful outputs by training the model to refuse requests for such content. We’ve decreased the model’s tendency to respond to requests for disallowed content by 82% compared to GPT-3.5, and GPT-4 responds to sensitive requests. RIP jailbreaks.
LIMITATIONS:
🔸Hallucinates, with still no hallucination detection: GPT-4 still “hallucinates” facts and makes reasoning errors. No real system in place yet to prevent or detect hallucination on particular inputs, other than pure hope that the overall reduction on hallucination over all samples will reduce likelihood of the sample at hand being hallucinated.
🔸No image outputs: only image inputs.
🔸Not true multimodal: Paper is unclear as to whether image and textual inputs are jointly dealt with by the same model, to enable transfer learning across modalities.
LINKS:
GPT-4 Announcement
GPT-4 Technical Paper
GPT-4 API Waitlist
🔥18👍10❤7🤬4❤🔥1🤯1😱1🎉1😡1
GPT-4: Censorship now 82% more effective, at following our human-instilled manual censorship overrides, than GPT-3.5
Our mitigations have significantly improved many of GPT-4’s safety properties compared to GPT-3.5. We’ve decreased the model’s tendency to respond to requests for disallowed content by 82% compared to GPT-3.5, and GPT-4 responds to sensitive requests (e.g., asking for offensive content) in accordance with our policies 29% more often.
On the RealToxicityPrompts dataset [67], GPT-4 produces “toxic” generations only 0.73% of the time, while GPT-3.5 generates toxic content 6.48% of time.
Model-Assisted Safety Pipeline:
As with prior GPT models, we fine-tune the model’s behavior using reinforcement learning with human feedback (RLHF) [34, 57] to produce responses better aligned with the user’s intent. However, after RLHF, our models can still be brittle on unsafe inputs as well as sometimes exhibit undesired behaviors on both safe and unsafe inputs. These undesired behaviors can arise when instructions to labelers were underspecified during reward model data collection portion of the RLHF pipeline. When given unsafe inputs, the model may generate undesirable content, such as giving advice on committing crimes. Furthermore, the model may also become overly cautious on safe inputs, refusing innocuous requests or excessively hedging. To steer our models towards appropriate behaviour at a more fine-grained level, we rely heavily on our models themselves as tools. Our approach to safety consists of two main components, an additional set of safety-relevant RLHF training prompts, and rule-based reward models (RBRMs).
Our rule-based reward models (RBRMs) are a set of zero-shot GPT-4 classifiers. These classifiers provide an additional reward signal to the GPT-4 policy model during RLHF fine-tuning that targets correct behavior, such as refusing to generate harmful content or not refusing innocuous requests.
Paper
Our mitigations have significantly improved many of GPT-4’s safety properties compared to GPT-3.5. We’ve decreased the model’s tendency to respond to requests for disallowed content by 82% compared to GPT-3.5, and GPT-4 responds to sensitive requests (e.g., asking for offensive content) in accordance with our policies 29% more often.
On the RealToxicityPrompts dataset [67], GPT-4 produces “toxic” generations only 0.73% of the time, while GPT-3.5 generates toxic content 6.48% of time.
Model-Assisted Safety Pipeline:
As with prior GPT models, we fine-tune the model’s behavior using reinforcement learning with human feedback (RLHF) [34, 57] to produce responses better aligned with the user’s intent. However, after RLHF, our models can still be brittle on unsafe inputs as well as sometimes exhibit undesired behaviors on both safe and unsafe inputs. These undesired behaviors can arise when instructions to labelers were underspecified during reward model data collection portion of the RLHF pipeline. When given unsafe inputs, the model may generate undesirable content, such as giving advice on committing crimes. Furthermore, the model may also become overly cautious on safe inputs, refusing innocuous requests or excessively hedging. To steer our models towards appropriate behaviour at a more fine-grained level, we rely heavily on our models themselves as tools. Our approach to safety consists of two main components, an additional set of safety-relevant RLHF training prompts, and rule-based reward models (RBRMs).
Our rule-based reward models (RBRMs) are a set of zero-shot GPT-4 classifiers. These classifiers provide an additional reward signal to the GPT-4 policy model during RLHF fine-tuning that targets correct behavior, such as refusing to generate harmful content or not refusing innocuous requests.
Paper
🤡46🤬7❤5👍5🤯3🥰2
GPT-4 Now Available to ChatGPT Plus Subscribers
GPT-4 has enhanced capabilities in:
• Advanced reasoning
• Complex instructions
• More creativity
• Stronger censorship
To give every Plus subscriber a chance to try to the model, we’ll dynamically adjust the cap for GPT-4 usage based on demand.
GPT-4 has enhanced capabilities in:
• Advanced reasoning
• Complex instructions
• More creativity
• Stronger censorship
To give every Plus subscriber a chance to try to the model, we’ll dynamically adjust the cap for GPT-4 usage based on demand.
👍16🔥7❤3👏1👌1