Check out this 3 year old tool trained on GPT-2 data.
Work for you guys?
https://huggingface.co/openai-detector
Work for you guys?
https://huggingface.co/openai-detector
parth007_96βs brilliant notes on reverse-engineering GitHub Copilot:
https://thakkarparth007.github.io/copilot-explorer/posts/copilot-internals
https://thakkarparth007.github.io/copilot-explorer/posts/copilot-internals
GPT-3/LLMs' Achilles heel is short context length - how many "in-context" examples they can consume to learn a new task.
Enter "Structured Prompting": scale your examples from dozens => 1,000+
Here's how:
=> Get 1000s of in-context samples
=> split them into M groups, each small enough to fit in regular context length
=> encode each of M groups using LLM encoder
=> combine these encoded groups and attend over a scaled version of the combination simultaneously
Paper: https://arxiv.org/pdf/2212.06713.pdf
Code: https://github.com/microsoft/LMOps
Enter "Structured Prompting": scale your examples from dozens => 1,000+
Here's how:
=> Get 1000s of in-context samples
=> split them into M groups, each small enough to fit in regular context length
=> encode each of M groups using LLM encoder
=> combine these encoded groups and attend over a scaled version of the combination simultaneously
Paper: https://arxiv.org/pdf/2212.06713.pdf
Code: https://github.com/microsoft/LMOps
π1