"Evaluating Large Language Models Trained on Code",
arxiv.org/abs/2107.03374A detailed description of early (pre-GitHub Copilot) versions of OpenAI Codex. This is the "paper of the year" so far: we finally have real progress in AI-assisted computer programming (and difficulties of computer programming form the key bottleneck limiting the speed of
(
Read more... )
Comments 10
An early-GPT-4-level model is now open source, reasonably compact, fine-tunable (LLama-3-70B in base and instruct configurations).
GPT-4o is two "early GPT-4 minus GPT-3.5" deltas above that, and OpenAI has something much stronger than that internally.
A lot of novel architectural decisions were being published in recent months.
Reply
no, actually, only 1.5 deltas above that
Reply
Reply
o1-preview and o1-mini (reasoning and deliberation; System 2 "slow thinking" over System 1 "fast LLM thinking")
a lot of interesting alternative architectures and algorithms
multi-agents over LLMs
Reply
Leave a comment