Submitted by enryu42 t3_122ppu0 in MachineLearning
https://medium.com/@enryu9000/gpt4-and-coding-problems-8fbf04fa8134
Apparently it cannot solve coding problems which require any amount of thinking. LeetCode examples were most likely data leakage.
Such drastic gap between MMLU performance and end-to-end coding is somewhat surprising. <sarcasm>Looks like AGI is not here yet.</sarcasm> Thoughts?
addition t1_jdrsas2 wrote
I’ve become increasingly convinced that the next step for AI is adding some sort of feedback loop so that the AI can react to its own output.
There is increasing evidence that this is true. Chain-of-thought prompting, reflexon, and Anthropic’s constitutional AI all point in this direction.
I find constitutional AI to be particularly interesting because it suggests that after an LLM reaches a certain threshold of language understanding that it can start to assess its own outputs during training.