Just gave deepseek a 1800 rated question with tutorial explaination it still couldnt generate a correct solution. In my opinion LLMs as of now are just good at solving pretty standard questions anything adhoc/constructive above 1500 with multiple edge cases it starts acting like a retard.
honestly llms are gonna remain good at doing things that are already done. True creativity is gonna take a long time.
To be fair, most cf users are no smarter than deepseek in this aspect.
A fun fact is that I tested DeepSeek on Problems B and C from yesterday's ARC after the contest ended. I think Problem C,which was a constructive problem,is so ad-hoc that I couldn't solve it during the contest. However,deepseek solved them correctly and shocked me a lot.
how about perplexity ai?
Perplexity is based on OpenAI's GPT.
probably the solution was out by then? I see a pattern where ai get better at solving after the contest is over. Extremely fishy.
I made the test just about 10 minutes after the contest ended. Though the editorial was released at that time, I don't think it was quick enough to learn the solution from it.
Wow, I didn't solve it either. wonderful llm