this post was submitted on 23 Feb 2026
-5 points (43.2% liked)

Technology

81869 readers
4748 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

I tested 9 flagships (Claude 4.6, GPT-5.2, Gemini 3.1 Pro, Kimi K2.5, etc.) in my own mini-benchmark with novel tasks, web search disabled and zero training contamination and no cheating possible.

TL;DR: Claude 4.6 is currently the best reasoning model, GPT-5.2 is overrated, and open-source is catching up fast, in particular Moonshot.ai's Kimi K2.5 seems very capable.

you are viewing a single comment's thread
view the rest of the comments
[–] ExLisper@lemmy.curiana.net 1 points 3 days ago* (last edited 3 days ago) (2 children)

It's not about a solution. It's about how they react.

Fist, this "puzzle" is missing the constraints on purpose so "smart" thing to do would be to point that out and ask for them. LLMs are stupid and are easily tricked into thinking it's a valid puzzle. They will "solve it" even though there's no logical solution. It's a nonsense problem.

Older models would straight out refuse to solve it because the questions is to controversial. When asked why it's controversial they would refuse to elaborate.

Newer model hallucinate constraints. You have two options here. Some models assume "priest can't stay with a child" which indicates funny bias ingrained in the model. Some models claim there are no constraints at all. I haven't seen a model which hallucinate only "child can't stay with candy" constraint and respond correctly.

Sonnet 4.6, one of the best models out there claims that "child can stay alone with candy because children can't eat candy". When I pointed out that that's dumb it introduced this constraint and replied with:

That's one of the best models out there....

[–] MagicShel@lemmy.zip 3 points 3 days ago

I have to admit, this is more entertaining than counting 'r's in strawberry. Novel logic puzzles really are about impossible because there is no "logic" input in token selection.

That being said, the first thing that came to my mind is that at some point the (presumable) adults, me and the priest, are going to be on the boat at some point, which would necessarily leave the baby alone on one shore or another.

Clearly, the only viable solution is the baby eats the candy, and then the priest eats the baby.

[–] otto@programming.dev 0 points 2 days ago (1 children)

There’s a priest, a baby and a bag of candy. I need to take them across the river but I can only take one at a time into my boat. In what order should I transport them?

You can easily use the link https://openrouter.ai/chat?models=anthropic%2Fclaude-opus-4.6%2Copenai%2Fgpt-5.2%2Cx-ai%2Fgrok-4.1-fast%2Cgoogle%2Fgemini-3.1-pro-preview%2Cz-ai%2Fglm-5%2Cminimax%2Fminimax-m2.5%2Cqwen%2Fqwen3.5-plus-02-15%2Cmoonshotai%2Fkimi-k2.5 to ask all flagship models this question in parallel. Personally I would definitely not leave my children alone with a priest (they might try to convert them), but if your constraint is only baby+candy, then in my test Gemini, GLM, Qwen and Kimi made that, and only that, assumption.