I evaluated three AI systems—Manus, Claude Chrome Extension, and Claude Cowork—on a complex planning task for an Akkadian translation competition, and none of them produced genuinely useful strategic analysis. All three independently generated the same obvious recommendation (fine-tune ByT5, augment data, run Optuna) with the same fabricated probability estimates, revealing that they're pattern-matching to surface-level solutions rather than doing real analytical work. The differences between them are purely cosmetic: one writes like an academic, one scrapes leaderboard numbers, one includes boilerplate code—but underneath the formatting, there's no there there.
Write up a well-researched plan for competing in https://www.kaggle.com/competitions/deep-past-initiative-machine-translation the Deep Past Challenge - Translate Akkadian to English on Kaggle. Read through all discussions https://www.kaggle.com/competitions/deep-past-initiative-machi