Researchers at ByteDance have released DAComp, a benchmark created to test how well AI can manage the full analytics lifecycle ā from data processing to insight and recommendation. And they've looked at how well state-of-the art models perform on these benchmarks today.ā£
ā£
šššš: AI is now an OK analyst (for less complex tasks) and a terrible strategist.ā£
ā£
Their paper looks at AI performance across two dimensions:ā£
ā£
(š) šš”š āš”šš«šā šš¢š¦šš§š¬š¢šØš§ (šššš”š§š¢ššš„ šš±ššš®šš¢šØš§)ā£
ā£š» This includes breaking down a business question, planning analyses, writing queries, running calculations, and producing charts.⣠AI can handle many of these steps, but performance is uneven ā it does well on straightforward descriptive tasks, yet reliability drops when logic becomes layered or when multiple analytical steps must be sequenced correctly.ā£
ā£
(š) šš”š āš¬šØššā šš¢š¦šš§š¬š¢šØš§ (šš§šš„š²šš¢ššš„ š«ššš¬šØš§š¢š§š & š¢š§ššš«š©š«ššššš¢šØš§)ā£
ā£š§ This involves interpreting results, identifying drivers, weighing explanations, and turning findings into clear recommendations.⣠Here, AI still struggles: it can summarise outputs, but often cannot deliver grounded diagnosis or meaningful strategic direction.ā£
ā£
My take-away is that AI accelerates parts of the analytical workflow, but its limitations grow as tasks become more complex and interpretive, and human expertise remains essential for context, judgment, and synthesis.ā£
ā£
Good initiative to create a dedicated benchmark for this.ā£
Link to paper:https://arxiv.org/pdf/2512.04324


