A live agent spends hours each week manually documenting routine interactions. Another combs through multiple knowledge bases to find the right solution, scrambling to piece it together while the customer waits on hold. A third types out the same response they’ve written dozens of times before. These repetitive tasks can be draining, leaving less time…
This is today’s edition of The Download, our weekday newsletter that provides a daily dose of what’s going on in the world of technology. The way we measure progress in AI is terrible Every time a new AI model is released, it’s typically touted as acing its performance against a series of benchmarks. OpenAI’s GPT-4o, for example,…
Generative AI models have become remarkably good at conversing with us, and creating images, videos, and music for us, but they’re not all that good at doing things for us. AI agents promise to change that. Think of them as AI models with a script and a purpose. They tend to come in one of…
Every time a new AI model is released, it’s typically touted as acing its performance against a series of benchmarks. OpenAI’s GPT-4o, for example, was launched in May with a compilation of results that showed its performance topping every other AI company’s latest model in several tests. The problem is that these benchmarks are poorly…