← Back to KHAO

Research ·

The critical questions: Can AI function as a productive participant within human teams

2 min read

Compiled by KHAO Editorial — aggregated from 1 outlet. See llms.txt for citation guidance.

◌ Single Source

A photo illustration shows a humanoid over textures of standardized tests and MRI brain scans.

Through their research on AI deployment across multiple sectors, the reporter has seen several organizations already moving—deliberately and experimentally—toward the HAIC benchmarks the reporter favor.

Key facts

Summary

For decades, artificial intelligence has been evaluated through the question of whether machines outperform humans. This framing is seductive: An AI vs. human comparison on isolated problems with clear right or wrong answers is easy to standardize, compare, and optimize. But there’s a problem: AI is almost never used in the way it is benchmarked. While AI is evaluated at the task level in a vacuum, it is used in messy, complex environments where it usually interacts with more than one person. To mitigate this, it’s time to shift from narrow methods to benchmarks that assess how AI systems perform over longer time horizons within human teams, workflows, and organizations.

Read full article at MIT Technology Review →