“Perhaps the most important news of our day is that datasets — not algorithms — might be the key limiting factor to development of human-level artificial intelligence,” according to Alexander Wissner-Gross in a written response to the question posed by Edge: “What do you consider the most interesting recent scientific news?”
At the dawn of the dield of artificial intelligence, two of its founders famously predicted that solving the problem of machine vision would only take a summer. We now know that they were off by half a century. Wissner-Gross began to ponder the question of: “What took the AI revolution so long?” By reviewing the timing of the most publicized AI advances over the past 30 years, he found evidence that suggests a provocative explanation: perhaps many major AI breakthroughs have actually been constrained by the availability of high-quality training datasets, and not by algorithmic advances. Here we summarize the key AI milestones:
If true, this hypothesis have foundational implications for future progress in AI. For example, prioritizing the cultivation of high-quality training datasets might allow an order-of-magnitude speedup in AI breakthroughs over purely algorithmic advances. After all, focusing on dataset rather than algorithm is a potentially simpler approach. “Although new algorithms receive much of the public credit for ending the last AI winter,” concluded Alexander Wissner-Gross, “the real news might be that prioritizing the cultivation of new datasets and research communities around them could be essential to extending the present AI summer.”
We wonder if algorithmic trading systems might similarly benefit from the cultivation of new datasets and research communities around them. What might that look like? How do we learn to work with imperfect data? What are the risks of trusting the data too much?