Featured Product
This Week in Quality Digest Live
Innovation Features
Jesse Lyn Stoner
No athlete ever achieved a gold medal by going for the bronze
Caroline Preston
Doubling down on technical skills isn’t the answer
Jennifer Chu
Techniques could lead to personalized wearable and implantable devices
Knowledge at Wharton
From global internet to innovation labs, it all starts 50 miles above sea level
Sarah Webb
Technology is enabling increasingly lifelike models of organs to help doctors practice operations

More Features

Innovation News
Inkbit is overcoming traditional constraints to 3D printing by giving its machines ‘eyes and brains’
Deep Reality Viewer creates stereo, high-definition 3D images without using a monitor
Wing’s shape shifts automatically for aerodynamic loads (e.g., takeoff, landing, cruising)
Process concerns technology feasibility, commercial potential, and transition to marketplace
Identifying the 252 needs for workforce development to meet our future is a complex, wicked, and urgent problem
How established companies turn the tables on digital disruptors
Building organizational capability and capacity to create outcomes that matter most
‘What if we could develop electronic systems that we wrap around a bridge or cover the walls of our office?’
System breaks down complex designs into easily modifiable shapes for custom manufacturing and 3D printing

More News

Theodoros Evgeniou


The Pivotal Management Challenge of the AI Era

Managers and leaders have nothing to fear from AI—except missing out

Published: Monday, April 22, 2019 - 12:03

History indicates that major technological changes can take about half a century to go from the first lab drawings to society. Alan Turing first proposed the Turing machine, laying the foundations of computation, in 1936; the first general-purpose “Turing-complete” system was built in 1945, and “The Computer” was only named “Machine of the Year” by Time in 1982, about half a century later.

The foundations of the internet were laid out during the 1960s, but consumers did’t get to broadly use and benefit from it until the mid- to late 1990s. For most people, artificial intelligence was strictly a sci-fi concept until recent years.

Yet, if you go by the above timeline, the AI revolution may actually be running at least two decades late. It has been more than 70 years since the famous 1956 Dartmouth workshop with Allen Newell, Herbert Simon, John McCarthy, and Marvin Minsky—the last of them passing away in 2016—during which the first AI program was officially unveiled. But statistical learning theory, the foundation of modern AI and machine learning, arrived a little ahead of the 50-year deadline. The field (whose luminaries included Vladimir Vapnik, Tomaso Poggio, and Steve Smale) cross-pollinated statistics, mathematics, and computer science to produce a flowering of breakthroughs, leading directly to today’s AI revolution.

AI and its effect on business, governments, and society may be today where physics was at the turn of the previous century. After breakthroughs in physics at that time, and later in biology and other fields, the world became very different from what it was for thousands of years before.

At the moment, companies and countries are scrambling to unleash AI’s pent-up transformational potential ahead of competitors. Due to the rapid pace of technological change and the increasingly winner-take-all nature of the innovation economy, the victors of the AI race may capture the lion’s share of the spoils—even if they win by only a hair.

A recipe for winning the AI race

So who has the advantage in the AI race? Managing technology adoption is a very old topic. Philosophers, sociologists, management scholars, economists, and engineers have pondered this matter at least since the invention of the computer, if not long before, in the case of other types of technologies such as the mechanization of mining or other analog technologies.

Surveying this body of work, what sticks out is a number of people-focused features that determine whether new tech will thrive in a given context. At the organizational level, it is essential for leaders to foster technological skills; IT infrastructure and governance; data literacy; innovation culture; norms that adhere to best practices; and most important, the ability to align the capacities of new technologies with the needs of the core business. At the country level, the same required competencies apply, but there are two important additions to the list: regulations that keep pace with technological developments, such as the General Data Protection Regulation (GDPR), and an education system that is strong not only in science, technology, engineering, and math (STEM), but across the board.

At this point, a sceptic may ask whether an AI race can be won by humans in the first place. Much of the media discourse about AI has centered on the idea of robots taking work away from humans. Is AI simply a Pandora’s box that, when opened, dooms us to a future of irrelevancy? In response, I would point to the financial sector. Computers are now executing trades in the financial markets at speeds that were impossible to conceive just a couple of decades ago. This brought down trading costs and bid-ask spreads at levels never seen before and also making markets—the arteries of our modern economic system—impressively efficient. Indeed, some jobs have been displaced, but major trading decisions are still made, and communicated to clients, by humans. The main difference is that people are doing so under increasingly different conditions: The trading floors have long been a leading example of the modern workplace where humans and machines work together in increasingly intertwined ways. 

Kasparov’s law

Garry Kasparov, the Russian chess grandmaster who has been closely working with machines since his famous defeat by Deep Blue in 1997, summarized his 20 years of experience with what’s now called “Kasparov’s law”: “Weak human + machine + better process beats strong human + machine + inferior process.” The key component in this law is the “better process.”

Management and leadership scholars have spent the majority of the 20th century developing the principles of “people managing people.” The central management question of the future will be “How can we get humans and machines to best work together?” It is not about AI vs. people; it is about AI and people. At INSEAD, we are now working on this question in contexts ranging from finance and investment decision making, to consumer understanding and purchasing, to medical diagnosis and healthcare.

Can human work be massively enhanced with the right use of AI? Absolutely. Can, in a more “science fiction” way, humans themselves be enhanced using the results of scientific and technological advances driven by AI—from fields like computational biology or computational neuroscience? Remains to be seen. You can choose to view this as scary or hopeful, but it is definitely not boring.

First published April 8, 2019, on the INSEAD Knowledge blog.


About The Author

Theodoros Evgeniou’s picture

Theodoros Evgeniou

Theodoros Evgeniou is a professor of decision sciences and technology management at INSEAD, and academic director of INSEAD elab, the research and analytics center of INSEAD that focuses on data analytics for business.