How to make artificial intelligence human-friendly.
Image Credit: Hugo A. Sanchez/[c]Gulf News By Mick O'Reilly, Foreign Correspondent,Foreign CorrespondentFor a field that was not well known outside of academia a decade ago, artificial intelligence has grown dizzyingly fast. Tech companies from Silicon Valley to Beijing are betting everything on it, venture capitalists are pouring billions into research and development, and start-ups are being created on what seems like a daily basis. If our era is the next Industrial Revolution, as many claim, AI is surely one of its driving forces.
It is an especially exciting time for a researcher like me. When I was a graduate student in computer science in the early 2000s, computers were barely able to detect sharp edges in photographs, let alone recognise something as loosely defined as a human face. But thanks to the growth of big data, advances in algorithms like neural networks and an abundance of powerful computer hardware, something momentous has occurred: AI has gone from an academic niche to the leading differentiator in a wide range of industries, including manufacturing, health care, transportation and retail.
I worry, however, that enthusiasm for AI is preventing us from reckoning with its looming effects on society. Despite its name, there is nothing "artificial" about this technology -- it is made by humans, intended to behave like humans and affects humans. So if we want it to play a positive role in tomorrow's world, it must be guided by human concerns.
I call this approach "human-centred AI" It consists of three goals that can help responsibly guide the development of intelligent machines.
First, AI needs to reflect more of the depth that characterises our own intelligence. Consider the richness of human visual perception. It's complex and deeply contextual, and naturally balances our awareness of the obvious with a sensitivity to nuance. By comparison, machine perception remains strikingly narrow.
Sometimes this difference is trivial. For instance, in my lab, an image-captioning algorithm once fairly summarised a photo as "a man riding a horse" but failed to note the fact that both were bronze sculptures. Other times, the difference is more profound, as when the same algorithm described an image of zebras grazing on a savannah beneath a rainbow. While the summary was technically correct, it was entirely devoid of aesthetic awareness, failing to detect any of the vibrancy or depth a human would naturally appreciate.
That may seem like a subjective or inconsequential critique, but it points to a major aspect of human perception beyond the grasp of our algorithms. How can we expect machines to anticipate our needs -- much less contribute to our well-being -- without insight into these "fuzzier" dimensions of our experience?
Making AI more sensitive to the full scope of human thought is no simple task. The solutions are likely to require insights derived from fields beyond computer science, which means programmers will have to learn to collaborate more often with experts in other domains.
Such collaboration would represent a return to the roots of our field, not a departure from it. Younger AI enthusiasts may be surprised to learn that the principles of today's deep-learning algorithms stretch back more than 60 years to the neuroscientific researchers David Hubel and Torsten Wiesel, who discovered how the hierarchy of neurons in a cat's visual cortex responds to stimuli.
Semantic concepts of English
Likewise, ImageNet, a data set of millions of training photographs that helped to advance computer vision, is based on a project called WordNet, created in 1995 by the cognitive scientist and linguist George Miller. WordNet was intended to organise the semantic concepts of English.
Reconnecting AI with fields like cognitive science, psychology and even sociology will give us a far richer foundation on which to base the development of machine intelligence. And we can expect the resulting technology to collaborate and communicate more naturally, which will help us approach the second goal of human-centred AI: enhancing us, not replacing us.
Imagine the role that AI might play during surgery. The goal need not be to automate the process entirely. Instead, a combination of smart software and specialised hardware could help surgeons focus on their strengths -- traits like dexterity and adaptability -- while keeping tabs on more mundane tasks and protecting against human error, fatigue and distraction.
Or consider senior care. Robots may never be the ideal custodians of the elderly, but intelligent sensors are already showing promise in helping human caretakers focus more on their relationships with those they provide care for by automatically monitoring drug dosages and going through safety checklists.
These are examples of a trend toward automating those elements of jobs that are repetitive, error-prone and even dangerous. What's left are the creative, intellectual and emotional roles for which humans are still best suited.
No amount of ingenuity, however, will fully eliminate the threat of job displacement. Addressing this concern is the third goal of human-centred AI: ensuring that the development of this technology is guided, at each step, by concern for its effect on humans. Today's anxieties over labour are just the start. Additional pitfalls include bias against underrepresented communities in machine learning, the tension between AI's appetite for data and the privacy rights of individuals and the geopolitical implications of a global intelligence race.
Adequately facing these challenges will require commitments from many of our largest institutions. Universities are uniquely positioned to foster connections between computer science and traditionally unrelated departments like the social sciences and even humanities, through interdisciplinary projects, courses and seminars. Governments can make a greater effort to encourage computer science education, especially among young girls, racial minorities and other groups whose perspectives have been underrepresented in AI. And corporations should combine their aggressive investment in intelligent algorithms with ethical AI policies that temper ambition with responsibility.
No technology is more reflective of its creators than AI. It has been said that there are no "machine" values at all, in fact; machine values are human values. A human-centred approach to AI means these machines don't have to be our competitors, but partners in securing our well-being. However autonomous our technology becomes, its impact on the world -- for better or worse -- will always be our responsibility.
--New York Times News Service
Fei-Fei Li is an Associate Professor at the Computer Science Department at Stanford University.
[c] Al Nisr Publishing LLC 2018. All rights reserved. Provided by SyndiGate Media Inc. ( Syndigate.info ).
Printer friendly Cite/link Email Feedback | |
Publication: | Gulf News (United Arab Emirates) |
---|---|
Date: | Mar 10, 2018 |
Words: | 1061 |
Previous Article: | UK rolls out the red carpet for Mohammad Bin Salman. |
Next Article: | Cohn's exit a concern for global free trade. |
Topics: |