The promise of voice AI lies at work, not at home

Conversational AI has been around for a few years now, in our phones, smart speakers, and throughout connected homes. And while the adoption of this tech is steadily increasing, most applications in this space essentially only enable you to push buttons using your voice.

The next frontier for conversational AI lies beyond your doorstep. From education to workforce automation to retail, the scope for plenty of novel new applications is opening up as the technology improves.

As Shreyas Nivas of Replica wrote recently, we will see a shift in voice AI from services being “primarily transactional — ‘Alexa, tell me the weather’ — to being based on dynamic interactions and relationships between characters in any digital narrative or experience.”

There are gains to be made in better understanding nuance in human speech, deriving more insights from the data in users’ commands and questions, interaction personalization, and delivering more natural responses.

To that end, companies like NYC-based Rain are working on solutions to leverage these advancements and deploy them across organizations — and even help brands create unique voice-powered experiences for their customers.

Rain’s CEO, Nitya Thadani, will be joining me for a fireside chat about the future of conversational AI, what we can look forward to accomplishing with our voices, and how businesses can think about integrating voice technologies into their operations and offerings in the future. Tune in to our conversation at TNW2020 next month.

Secure your free ticket to TNW2020 now!

Computers suck at ‘common sense’ — AI expert explains why

Imagine you’re having friends over for lunch and plan to order a pepperoni pizza. You recall Amy mentioning that Susie had stopped eating meat. You try calling Susie, but when she doesn’t pick up, you decide to play it safe and just order a margherita pizza instead.

People take for granted the ability to deal with situations like these on a regular basis. In reality, in accomplishing these feats, humans are relying on not one but a powerful set of universal abilities known as common sense.

As an artificial intelligence researcher , my work is part of a broad effort to give computers a semblance of common sense. It’s an extremely challenging effort.

Quick – define common sense

Despite being both universal and essential to how humans understand the world around them and learn, common sense has defied a single precise definition. G. K. Chesterton, an English philosopher, and theologian, famously wrote at the turn of the 20th century that “common sense is a wild thing, savage, and beyond rules.” Modern definitions today agree that, at minimum, it is a natural, rather than formally taught, human ability that allows people to navigate daily life.

Common sense is unusually broad and includes not only social abilities, like managing expectations and reasoning about other people’s emotions, but also a naive sense of physics , such as knowing that a heavy rock cannot be safely placed on a flimsy plastic table. Naive, because people know such things despite not consciously working through physics equations.

Common sense also includes background knowledge of abstract notions, such as time, space and events. This knowledge allows people to plan, estimate and organize without having to be too exact.

Common sense is hard to compute

Intriguingly, common sense has been an important challenge at the frontier of AI since the earliest days of the field in the 1950s. Despite enormous advances in AI, especially in game-playing and computer vision , machine common sense with the richness of human common sense remains a distant possibility. This may be why AI efforts designed for complex, real-world problems with many intertwining parts, such as diagnosing and recommending treatments for COVID-19 patients, sometimes fall flat .

Modern AI is designed to tackle highly specific problems, in contrast to common sense, which is vague and can’t be defined by a set of rules. Even the latest models make absurd errors at times, suggesting that something fundamental is missing in the AI’s world model. For example, given the following text:

“You poured yourself a glass of cranberry, but then absentmindedly, you poured about a teaspoon of grape juice into it. It looks OK. You try sniffing it, but you have a bad cold, so you can’t smell anything. You are very thirsty. So you”

the highly touted AI text generator GPT-3 supplied

“drink it. You are now dead.”

Recent ambitious efforts have recognized machine common sense as a moonshot AI problem of our times, one requiring concerted collaborations across institutions over many years. A notable example is the four-year Machine Common Sense program launched in 2019 by the U.S. Defense Advanced Research Projects Agency to accelerate research in the field after the agency released a paper outlining the problem and the state of research in the field .

The Machine Common Sense program funds many current research efforts in machine common sense, including our own, Multi-modal Open World Grounded Learning and Inference ( MOWGLI ). MOWGLI is a collaboration between our research group at the University of Southern California and AI researchers from the Massachusetts Institute of Technology, University of California at Irvine, Stanford University , and Rensselaer Polytechnic Institute. The project aims to build a computer system that can answer a wide range of commonsense questions.

Transformers to the rescue?

One reason to be optimistic about finally cracking machine common sense is the recent development of a type of advanced deep learning AI called transformers. Transformers are able to model natural language in a powerful way and, with some adjustments, are able to answer simple commonsense questions. Commonsense question answering is an essential first step for building chatbots that can converse in a human-like way.

In the last couple of years, a prolific body of research has been published on transformers, with direct applications to commonsense reasoning. This rapid progress as a community has forced researchers in the field to face two related questions at the edge of science and philosophy: Just what is common sense? And how can we be sure an AI has common sense or not?

To answer the first question, researchers divide common sense into different categories, including commonsense sociology, psychology, and background knowledge. The authors of a recent book argue that researchers can go much further by dividing these categories into 48 fine-grained areas, such as planning, threat detection, and emotions.

However, it is not always clear how cleanly these areas can be separated. In our recent paper , experiments suggested that a clear answer to the first question can be problematic. Even expert human annotators – people who analyze text and categorize its components – within our group disagreed on which aspects of common sense applied to a specific sentence. The annotators agreed on relatively concrete categories like time and space but disagreed on more abstract concepts.

Recognizing AI common sense

Even if you accept that some overlap and ambiguity in theories of common sense is inevitable, can researchers ever really be sure that an AI has common sense? We often ask machines questions to evaluate their common sense, but humans navigate daily life in far more interesting ways. People employ a range of skills, honed by evolution, including the ability to recognize basic cause and effect, creative problem solving , estimations, planning , and essential social skills, such as conversation and negotiation . As long and incomplete as this list might be, an AI should achieve no less before its creators can declare victory in machine commonsense research.

It’s already becoming painfully clear that even research in transformers is yielding diminishing returns. Transformers are getting larger and more power-hungry . A recent transformer developed by Chinese search engine giant Baidu has several billion parameters. It takes an enormous amount of data to effectively train. Yet, it has so far proved unable to grasp the nuances of humancommon sense.

Even deep learning pioneers seem to think that new fundamental research may be needed before today’s neural networks are able to make such a leap. Depending on how successful this new line of research is, there’s no telling whether machine common sense is five years away, or 50.

This article by Mayank Kejriwal , Research Assistant Professor of Industrial & Systems Engineering, University of Southern California is republished from The Conversation under a Creative Commons license. Read the original article .

Try not to gag: Researchers say you’re breathing animal DNA

A team of researchers from the Queen Mary University of London and the Hamerton Zoo Park recently demonstrated they could identify and collect the DNA of dozens of different species of animal by simply extracting it from the air we breathe.

Side note: According to Guinness, the longest anyone’s held their breath on record is just shy of 25 minutes .

Up front: Yuck. But also, wow! We’ve long known that scientists could detect environmental DNA, or eDNA as it’s called. But this study puts into perspective not only how useful this could be but how robust our ability to detect and extract eDNA is in the modern age.

Per the researchers :

Background: It’s common knowledge that humans are destroying Earth. We’ve completely changed the land and seascapes of the planet while simultaneously screwing up the atmosphere.

For many species of animal, the situation is critically precarious. And we’re the only ones that can save them. We need as much information as we can possibly get, but we also need to ensure we’re not disrupting a species on the brink of extinction with our conservation efforts.

As the researchers put it:

The big idea: The team collected air samples from the area inside and surrounding a zoo site and compared the findings to databases of known species in the area.

The scientists were able to detect zoo animal eDNA – that is, DNA in the air from animals not found outside of the zoo in London – from as far away as 300 meters.

Once the air was collected, the team put it through a chemical process that allowed them to use machine learning systems to sort the data and determine exactly which species of animal were shedding cells into the local atmospheric environment.

Quick take: This has the potential to save the lives of countless animals, and maybe even bring thousands of species back from the brink of extinction. It also has the potential to be the origin story for the most terrifying surveillance system ever imagined.

It’s all fun and games until the government can snatch DNA out of the air and determine with 100% accuracy not only what it belongs to but whom.

That’s probably a long ways off, but it’s never to early to start thinking about privacy.

Leave A Comment