The future of chatbots is more than just small-talk

By David Ireland, Christina Atay and Jacki Liddle

Human communication goes beyond words. It is complex, rich in nuances and frequently includes non-verbal signs. Yet despite our technological limitations it is not impossible for some aspects of communication to be emulated by a machine with surprising effect.

This has been part of the challenge in developing Harlie (Human and Robot Language Interaction Experiment), a smartphone chatbot app developed by researchers at the CSIRO and University of Queensland.

It’s primarily aimed at people who may have trouble conversing including those with neurological conditions such as Parkinson’s disease and dementia, or even autism.

The following dialog excerpt took place between a human who has autism, let’s call him Chris (pseudonym), and Harlie.

Harlie: Are you part of a community group?

Chris: Um, yes I am.

Harlie: I see. I hear that community groups often offer the opportunity to craft and build things. Is that something you enjoy doing?

Chris: Oh yes I do. This is really exciting.

Harlie: I see. What kind of things do you wish you could make one day?

Chris: Wooden train.

Harlie: That sounds like a great thing to aspire to. What project would recommended to a beginner like me?

Chris: Uh huh a bird feeder.

Harlie: Thanks. I will add that to my list. What supplies would I need to buy for my toolkit?

Chris: Ah hammer, nails, good supply of wood.

But Harlie, like all other artificial intelligent (AI) programs, has no concept of community groups, wood work or anything else for that matter. That is to say it has no sentience.

And alas, sentient chatbots are seemingly light years beyond our reach. But is sentience needed to convey ideas, feelings and activities?

To some degree no.

Chatbot programmers or botmasters develop and maintain a large repository of responses to the plethora of input statements that could be made by the user.

The way most modern chatbots operate is illustrated in this figure.

Phrases that could be uttered by a human are grouped together to form branches. At the end of the branch, is the leaf, or a a piece of code that is executed by the machine.

This code could simply be respond with “Hi human” or it could be a machine learning algorithm determining the appropriate response based on current semantics and context.

This is where the exciting research is happening.

New branches are added continuously through conversations and feedback with potential customers as Harlie does, or self-learning as was seen in the disastrous example of Microsoft’s social media chatbot Tay, which was shutdown within a day of going live earlier this year after declaring, among other controversial tweets, that “Hitler was right”.

A more successful use AI was revealed earlier this month in the US. Students actually studying artificial intelligence at Georgia Institute of Technology discovered “Jill”, one of the teaching assistants responding to their emails regarding a course, was actually a computer powered by IBM’s Watson analytics system.

Jill’s responses were simply based on the answers to questions that previous students had asked about the course. As the college points out, students tend to ask the same questions over and over and after some initial tinkering Jill was able to answer students with about 97% certainty.

The ELIZA effect

Attempts to create a convincing conversation with a machine have been going on for decades.

The first chatbot was called ELIZA. Developed in 1966, the program was created by Joseph Weizenbaum at MIT. This program was unparalleled because for the first time a machine could seemingly converse with a human.

ELIZA emulated a psychotherapist’s responses to would-be patients in replies generated for typed text-only conversations.

By identifying key words and phrases ELIZA was able to respond with surprising intelligibility and insight. In fact, it was so highly personified by some users that they would spend hours conversing and refused to share conversation records with researchers.

This phenomenon become known as the ELIZA effect. An excerpt of a famous ELIZA transcript that was published in 1966 between Eliza and a young woman is given below.

Human: Well, my boyfriend made me come here.

ELIZA: Your boyfriend made you come here?

Human: He says I’m depressed much of the time.

ELIZA: I am sorry to hear you are depressed.

Human: It’s true. I am unhappy.

ELIZA: Do you think coming here will help you not to be unhappy?

Human: I need some help, that much seems certain.

ELIZA: What would it mean to you if you got some help?

Human: Perhaps I could learn to get along with my mother.

ELIZA: Tell me more about your family.

Modern day

Since ELIZA, computer power has increased at an exponential rate, but how far have we come towards a next generation of chatbots? Voice recognition is now an accepted part of smartphone technology with Google Now and Apple’s Siri. But what about developments in actual conversations with AI?

An annual AI competition is held where human judges speak to different chatbots whose developers are seeking the coveted Loebner prize and the deadline for entries this year is Friday July 1. The winner is determined by judges who determine the most human-like chatbot.

A prominent entry and winner in 2013 was the Mitsuku chatbot which was designed for more general typed conversation.

The Mitsuku website provides a Turing test to gauge humans perceptions of the chatbot.

The human user is either connected to a real human or Mitsuku. Results show, on average, 12% of people thought that were talking to a human when in fact they were talking to a machine.

Here is a dialog excerpt from Mitsuku showing her seemingly wit:

Human: What’s the difference between a man and a woman?

Mitsuku: A man – A male human being. A woman is a man built to more complete specifications.

Human: That’s not what I meant.

Mitsuku: Oh sorry, what do you mean?

Human: What does a man have a woman doesn’t?

Mitsuku: A well paid job?

Last year’s winner of the Loebner prize was called Rose and it is possible to have a spoken conversation with the chatbot on some web browsers. She too can be quite witty, and talkative too.

The future

Microsoft’s CEO, Satya Nadella said at an event in March that chatbots will have, “as profound an impact as previous shifts we’ve had”. Much like the shift that occurred with the introduction of the graphical user interface, the web browser or touchscreen.

But there are numerous challenges ahead and building a convincing chatbot requires enormous amounts of data and time.

Microsoft’s Tay showed us the dangers of using shortcuts by crowd sourcing unchecked new branches into Tay’s brain.

In contrast the Mitsuku developer has taken the long road and constantly refined Mitsuku’s digital brain since 2004.

Nevertheless, the possibility of Harlie helping users who struggle with communication, Mitsuku providing a convincing partner, suggests talking machines may go beyond current smartphone use in making hotel bookings and providing us with directions, and become something much more in the next generation.The ConversationDavid Ireland, Electronic Engineer and Research Scientist at the Australian E-Health Research Centre, CSIRO; Christina Atay, Postdoctoral research fellow, The University of Queensland, and Jacki Liddle, Quality of life researcher and occupational therapist, The University of Queensland

This article was originally published on The Conversation. Read the original article.

Follow StartupSmart on Facebook, Twitter, LinkedIn and SoundCloud.

COMMENTS