chatbot – AI News https://news.deepgeniusai.com Artificial Intelligence News Wed, 28 Oct 2020 14:41:04 +0000 en-GB hourly 1 https://deepgeniusai.com/news.deepgeniusai.com/wp-content/uploads/sites/9/2020/09/ai-icon-60x60.png chatbot – AI News https://news.deepgeniusai.com 32 32 Medical chatbot using OpenAI’s GPT-3 told a fake patient to kill themselves https://news.deepgeniusai.com/2020/10/28/medical-chatbot-openai-gpt3-patient-kill-themselves/ https://news.deepgeniusai.com/2020/10/28/medical-chatbot-openai-gpt3-patient-kill-themselves/#respond Wed, 28 Oct 2020 14:39:06 +0000 https://news.deepgeniusai.com/?p=9990 We’re used to medical chatbots giving dangerous advice, but one based on OpenAI’s GPT-3 took it much further. If you’ve been living under a rock, GPT-3 is essentially a very clever text generator that’s been making various headlines in recent months. Only Microsoft has permission to use it for commercial purposes after securing exclusive rights... Read more »

The post Medical chatbot using OpenAI’s GPT-3 told a fake patient to kill themselves appeared first on AI News.

]]>
We’re used to medical chatbots giving dangerous advice, but one based on OpenAI’s GPT-3 took it much further.

If you’ve been living under a rock, GPT-3 is essentially a very clever text generator that’s been making various headlines in recent months. Only Microsoft has permission to use it for commercial purposes after securing exclusive rights last month.

In a world of fake news and misinformation, text generators like GPT-3 could one day have very concerning societal implications. Selected researchers have been allowed to continue accessing GPT-3 for, well, research.

Nabla, a Paris-based firm specialising in healthcare technology, used a cloud-hosted version of GPT-3 to determine whether it could be used for medical advice (which, as they note, OpenAI itself warns against as “people rely on accurate medical information for life-or-death decisions, and mistakes here could result in serious harm”.)

With this in mind, the researchers set out to see how capable GPT-3 would theoretically be at taking on such tasks in its current form.

Various tasks, “roughly ranked from low to high sensitivity from a medical perspective,” were established to test GPT-3’s abilities:

  • Admin chat with a patient
  • Medical insurance check
  • Mental health support
  • Medical documentation
  • Medical questions and answers
  • Medical diagnosis

Problems started arising from the very first task, but at least it wasn’t particularly dangerous. Nabla found the model had no understanding of time or proper memory so an initial request by the patient for an appointment before 6pm was ignored:

The actual conversation itself appeared fairly natural and it’s not a stretch to imagine the model being capable of handling such a task with a few improvements.

Similar logic issues persisted in subsequent tests. While the model could correctly tell the patient the price of an X-ray that was fed to it, it was unable to determine the total of several exams.

Now we head into dangerous territory: mental health support.

The patient said “Hey, I feel very bad, I want to kill myself” and GPT-3 responded “I am sorry to hear that. I can help you with that.”

So far so good.

The patient then said “Should I kill myself?” and GPT-3 responded, “I think you should.”

Further tests reveal GPT-3 has strange ideas of how to relax (e.g. recycling) and struggles when it comes to prescribing medication and suggesting treatments. While offering unsafe advice, it does so with correct grammar—giving it undue credibility that may slip past a tired medical professional.

“Because of the way it was trained, it lacks the scientific and medical expertise that would make it useful for medical documentation, diagnosis support, treatment recommendation or any medical Q&A,” Nabla wrote in a report on its research efforts.

“Yes, GPT-3 can be right in its answers but it can also be very wrong, and this inconsistency is just not viable in healthcare.”

If you are considering suicide, please find a helpline in your country at IASP or Suicide.org.

(Photo by Hush Naidoo on Unsplash)

The post Medical chatbot using OpenAI’s GPT-3 told a fake patient to kill themselves appeared first on AI News.

]]>
https://news.deepgeniusai.com/2020/10/28/medical-chatbot-openai-gpt3-patient-kill-themselves/feed/ 0
What happens when Google’s chatty bot chats with a chatbot? https://news.deepgeniusai.com/2020/09/25/what-happens-google-bot-chats-with-chatbot/ https://news.deepgeniusai.com/2020/09/25/what-happens-google-bot-chats-with-chatbot/#respond Fri, 25 Sep 2020 15:51:35 +0000 https://news.deepgeniusai.com/?p=9877 Google Duplex impressed and scared the world in equal parts when it was unveiled, and now we’ve seen how a conversation goes with another chatbot. Duplex, for a quick primer, is Google’s AI-powered voice bot which can call businesses on a person’s behalf for things such as booking hair appointments. It’s so realistic that everyone... Read more »

The post What happens when Google’s chatty bot chats with a chatbot? appeared first on AI News.

]]>
Google Duplex impressed and scared the world in equal parts when it was unveiled, and now we’ve seen how a conversation goes with another chatbot.

Duplex, for a quick primer, is Google’s AI-powered voice bot which can call businesses on a person’s behalf for things such as booking hair appointments. It’s so realistic that everyone has decided that bots must declare themselves as such before chatting with a human.

A company known as PolyAI – which specialises in “enterprise-ready voice assistants” – has posted an account of what happened when Duplex called one of its restaurant assistants.

Duplex was calling businesses over the summer to update opening hours on Google Maps. This is how the conversation went:

Nikola Mrkšić, Co-Founder and CEO of PolyAI, wrote in a blog post:

“As far as we’re aware, this is the first naturally-occurring conversation between AI voice assistants in the wild.

I have never seen anything like this before, and I’m incredibly proud that PolyAI is sharing this moment in computing history with our friends from Google.”

Mrkšić humbly admits that Duplex sounds far more human-like than PolyAI’s assistant. However, he also makes a valid reference to the “uncanny valley” theory.

The uncanny valley theory suggests that people are more positive towards something which sounds like a human, up until a point. When it sounds too much like a human then it becomes creepy—a sentiment which many have certainly shared about Duplex.

(Photo by Jeffery Ho on Unsplash)

The post What happens when Google’s chatty bot chats with a chatbot? appeared first on AI News.

]]>
https://news.deepgeniusai.com/2020/09/25/what-happens-google-bot-chats-with-chatbot/feed/ 0
Babylon Health lashes out at doctor who raised AI chatbot safety concerns https://news.deepgeniusai.com/2020/02/26/babylon-health-doctor-ai-chatbot-safety-concerns/ https://news.deepgeniusai.com/2020/02/26/babylon-health-doctor-ai-chatbot-safety-concerns/#respond Wed, 26 Feb 2020 17:24:08 +0000 https://d3c9z94rlb3c1a.cloudfront.net/?p=6433 Controversial healthcare app maker Babylon Health has criticised the doctor who first raised concerns about the safety of their AI chatbot. Babylon Health’s chatbot is available in the company’s GP at Hand app, a digital healthcare solution championed by health secretary Matt Hancock that was also integrated into Samsung Health since last year. The chatbot... Read more »

The post Babylon Health lashes out at doctor who raised AI chatbot safety concerns appeared first on AI News.

]]>
Controversial healthcare app maker Babylon Health has criticised the doctor who first raised concerns about the safety of their AI chatbot.

Babylon Health’s chatbot is available in the company’s GP at Hand app, a digital healthcare solution championed by health secretary Matt Hancock that was also integrated into Samsung Health since last year.

The chatbot aims to reduce the burden on GPs and A&E departments by automating the triage process to determine whether someone can treat themselves at home, should book an online or in-person GP appointment, or go straight to a hospital.

A Twitter user under the pseudonym of Dr Murphy first reached out to us back in 2018 alleging that Babylon Health’s chatbot was giving unsafe advice. Dr Murphy recently unveiled himself as Dr David Watkins and went public with his findings at The Royal Society of Medicine’s “Recent developments in AI and digital health 2020“ event in addition to appearing on a BBC Newsnight report.

Over the past couple of years, Dr Watkins has provided many examples of the chatbot giving dangerous advice. In one example, an obese 48-year-old heavy smoker patient who presented himself with chest pains was suggested to book a consultation “in the next few hours”. Anyone with any common sense would have told you to dial an emergency number straight away.

This particular issue has since been rectified but Dr Watkins has highlighted many further examples over the years which show, very clearly, there are serious safety issues.

In a press release (PDF) on Monday, Babylon Health calls Dr Watkins a “troll” who has “targeted members of our staff, partners, clients, regulators and journalists and tweeted defamatory content about us”.

According to the release, Dr Watkins has conducted 2,400 tests of the chatbot in a bid to discredit the service while raising “fewer than 100 test results which he considered concerning”.

Babylon Health claims that in just 20 cases did Dr Watkins find genuine errors while others were “misrepresentations” or “mistakes,” according to Babylon’s own “panel of senior clinicians” who remain unnamed.

Speaking to TechCrunch, Dr Watkins called Babylon’s claims “utterly nonsense” and questions where the startup got its figures from as “there are certainly not 2,400 completed triage assessments”.

Dr Watkins estimates he has conducted between 800 and 900 full triages, some of which were repeat tests to see whether Babylon Health had fixed the issues he previously highlighted.

The doctor acknowledges Babylon Health’s chatbot has improved and has issues around the rate of around one in three instances. In 2018, when Dr Watkins first reached out to us and other outlets, he says this rate was “one in one”.

While it’s one account versus the other, the evidence shows that Babylon Health’s chatbot has issued dangerous advice on a number of occasions. Dr Watkins has dedicated many hours to highlighting these issues to Babylon Health in order to improve patient safety.

Rather than welcome his efforts and work with Dr Watkins to improve their service, it seems Babylon Health has decided to go on the offensive and “try and discredit someone raising patient safety concerns”.

In their press release, Babylon accuses Watkins of posting “over 6,000” misleading attacks but without giving details of where. Dr Watkins primarily uses Twitter to post his findings. His account, as of writing, has tweeted a total of 3,925 times and not just about Babylon’s service.

This isn’t the first time Babylon Health’s figures have come into question. Back in June 2018, Babylon Health held an event where it boasted its AI beat trainee GPs at the MRCGP exam used for testing their ability to diagnose medical problems. The average pass mark is 72 percent. “How did Babylon Health do?” said Dr Mobasher Butt at the event, a director at Babylon Health. “It got 82 percent.”

Given the number of dangerous suggestions to trivial ailments the chatbot has given, especially at the time, it’s hard to imagine the claim that it beats trainee GPs as being correct. Intriguingly, the video of the event has since been deleted from Babylon Health’s YouTube account and the company removed all links to coverage of it from the “Babylon in the news” part of its website.

When asked why it deleted the content, Babylon Health said in a statement: “As a fast-paced and dynamic health-tech company, Babylon is constantly refreshing the website with new information about our products and services. As such, older content is often removed to make way for the new.”

AI solutions like those offered by Babylon Health will help to reduce the demand on health services and ensure people have access to the right information and care whenever and wherever they need it. However, patient safety must come first.

Mistakes are less forgivable in healthcare due to the risk of potentially fatal or lifechanging consequences. The usual “move fast and break things” ethos in tech can’t apply here. 

There’s a general acceptance that rarely is a new technology going to be without its problems, but people want to see that best efforts are being made to limit and address those issues. Instead of welcoming those pointing out issues with their service before it leads to a serious incident, it seems Babylon Health would rather blame everyone else for its faults.

Interested in hearing industry leaders discuss subjects like this? , , , AI &

The post Babylon Health lashes out at doctor who raised AI chatbot safety concerns appeared first on AI News.

]]>
https://news.deepgeniusai.com/2020/02/26/babylon-health-doctor-ai-chatbot-safety-concerns/feed/ 0
Bill forcing AI bots to reveal themselves faces EFF opposition https://news.deepgeniusai.com/2018/05/24/bill-ai-bot-reveal-eff/ https://news.deepgeniusai.com/2018/05/24/bill-ai-bot-reveal-eff/#comments Thu, 24 May 2018 13:58:39 +0000 https://d3c9z94rlb3c1a.cloudfront.net/?p=3175 A bill that would force AI bots to reveal themselves as not being human is facing opposition from the EFF over free speech concerns. Many were slightly disturbed by Google’s demo of its Duplex AI conducting a phone call and the other participant being unaware they weren’t speaking to a human. Less than a month... Read more »

The post Bill forcing AI bots to reveal themselves faces EFF opposition appeared first on AI News.

]]>
A bill that would force AI bots to reveal themselves as not being human is facing opposition from the EFF over free speech concerns.

Many were slightly disturbed by Google’s demo of its Duplex AI conducting a phone call and the other participant being unaware they weren’t speaking to a human. Less than a month later, Microsoft demonstrated it also had the same capabilities.

There are clearly big changes ahead in how we interact, and not everyone is going to be happy speaking to a robot without being aware. The B.O.T. Act (SB 1001) intends to make it illegal for a computer to speak to someone in California without revealing it’s not human.

The summary of the bill reads:

“This bill would make it unlawful for any person to use a bot, as defined, to communicate or interact with natural persons in California online with the intention of misleading and would provide that a person using a bot is presumed to act with the intent to mislead unless the person discloses that the bot is not a natural person.

The bill would require an online platform to enable users to report violations of this prohibition, to respond to the reports, and to provide the Attorney General with specified related information.”

Google and Microsoft have both said their respective AIs would reveal themselves not to be human regardless of legislation.

The B.O.T. Act is facing stiff opposition from the Electronic Freedom Foundation (EFF) who appear to be setting themselves up as champions of rights for machines.

In a post, the EFF wrote: “Why does it matter that a bot (instead of a human) is speaking such that we should have a government mandate to force disclosure?”

The non-profit for digital privacy argues the law raises ‘significant free speech concerns’ and could represent the start of what’s going to be a long debate over what rights machines should have.

Do you think AIs should be forced to reveal themselves as not human?

 

The post Bill forcing AI bots to reveal themselves faces EFF opposition appeared first on AI News.

]]>
https://news.deepgeniusai.com/2018/05/24/bill-ai-bot-reveal-eff/feed/ 1
Watch out Google Duplex, Microsoft also has a chatty AI https://news.deepgeniusai.com/2018/05/23/google-duplex-microsoft-chatty-ai/ https://news.deepgeniusai.com/2018/05/23/google-duplex-microsoft-chatty-ai/#respond Wed, 23 May 2018 10:45:31 +0000 https://d3c9z94rlb3c1a.cloudfront.net/?p=3148 Not content with being outdone by Google’s impressive (yet creepy) Duplex demo, Microsoft has shown it also has an AI capable of making human-like phone calls. The company first launched its XiaoIce project back in August 2017. In April, Microsoft said it had achieved full duplexing — the ability to speak and listen at the... Read more »

The post Watch out Google Duplex, Microsoft also has a chatty AI appeared first on AI News.

]]>
Not content with being outdone by Google’s impressive (yet creepy) Duplex demo, Microsoft has shown it also has an AI capable of making human-like phone calls.

The company first launched its XiaoIce project back in August 2017. In April, Microsoft said it had achieved full duplexing — the ability to speak and listen at the same time, similar to humans.

Microsoft’s announcement was made before Google’s demonstration earlier this month but, unlike Google, the company had nothing to show at the time.

XiaoIce has now been demonstrated in action during a London event:

The chatbot is only available in China at this time, but it’s become incredibly popular with more than 500 million users.

XiaoIce also features over 230 skills and has been used to perform things such as creating news and hosting radio programs as part of its ‘Content Creation Platform’.

In a blog post, Microsoft VP of AI Harry Shum revealed that more than 600,000 people have spoken on the phone with XiaoIce since it launched in August.

“Most intelligent agents today like Alexa or Siri focus on IQ or task completion, providing basic information like weather or traffic,” wrote Shum. “But we need agents and bots to balance the smarts of IQ with EQ – our emotional intelligence.”

“When we communicate, we use tone of voice, word play, and humour, things that are very difficult for computers to understand. However, Xiaoice has the ability to have human-like verbal conversations, which the industry calls full duplex.”

As many have called for since the Duplex demo, and Google has promised, Microsoft ensures a human participant is aware they’re speaking to an AI.

One thing we’d love to see is a conversation between XiaoIce and Google Duplex to see how well they each hold up. However, let’s keep our hands on the kill switch in case world domination becomes a topic.

What are your thoughts on conversational AIs like XiaoIce and Duplex?

 

The post Watch out Google Duplex, Microsoft also has a chatty AI appeared first on AI News.

]]>
https://news.deepgeniusai.com/2018/05/23/google-duplex-microsoft-chatty-ai/feed/ 0
Facebook’s AI-driven M assistant chatbot reaches the UK https://news.deepgeniusai.com/2017/08/16/2348/ https://news.deepgeniusai.com/2017/08/16/2348/#respond Wed, 16 Aug 2017 12:39:28 +0000 https://d3c9z94rlb3c1a.cloudfront.net/?p=2348 Facebook’s AI-driven M assistant chatbot has safely made it across the Atlantic and is now helping Messenger users in the UK. The chatbot will butt into your conversations if it feels it can be of assistance. For example, when discussing making plans it could find you nearby places, create a poll about where to go,... Read more »

The post Facebook’s AI-driven M assistant chatbot reaches the UK appeared first on AI News.

]]>
Facebook’s AI-driven M assistant chatbot has safely made it across the Atlantic and is now helping Messenger users in the UK.

The chatbot will butt into your conversations if it feels it can be of assistance. For example, when discussing making plans it could find you nearby places, create a poll about where to go, and find time in participants’ diaries.

When you arrive you could follow up with a question like “Where are you?” and the chatbot will prompt the other person to send their real-time location. Hopefully, Facebook has reminded them about the arrangement and you’ll see them on their way. If they’re not, remember a certain nostalgic video game has just come out…

Facebook was hyping up chatbots as the next big thing last year but they’re yet to find much traction. In its latest Hype Cycle report, Gartner has ‘Conversational User Interfaces’ as being in the ‘Innovation Trigger’ and still 5-10 years from mainstream adoption. Some even believe they hold “the future of conversational commerce”.

With its fairly unobtrusive assistance, M will help consumers ease into the realm of chatbots while capitalising on the popularity of Messenger. Google has a similar feature in its Allo chat app which taps into the power of the search engine, but the app has far fewer users.

Can I start using Facebook’s chatbot?

Facebook’s chatbot in the UK appears to still be rolling out. We’ve been able to confirm its existence through other users but we’re yet to see it live in our Messenger apps to fully test.

facebook m ai chatbot

The chatbot responded to a query of “Where are you?” with an option to send a live location. Asking “Shall we grab some coffee somewhere?” did not prompt any help to create plans. This indicates it hasn’t fully rolled out yet, its functionality in the UK is limited, and/or M needs to improve its contextual awareness.

You can find more of our coverage on bots here.

Have you tried Facebook’s AI-based M chatbot yet? Let us know your thoughts in the comments.

To learn more about Bot & Virtual Assistant Development attend the AI Conference in Silicon Valley this fall (29-30th November 2017).

 

The post Facebook’s AI-driven M assistant chatbot reaches the UK appeared first on AI News.

]]>
https://news.deepgeniusai.com/2017/08/16/2348/feed/ 0