Jump to content
  • US Air Force denies AI drone attacked operator in test

    aum

    • 1 comment
    • 437 views
    • 3 minutes
     Share


    • 1 comment
    • 437 views
    • 3 minutes

    A US Air Force colonel "mis-spoke" when describing an experiment in which an AI-enabled drone opted to attack its operator in order to complete its mission, the service has said.

     

    Colonel Tucker Hamilton, chief of AI test and operations in the US Air Force, was speaking at a conference organised by the Royal Aeronautical Society.

     

    A report about it went viral.

     

    The Air Force says no such experiment took place.

     

    In his talk, he had described a simulation in which an AI-enabled drone was repeatedly stopped from completing its task of destroying Surface-to-Air Missile sites by its human operator.

     

    He said that in the end, despite having been trained not to kill the operator, the drone destroyed the communication tower so that the operator could no longer communicate with it.

     

    "We've never run that experiment, nor would we need to in order to realise that this is a plausible outcome," Col Hamilton later clarified in a statement to the Royal Aeronautical Society.

     

    He added that it was a "thought experiment" rather than anything which had actually taken place.

     

    There have been a number of warnings about the threat to humanity posed by AI issued recently by people working in the sector, although not all experts agree how serious a risk it is.

     

    Speaking to the BBC earlier this week, Prof Yoshua Bengio, one of three computer scientists described as the "godfathers" of AI after winning a prestigious Turing Award for their work, said he thought the military should not be allowed to have AI powers at all.

     

    He described it as "one of the worst places where we could put a super-intelligent AI".


    A pre-planned scenario?

     

    I spent several hours this morning speaking to experts in both defence and AI, all of whom were very sceptical about Col Hamilton's claims, which were being widely reported before his clarification.

     

    One defence expert told me Col Hamilton's original story seemed to be missing "important context", if nothing else.

     

    There were also suggestions on social media that had such an experiment taken place, it was more likely to have been a pre-planned scenario rather than the AI-enabled drone being powered by machine learning during the task - which basically means it would not have been choosing its own outcomes as it went along, based on what had happened previously.

     

    Steve Wright, professor of aerospace engineering at the University of the West of England, and an expert in unmanned aerial vehicles, told me jokingly that he had "always been a fan of the Terminator films" when I asked him for his thoughts about the story.

     

    "In aircraft control computers there are two things to worry about: 'do the right thing' and 'don't do the wrong thing', so this is a classic example of the second," he said.

     

    "In reality we address this by always including a second computer that has been programmed using old-style techniques, and this can pull the plug as soon as the first one does something strange."

     

    Source


    User Feedback

    Recommended Comments



    Join the conversation

    You can post now and register later. If you have an account, sign in now to post with your account.
    Note: Your post will require moderator approval before it will be visible.

    Guest
    Add a comment...

    ×   Pasted as rich text.   Paste as plain text instead

      Only 75 emoji are allowed.

    ×   Your link has been automatically embedded.   Display as a link instead

    ×   Your previous content has been restored.   Clear editor

    ×   You cannot paste images directly. Upload or insert images from URL.


  • Recently Browsing   0 members

    • No registered users viewing this page.
×
×
  • Create New...