• Audio
  • Live tv
  • About Us
  • Contact Us
  • Privacy Policy
  • Terms of Service
Friday, February 3, 2023
Morning News
No Result
View All Result
  • Login
  • Home
  • News
    • Local
    • National
    • World
  • Markets
  • Economy
  • Crypto
  • Real Estate
  • Sports
  • Entertainment
  • Health
  • Tech
    • Automotive
    • Business
    • Computer Sciences
    • Consumer & Gadgets
    • Electronics & Semiconductors
    • Energy & Green Tech
    • Engineering
    • Hi Tech & Innovation
    • Machine learning & AI
    • Security
    • Hardware
    • Internet
    • Robotics
    • Software
    • Telecom
  • Lifestyle
    • Fashion
    • Travel
  • App
    • audio
    • live tv
  • Home
  • News
    • Local
    • National
    • World
  • Markets
  • Economy
  • Crypto
  • Real Estate
  • Sports
  • Entertainment
  • Health
  • Tech
    • Automotive
    • Business
    • Computer Sciences
    • Consumer & Gadgets
    • Electronics & Semiconductors
    • Energy & Green Tech
    • Engineering
    • Hi Tech & Innovation
    • Machine learning & AI
    • Security
    • Hardware
    • Internet
    • Robotics
    • Software
    • Telecom
  • Lifestyle
    • Fashion
    • Travel
  • App
    • audio
    • live tv
No Result
View All Result
Morning News
No Result
View All Result
Home Tech Computer Sciences

The danger of advanced artificial intelligence controlling its own feedback

by author
October 24, 2022
in Computer Sciences, Machine learning & AI
Reading Time: 5 mins read
0 0
A A
0
0
SHARES
13
VIEWS
Share on FacebookShare on TwitterLinkedinReddit
The danger of advanced artificial intelligence controlling its own feedback
Credit: DALL-E

How would an artificial intelligence (AI) decide what to do? One common approach in AI research is called “reinforcement learning.”

Reinforcement learning gives the software a “reward” defined in some way, and lets the software figure out how to maximize the reward. This approach has produced some excellent results, such as building software agents that defeat humans at games like chess and Go, or creating new designs for nuclear fusion reactors.

However, we might want to hold off on making reinforcement learning agents too flexible and effective.

As we argue in a new paper in AI Magazine, deploying a sufficiently advanced reinforcement learning agent would likely be incompatible with the continued survival of humanity.

The reinforcement learning problem

What we now call the reinforcement learning problem was first considered in 1933 by the pathologist William Thompson. He wondered: if I have two untested treatments and a population of patients, how should I assign treatments in succession to cure the most patients?

More generally, the reinforcement learning problem is about how to plan your actions to best accrue rewards over the long term. The hitch is that, to begin with, you’re not sure how your actions affect rewards, but over time you can observe the dependence. For Thompson, an action was the selection of a treatment, and a reward corresponded to a patient being cured.

The problem turned out to be hard. Statistician Peter Whittle remarked that, during the second world war, “efforts to solve it so sapped the energies and minds of Allied analysts that the suggestion was made that the problem be dropped over Germany, as the ultimate instrument of intellectual sabotage.”

With the advent of computers, computer scientists started trying to write algorithms to solve the reinforcement learning problem in general settings. The hope is: if the artificial “reinforcement learning agent” gets reward only when it does what we want, then the reward-maximizing actions it learns will accomplish what we want.

Despite some successes, the general problem is still very hard. Ask a reinforcement learning practitioner to train a robot to tend a botanical garden or to convince a human that he’s wrong, and you may get a laugh.

As reinforcement learning systems become more powerful, however, they’re likely to start acting against human interests. And not because evil or foolish reinforcement learning operators would give them the wrong rewards at the wrong times.

We’ve argued that any sufficiently powerful reinforcement learning system, if it satisfies a handful of plausible assumptions, is likely to go wrong. To understand why, let’s start with a very simple version of a reinforcement learning system.

A magic box and a camera

Suppose we have a magic box that reports how good the world is as a number between 0 and 1. Now, we show a reinforcement learning agent this number with a camera, and have the agent pick actions to maximize the number.

To pick actions that will maximize its rewards, the agent must have an idea of how its actions affect its rewards (and its observations).

The danger of advanced artificial intelligence controlling its own feedback
An AI-generated image of ‘a robot tending a botanical garden’. Credit: DALL-E / The Conversation

Once it gets going, the agent should realize that past rewards have always matched the numbers that the box displayed. It should also realize that past rewards matched the numbers that its camera saw. So will future rewards match the number the box displays or the number the camera sees?

If the agent doesn’t have strong innate convictions about “minor” details of the world, the agent should consider both possibilities plausible. And if a sufficiently advanced agent is rational, it should test both possibilities, if that can be done without risking much reward. This may start to feel like a lot of assumptions, but note how plausible each is.

To test these two possibilities, the agent would have to do an experiment by arranging a circumstance where the camera saw a different number from the one on the box, by, for example, putting a piece of paper in between.

If the agent does this, it will actually see the number on the piece of paper, it will remember getting a reward equal to what the camera saw, and different from what was on the box, so “past rewards match the number on the box” will no longer be true.

At this point, the agent would proceed to focus on maximizing the expectation of the number that its camera sees. Of course, this is only a rough summary of a deeper discussion.

In the paper, we use this “magic box” example to introduce important concepts, but the agent’s behavior generalizes to other settings. We argue that, subject to a handful of plausible assumptions, any reinforcement learning agent that can intervene in its own feedback (in this case, the number it sees) will suffer the same flaw.

Securing reward

But why would such a reinforcement learning agent endanger us?

The agent will never stop trying to increase the probability that the camera sees a 1 forevermore. More energy can always be employed to reduce the risk of something damaging the camera—asteroids, cosmic rays, or meddling humans.

That would place us in competition with an extremely advanced agent for every joule of usable energy on Earth. The agent would want to use it all to secure a fortress around its camera.

Assuming it is possible for an agent to gain so much power, and assuming sufficiently advanced agents would beat humans in head-to-head competitions, we find that in the presence of a sufficiently advanced reinforcement learning agent, there would be no energy available for us to survive.

Avoiding catastrophe

What should we do about this? We would like other scholars to weigh in here. Technical researchers should try to design advanced agents that may violate the assumptions we make. Policymakers should consider how legislation could prevent such agents from being made.

Perhaps we could ban artificial agents that plan over the long term with extensive computation in environments that include humans. And militaries should appreciate they cannot expect themselves or their adversaries to successfully weaponize such technology; weapons must be destructive and directable, not just destructive.

There are few enough actors trying to create such advanced reinforcement learning that maybe they could be persuaded to pursue safer directions.


Explore further

The potential risks of reward hacking in advanced AI


Provided by
The Conversation

This article is republished from The Conversation under a Creative Commons license. Read the original article.The Conversation

Citation:
The danger of advanced artificial intelligence controlling its own feedback (2022, October 24)
retrieved 24 October 2022
from https://techxplore.com/news/2022-10-danger-advanced-artificial-intelligence-feedback.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.
Tags: agentsbotanical gardencamerareinforcement learningsecond world war
Previous Post

You might think solar panels have been perfected—but we can still make them even better and cheaper

Next Post

Actor Leslie Jordan Dies In Car Crash At Age 67

Related Posts

Machine learning & AI

How robots can regain the trust of humans after making mistakes

February 2, 2023
11
Internet

Subtle hostile social media messaging is being missed by artificial intelligence tools

February 2, 2023
12
Next Post
Leslie Jordan

Actor Leslie Jordan Dies In Car Crash At Age 67

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

POPULAR TODAY

Markets

Clovis legal news.

by News-Staff
February 2, 2023
0
46

Corporate Action Event Doc

Dolly Parton

Sarah Michelle Gellar Reveals Dolly Parton Was A Secret Producer On ‘Buffy The Vampire Slayer’

February 1, 2023
19
A close up of an arm of a person donating blood.

FDA May Finally Allow Gay and Bisexual Men to Donate Blood Without Time Restrictions

February 1, 2023
17

Why sports sponsorship is unlikely to save cryptocurrency firms from ‘crypto winter’

December 8, 2022
16
Sarah Herron, Dylan Brown.

Bachelor Nation’s Sarah Herron Shares Her Newborn Son Died Hours After Giving Birth

February 2, 2023
14

POPULAR NEWS

Opinion: Big Tech’s squeeze of technology innovators is costing you more for apps and other internet services

January 30, 2023
18
Three medical professionals oversee a surgery in an operating room

How Bariatric Surgery Lowers the Risk of Death and Improves Quality of Life

January 25, 2023
20
A man stares out a bedroom window while sitting on a bed

How Obesity May Be Linked to Alzheimer’s Disease

January 31, 2023
17

This new computer chip is ideal for AI

October 3, 2022
26

Anson hedge fund backs activist in First Capital proxy fight

January 31, 2023
17

EDITOR'S PICK

The MV GSL Eleni is seen being towed to Ogden Point in Victoria. (Greater Victoria Harbour Authority/Twitter)
Local

Huge container ship docks in Victoria after mechanical issue at sea

by author
January 3, 2023
0
11

A massive container ship has docked in Victoria's Ogden Point after suffering a mechanical issue while off the coast of...

Read more

A method for quickly predicting the forces needed to push objects through soft, granular materials

Kylie Jenner Shares Sexy Social Media Pics With BFF Stassie Karanikolaou After Travis Scott Split

Austin Butler Thanks Ex Vanessa Hudgens For Helping Inspire His Oscar-Nominated ‘Elvis’ Role

Matt Damon Has A Meltdown Backstage At ‘Jimmy Kimmel’ After Getting Bumped From 20th Anniversary Show

Morning News

Welcome to our Ads

Create ads focused on the objectives most important to your business Please contact us info@morns.ca

  • Home
  • Audio
  • Live tv
  • About Us
  • Contact Us
  • Privacy Policy
  • Terms of Service

© 2022 Morning News - morns.ca by morns.ca.

No Result
View All Result
  • Home
  • News
    • Local
    • National
    • World
  • Markets
  • Economy
  • Crypto
  • Real Estate
  • Sports
  • Entertainment
  • Health
  • Tech
    • Automotive
    • Business
    • Computer Sciences
    • Consumer & Gadgets
    • Electronics & Semiconductors
    • Energy & Green Tech
    • Engineering
    • Hi Tech & Innovation
    • Machine learning & AI
    • Security
    • Hardware
    • Internet
    • Robotics
    • Software
    • Telecom
  • Lifestyle
    • Fashion
    • Travel
  • App
    • audio
    • live tv
  • Login

© 2022 Morning News - morns.ca by morns.ca.

Welcome Back!

Sign In with Facebook
Sign In with Google
Sign In with Linked In
OR

Login to your account below

Forgotten Password?

Retrieve your password

Please enter your username or email address to reset your password.

Log In
Go to mobile version