- 4
- 2
The guy https://en.wikipedia.org/wiki/Geoffrey_Hinton
Geoffrey Hinton is a British-Canadian computer scientist and cognitive psychologist, most noted for his work on artificial neural networks. From 2013 to 2023, he divided his time working for Google (Google Brain) and the University of Toronto, before publicly announcing his departure from Google in May 2023, citing concerns about the risks of artificial intelligence (AI) technology.
We'll need universal basic income - AI 'godfather'
The computer scientist regarded as the “godfather of artificial intelligence” says the government will have to establish a universal basic income to deal with the impact of AI on inequality.
Professor Geoffrey Hinton told BBC Newsnight that a benefits reform giving fixed amounts of cash to every citizen would be needed because he was “very worried about AI taking lots of mundane jobs”.
“I was consulted by people in Downing Street and I advised them that universal basic income was a good idea,” he said.
He said while he felt AI would increase productivity and wealth, the money would go to the rich “and not the people whose jobs get lost and that's going to be very bad for society”.
Professor Hinton is the pioneer of neural networks, which form the theoretical basis of the current explosion in artificial intelligence.
Until last year he worked at Google, but left the tech giant so he could talk more freely about the dangers from unregulated AI.
The concept of a universal basic income amounts to the government paying all individuals a set salary regardless of their means.
Critics say it would be extremely costly and divert funding away from public services, while not necessarily helping to alleviate poverty.
A government spokesman said there were "no plans to introduce a universal basic income".
[article continued]
Professor Hinton reiterated his concern that there were human extinction-level threats emerging.
Developments over the last year showed governments were unwilling to rein in military use of AI, he said, while the competition to develop products rapidly meant there was a risk tech companies wouldn't “put enough effort into safety”.
Professor Hinton said "my guess is in between five and 20 years from now there's a probability of half that we'll have to confront the problem of AI trying to take over".
This would lead to an “extinction-level threat” for humans because we could have “created a form of intelligence that is just better than biological intelligence… That's very worrying for us”.
AI could “evolve”, he said, “to get the motivation to make more of itself” and could autonomously “develop a sub-goal of getting control”.
He said there was already evidence of large language models - a type of AI algorithm used to generate text - choosing to be deceptive.
He said recent applications of AI to generate thousands of military targets were the “thin end of the wedge”.
“What I'm most concerned about is when these can autonomously make the decision to kill people," he said.
Professor Hinton said something similar to the Geneva Conventions - the international treaties that establish legal standards for humanitarian treatment in war - may be needed to regulate the military use of AI.
"But I don't think that's going to happen until after very nasty things have happened,” he added.
Asked if the West was in a Manhattan Project-style race - referring to nuclear weapons research during World War Two - with autocracies such as Russia and China on the military use of AI, Professor Hinton replied: “[Russian President Vladimir] Putin said some years ago that whoever controls AI controls the world. So I imagine they're working very hard.
"Fortunately, the West is probably well ahead of them in research. We're probably still slightly ahead of China. But China's putting more resources in. And so in terms of military uses I think there's going to be a race”.
He said a better solution would be a prohibition on military uses of AI.
- B64 : Gross
- 2
- 3
My favorite part is from 2:03-2:32
Favorite part from 0:59-1:27
Idk someone might appreciate it!
- WeihnachtenSalvador : B-word promised to kill her self by Christmas 2022
- 3
- 3
JUST IN: 29-year-old Dutch woman who is perfectly healthy has been granted to k*ll herself by assisted su*cide.
— Collin Rugg (@CollinRugg) May 17, 2024
Insane.
Zoraya ter Beek will soon be dead after officials in the Netherlands granted her request to take her own life.
Ter Beek has been wanting to take her own life… pic.twitter.com/gLLwVFMDOM
- 1
- 3
There are anti-communist protests happening within “Bloxburg” which is a game on Roblox.
— The Redheaded libertarian (@TRHLofficial) May 16, 2024
Let’s break it down.
So essentially, it’s your child’s responsibility to keep their character alive in this game— think Sims or a Tomagachi. These characters used to be able to get jobs…
- 2
- 4
been killing it on tiktok lately pic.twitter.com/AyaldwTTAK
— tdx (@realtdx) May 18, 2024
- 1
- 4
- 1
- 4
Answer was quite simple after saying the patient was incorrect about denials. She forgot she had recently gotten a flu shot and Covid booster. Checked her vaccination record because EMR stated she had these done concurrently, left the examination room to confirm dates and lined up. This led to acute lymphadenopathy of her right armpit. Patient was able to RTC 7 days later with full reduction and deemed just an adverse side effect of the vaccinations. Patient never returned but was seen by sister hospital alive and well for her yearly physical. That was WNR for everything.
- 2
- 5
- 4
- 5
I don't like the use of the word r-slur to describe developmentally disabled students, either. It's not their fault that their parents drank while pregnant.