upcarta
  • Sign In
  • Sign Up
  • Explore
  • Search

David Krueger

davidscottkrueger.com
No followers
community-curated profile

Cambridge faculty - AI alignment, deep learning, and existential safety.
Formerly Mila, FHI, DeepMind, ElementAI.

Overview Posts Content Recommendations
Popular Recent
  • Article
  • Tweet
  • Paper
David Krueger @DavidSKrueger · May 24, 2023
  • From Twitter

I don't want to summarize too much, since I think everyone should read this latest piece by Yoshua Bengio (Turing award winner for Deep Learning, along with Geoff Hinton and Yann Lecun): Suffice to say that this piece engages w/AI x-risk very seriously.

Article May 23, 2023
How Rogue AIs may Arise
by Yoshua Bengio
Post Add to Collection Mark as Completed
Recommended by 1 person
1 mention
Share on Twitter Repost
David Krueger @DavidSKrueger · May 8, 2023
  • From Twitter

BTW, 2 papers I often recommend: 1) "The alignment problem from a deep learning perspective" @RichardMCNgo et al. for research overview: 2) "Natural Selection Favors AIs over Humans" @DanHendrycks for argument for risk:

Paper Feb 22, 2023
The alignment problem from a deep learning perspective
by Richard Ngo and Sören Mindermann
Post Add to Collection Mark as Completed
Recommended by 1 person
2 mentions
Share on Twitter Repost
David Krueger @DavidSKrueger · May 8, 2023
  • From Twitter

BTW, 2 papers I often recommend: 1) "The alignment problem from a deep learning perspective" @RichardMCNgo et al. for research overview: 2) "Natural Selection Favors AIs over Humans" @DanHendrycks for argument for risk:

Paper Feb 22, 2023
Natural Selection Favors AIs over Humans
by Dan Hendrycks
Post Add to Collection Mark as Completed
Recommended by 1 person
1 mention
Share on Twitter Repost
David Krueger @DavidSKrueger · Apr 18, 2023
  • From Twitter

This is a good point / thread. Note that @MIRIBerkeley and @ESYudkowsky deserve credit for (purportedly) focusing on inner alignment from the beginning. The first public documentation of this I'm aware of are Alignment Forum posts from 2017, e.g.:

Tweet Apr 18, 2023
I find it kind of bizzare that people are saying that like the real AI safety argument wasn't wheather the system could understand you it was getting it to care. Like, maybe these people have always been saying that concrete problems in AI safety w
by Nathan Barnard
Post Add to Collection Mark as Completed
Recommended by 1 person
1 mention
Share on Twitter Repost
David Krueger @DavidSKrueger · Jan 22, 2023
  • From Twitter

Great piece from @_achan96_ about why "solving alignment" is not enough to guarantee good outcomes. IME, this is a common objection from ML researchers who are concerned about risks from advanced AI, but skeptical of AI Alignment.

Article Jan 11, 2023
Alignment is not enough
by Alan Chan
Post Add to Collection Mark as Completed
Recommended by 1 person
1 mention
Share on Twitter Repost
David Krueger @DavidSKrueger · Dec 19, 2022
  • From Twitter

Recommended! This is a well-written, concise (10 pages!) summary of the Alignment Problem. It helps explain: 1) The basic concerns of (many) alignment researchers 2) Why we don't think it's crazy to talk about AI killing everyone 3) What's distinctive about the field.

Paper Feb 22, 2023
The alignment problem from a deep learning perspective
by Richard Ngo and Sören Mindermann
Post Add to Collection Mark as Completed
Recommended by 1 person
2 mentions
Share on Twitter Repost
David Krueger @DavidSKrueger · Nov 13, 2022
  • From Twitter

This thread by @jachiam0 makes a lot of good points that seem underappreciated by many in AI x-safety. But I have several significant disagreements I want to highlight:

Tweet Nov 12, 2022
🧵 to clarify my views on AGI, timelines, and x-risk. TL;DR: My AGI timelines are short-ish (within two decades with things getting weird soon) and I think x-risk is real, but I think probabilities of doom by AGI instrumentally opting to kill us are g
by Joshua Achiam
Post Add to Collection Mark as Completed
Recommended by 1 person
1 mention
Share on Twitter Repost
David Krueger @DavidSKrueger · Oct 23, 2022
  • From Twitter

Another post by @jenner_erik and Johannes Treutlein presents a great response: Reflecting, I think Katja's post mostly gives me optimism about slow take-off; a lot of my pessimism there stems from general pessimism about advanced info-tech + society.

Article Oct 19, 2022
Response to Katja Grace's AI x-risk counterarguments - LessWrong
by Erik Jenner
Post Add to Collection Mark as Completed
Recommended by 1 person
1 mention
Share on Twitter Repost
  • upcarta ©2025
  • Home
  • About
  • Terms
  • Privacy
  • Cookies
  • @upcarta