Events2Join

A Psychological Take on AGI Alignment


A Psychological Take on AGI Alignment - Daystar Eld

Many more types of AGI seem predictably likely to lead to ruin, and as far as I'm concerned, until this “alignment problem” is solved, it's a ...

A Psychological Take on AGI Alignment : r/slatestarcodex - Reddit

The basic challenge of AGI alignment is predicting what sort of motivations will emerge from what reward function, which nobody knows how to do ...

A "Bitter Lesson" Approach to Aligning AGI and ASI

TL;DR: I discuss the challenge of aligning AGI/ASI, and outline an extremely simple approach to aligning an LLM: train entirely on a ...

RFC: Mental phenomena in AGI alignment - LessWrong

If we suppose AGI do have minds, then alignment schemes can also use philosophical methods to address the values, goals, models, and behaviors ...

A case for AI alignment being difficult - Unstable Ontology

This is an attempt to distill a model of AGI alignment that I have ... take over the world. If the AI can distinguish between “training ...

The Alignment Problem from a Deep Learning Perspective - arXiv

Abstract:In coming years or decades, artificial general intelligence (AGI) may surpass human capabilities at many critical tasks.

AI alignment with humans... but with which humans? — EA Forum

I agree that AI alignment with actual humans & groups needs to take ... B: Figuring out how to align an AGI with a group of human. C: Doing ...

A functional contextual, observer-centric, quantum mechanical, and ...

Some researchers exploring the alignment problem highlight three aspects that AGI (or AI) requires to help resolve this problem: (1) an interpretable values ...

A "Bitter Lesson" Approach to Aligning AGI and ASI - LessWrong

My proposal on instruction-following AGI and Max Harm's corrigibility are alignment targets with a basin of attraction so we don't need perfect ...

AI alignment - Wikipedia

... (AGI) and superhuman cognitive capabilities (ASI) and could endanger human ... Commercial organizations sometimes have incentives to take shortcuts on safety and ...

AI Safety: Alignment Is Not Enough | by Rob Whiteman - Medium

We need a path forward that keeps us safe while acknowledging that artificial general intelligence (AGI) is a question of when — not if.

Robustness to fundamental uncertainty in AGI alignment - PhilArchive

... alignment schemes that can be considered since an AGI without a mental ... If we suppose AGI do have minds, then alignment schemes can also use ...

AI Alignment: The Super Wicked Problem - LifeArchitect.ai

Do not claim to take any actions in the ... The psychology of modern LLMs (2024). AI theory by others. AI papers · AGI achieved internally: A re-written story

Current AIs Provide Nearly No Data Relevant to AGI Alignment

Take an AI Optimist who'd built up a solid model of how AIs trained by SGD work. Based on that, they'd concluded that the AGI Omnicide Risk ...

Assessing the Alignment of Large Language Models With Human ...

Large language models (LLMs) hold potential for mental health applications. However, their opaque alignment processes may embed biases that shape problematic ...

Robustness to Fundamental Uncertainty in AGI Alignment.

The AGI alignment problem has a bimodal distribution of outcomes with most outcomes clustering around the poles of total success and existential, catastrophic ...

The AI Alignment Problem and AGI - YouTube

AI alignment and AGI are a growing discourse. This video is a reading of Max Tegmark's essay "The 'Don't Look Up' Thinking That Could Doom ...

AI Alignment Problem - PhilPapers

A choice is typically not finalized until we take some irreversible action in the chosen direction, like buying a ticket to the country. In the case of Task AGI ...

Comments - Why I'm optimistic about our alignment approach

But eventually we may need to align AGI that is none of these things. Is the idea that this alignment research AI will discover/design alignment ...

What is the alignment problem with AI? - YouTube

... Psychologist). Contact: info ... Ilya Sutskever (OpenAI Chief Scientist) - Building AGI, Alignment, Spies, Microsoft, & Enlightenment.