The Nonlinear Library: LessWrong

By The Nonlinear Fund

Listen to a podcast, please open Podcast Republic app. Available on Google Play Store and Apple App Store.


Category: Education

Open in Apple Podcasts


Open RSS feed


Open Website


Rate for this podcast

Subscribers: 2
Reviews: 0
Episodes: 1001

Description

The Nonlinear Library allows you to easily listen to top EA and rationalist content on your podcast player. We use text-to-speech software to create an automatically updating repository of audio content from the EA Forum, Alignment Forum, LessWrong, and other EA blogs. To find out more, please visit us at nonlinear.org

Episode Date
LW - Prompts for Big-Picture Planning by Raemon
Apr 14, 2024
LW - What convincing warning shot could help prevent extinction from AI? by Charbel-Raphaël
Apr 13, 2024
LW - Things Solenoid Narrates by Solenoid Entity
Apr 13, 2024
LW - Carl Sagan, nuking the moon, and not nuking the moon by eukaryote
Apr 13, 2024
LW - MIRI's April 2024 Newsletter by Harlan
Apr 13, 2024
LW - UDT1.01: Plannable and Unplanned Observations (3/10) by Diffractor
Apr 12, 2024
LW - Generalized Stat Mech: The Boltzmann Approach by David Lorell
Apr 12, 2024
LW - A DandD.Sci Dodecalogue by abstractapplic
Apr 12, 2024
LW - Announcing Atlas Computing by miyazono
Apr 12, 2024
LW - DandD.Sci: The Mad Tyrant's Pet Turtles [Evaluation and Ruleset] by abstractapplic
Apr 10, 2024
LW - RTFB: On the New Proposed CAIP AI Bill by Zvi
Apr 10, 2024
LW - Ophiology (or, how the Mamba architecture works) by Danielle Ensign
Apr 09, 2024
LW - Conflict in Posthuman Literature by Martín Soto
Apr 09, 2024
LW - Medical Roundup #2 by Zvi
Apr 09, 2024
LW - Math-to-English Cheat Sheet by nahoj
Apr 09, 2024
LW - Gated Attention Blocks: Preliminary Progress toward Removing Attention Head Superposition by cmathw
Apr 09, 2024
LW - on the dollar-yen exchange rate by bhauth
Apr 08, 2024
LW - How We Picture Bayesian Agents by johnswentworth
Apr 08, 2024
LW - A Dozen Ways to Get More Dakka by Davidmanheim
Apr 08, 2024
LW - My intellectual journey to (dis)solve the hard problem of consciousness by Charbel-Raphaël
Apr 07, 2024
LW - "Fractal Strategy" workshop report by Raemon
Apr 07, 2024
LW - The 2nd Demographic Transition by Maxwell Tabarrok
Apr 07, 2024
LW - On Complexity Science by Garrett Baker
Apr 05, 2024
LW - Partial value takeover without world takeover by KatjaGrace
Apr 05, 2024
LW - New report: A review of the empirical evidence for existential risk from AI via misaligned power-seeking by Harlan
Apr 05, 2024
LW - AI #58: Stargate AGI by Zvi
Apr 05, 2024
LW - LLMs for Alignment Research: a safety priority? by abramdemski
Apr 04, 2024
LW - A gentle introduction to mechanistic anomaly detection by Erik Jenner
Apr 04, 2024
LW - What's with all the bans recently? by Gerald Monroe
Apr 04, 2024
LW - Best in Class Life Improvement by sapphire
Apr 04, 2024
LW - Notes on Dwarkesh Patel's Podcast with Sholto Douglas and Trenton Bricken by Zvi
Apr 02, 2024
LW - LessWrong: After Dark, a new side of LessWrong by So8res
Apr 02, 2024
LW - Gradient Descent on the Human Brain by Jozdien
Apr 02, 2024
LW - Coherence of Caches and Agents by johnswentworth
Apr 02, 2024
LW - Announcing Suffering For Good by Garrett Baker
Apr 01, 2024
LW - OMMC Announces RIP by Adam Scholl
Apr 01, 2024
LW - The Evolution of Humans Was Net-Negative for Human Values by Zack M Davis
Apr 01, 2024
LW - So You Created a Sociopath - New Book Announcement! by Garrett Baker
Apr 01, 2024
LW - A Selection of Randomly Selected SAE Features by CallumMcDougall
Apr 01, 2024
LW - Apply to be a Safety Engineer at Lockheed Martin! by yanni
Apr 01, 2024
LW - The Story of "I Have Been A Good Bing" by habryka
Apr 01, 2024
LW - The Best Tacit Knowledge Videos on Every Subject by Parker Conley
Mar 31, 2024
LW - SAE-VIS: Announcement Post by CallumMcDougall
Mar 31, 2024
LW - My simple AGI investment and insurance strategy by lc
Mar 31, 2024
LW - Back to Basics: Truth is Unitary by lsusr
Mar 30, 2024
LW - DandD.Sci: The Mad Tyrant's Pet Turtles by abstractapplic
Mar 30, 2024
LW - SAE reconstruction errors are (empirically) pathological by wesg
Mar 29, 2024
LW - How to safely use an optimizer by Simon Fischer
Mar 29, 2024
LW - [Linkpost] Practically-A-Book Review: Rootclaim $100,000 Lab Leak Debate by trevor
Mar 28, 2024
LW - Was Releasing Claude-3 Net-Negative? by Logan Riggs
Mar 28, 2024
LW - The Cognitive-Theoretic Model of the Universe: A Partial Summary and Review by jessicata
Mar 27, 2024
LW - Daniel Kahneman has died by DanielFilan
Mar 27, 2024
LW - AE Studio @ SXSW: We need more AI consciousness research (and further resources) by AE Studio
Mar 27, 2024
LW - Failures in Kindness by silentbob
Mar 27, 2024
LW - Modern Transformers are AGI, and Human-Level by abramdemski
Mar 26, 2024
LW - My Interview With Cade Metz on His Reporting About Slate Star Codex by Zack M Davis
Mar 26, 2024
LW - Should rationalists be spiritual / Spirituality as overcoming delusion by Kaj Sotala
Mar 26, 2024
LW - LessOnline (May 31 - June 2, Berkeley, CA) by Ben Pace
Mar 26, 2024
LW - On attunement by Joe Carlsmith
Mar 25, 2024
LW - My Detailed Notes and Commentary from Secular Solstice by Jeffrey Heninger
Mar 25, 2024
LW - On Lex Fridman's Second Podcast with Altman by Zvi
Mar 25, 2024
LW - Do not delete your misaligned AGI. by mako yass
Mar 25, 2024
LW - All About Concave and Convex Agents by mako yass
Mar 25, 2024
LW - Vipassana Meditation and Active Inference: A Framework for Understanding Suffering and its Cessation by Benjamin Sturgeon
Mar 24, 2024
LW - General Thoughts on Secular Solstice by Jeffrey Heninger
Mar 24, 2024
LW - A Teacher vs. Everyone Else by ronak69
Mar 23, 2024
LW - AI #56: Blackwell That Ends Well by Zvi
Mar 23, 2024
LW - ChatGPT can learn indirect control by Raymond D
Mar 22, 2024
LW - Vernor Vinge, who coined the term "Technological Singularity", dies at 79 by Kaj Sotala
Mar 21, 2024
LW - On green by Joe Carlsmith
Mar 21, 2024
LW - "Deep Learning" Is Function Approximation by Zack M Davis
Mar 21, 2024
LW - DeepMind: Evaluating Frontier Models for Dangerous Capabilities by Zach Stein-Perlman
Mar 21, 2024
LW - On the Gladstone Report by Zvi
Mar 21, 2024
LW - Stagewise Development in Neural Networks by Jesse Hoogland
Mar 20, 2024
LW - Monthly Roundup #16: March 2024 by Zvi
Mar 20, 2024
LW - Natural Latents: The Concepts by johnswentworth
Mar 20, 2024
LW - New report: Safety Cases for AI by joshc
Mar 20, 2024
LW - Increasing IQ by 10 Points is Possible by George3d6
Mar 19, 2024
LW - Inferring the model dimension of API-protected LLMs by Ege Erdil
Mar 19, 2024
LW - Experimentation (Part 7 of "The Sense Of Physical Necessity") by LoganStrohl
Mar 19, 2024
LW - Neuroscience and Alignment by Garrett Baker
Mar 19, 2024
LW - Toki pona FAQ by dkl9
Mar 19, 2024
LW - 5 Physics Problems by DaemonicSigil
Mar 18, 2024
LW - Measuring Coherence of Policies in Toy Environments by dx26
Mar 18, 2024
LW - Community Notes by X by NicholasKees
Mar 18, 2024
LW - On Devin by Zvi
Mar 18, 2024
LW - The Worst Form Of Government (Except For Everything Else We've Tried) by johnswentworth
Mar 17, 2024
LW - Anxiety vs. Depression by Sable
Mar 17, 2024
LW - My PhD thesis: Algorithmic Bayesian Epistemology by Eric Neyman
Mar 17, 2024
LW - Rational Animations offers animation production and writing services! by Writer
Mar 16, 2024
LW - Introducing METR's Autonomy Evaluation Resources by Megan Kinniment
Mar 16, 2024
LW - Constructive Cauchy sequences vs. Dedekind cuts by jessicata
Mar 15, 2024
LW - Conditional on Getting to Trade, Your Trade Wasn't All That Great by Ricki Heicklen
Mar 14, 2024
LW - Highlights from Lex Fridman's interview of Yann LeCun by Joel Burget
Mar 14, 2024
LW - How useful is "AI Control" as a framing on AI X-Risk? by habryka
Mar 14, 2024
LW - AI #55: Keep Clauding Along by Zvi
Mar 14, 2024
LW - Jobs, Relationships, and Other Cults by Ruby
Mar 14, 2024
LW - 'Empiricism!' as Anti-Epistemology by Eliezer Yudkowsky
Mar 14, 2024
LW - On the Latest TikTok Bill by Zvi
Mar 13, 2024
LW - What could a policy banning AGI look like? by TsviBT
Mar 13, 2024
LW - The Parable Of The Fallen Pendulum - Part 2 by johnswentworth
Mar 13, 2024
LW - Superforecasting the Origins of the Covid-19 Pandemic by DanielFilan
Mar 13, 2024
LW - OpenAI: The Board Expands by Zvi
Mar 12, 2024
LW - Be More Katja by Nathan Young
Mar 12, 2024
LW - "Artificial General Intelligence": an extremely brief FAQ by Steven Byrnes
Mar 11, 2024
LW - Some (problematic) aesthetics of what constitutes good work in academia by Steven Byrnes
Mar 11, 2024
LW - What do we know about the AI knowledge and views, especially about existential risk, of the new OpenAI board members? by Zvi
Mar 11, 2024
LW - Twelve Lawsuits against OpenAI by Remmelt
Mar 11, 2024
LW - "How could I have thought that faster?" by mesaoptimizer
Mar 11, 2024
LW - Simple versus Short: Higher-order degeneracy and error-correction by Daniel Murfet
Mar 11, 2024
LW - One-shot strategy games? by Raemon
Mar 11, 2024
LW - Notes from a Prompt Factory by Richard Ngo
Mar 10, 2024
LW - Closeness To the Issue (Part 5 of "The Sense Of Physical Necessity") by LoganStrohl
Mar 09, 2024
LW - Lies and disrespect from the EA Infrastructure Fund by Igor Ivanov
Mar 08, 2024
LW - Woods' new preprint on object permanence by Steven Byrnes
Mar 08, 2024
LW - AI #54: Clauding Along by Zvi
Mar 08, 2024
LW - MATS AI Safety Strategy Curriculum by Ryan Kidd
Mar 08, 2024
LW - Simple Kelly betting in prediction markets by jessicata
Mar 07, 2024
LW - Mud and Despair (Part 4 of "The Sense Of Physical Necessity") by LoganStrohl
Mar 07, 2024
LW - Social status part 1/2: negotiations over object-level preferences by Steven Byrnes
Mar 07, 2024
LW - Movie posters by KatjaGrace
Mar 07, 2024
LW - On Claude 3.0 by Zvi
Mar 06, 2024
LW - Vote on Anthropic Topics to Discuss by Ben Pace
Mar 06, 2024
LW - Using axis lines for good or evil by dynomight
Mar 06, 2024
LW - My Clients, The Liars by ymeskhout
Mar 05, 2024
LW - Research Report: Sparse Autoencoders find only 9/180 board state features in OthelloGPT by Robert AIZI
Mar 05, 2024
LW - Read the Roon by Zvi
Mar 05, 2024
LW - Claude 3 claims it's conscious, doesn't want to die or be modified by Mikhail Samin
Mar 05, 2024
LW - Housing Roundup #7 by Zvi
Mar 05, 2024
LW - The Broken Screwdriver and other parables by bhauth
Mar 04, 2024
LW - Are we so good to simulate? by KatjaGrace
Mar 04, 2024
LW - Self-Resolving Prediction Markets by PeterMcCluskey
Mar 04, 2024
LW - Grief is a fire sale by Nathan Young
Mar 04, 2024
LW - Agreeing With Stalin in Ways That Exhibit Generally Rationalist Principles by Zack M Davis
Mar 03, 2024
LW - The Defence production act and AI policy by NathanBarnard
Mar 02, 2024
LW - If you weren't such an idiot... by kave
Mar 02, 2024
LW - The World in 2029 by Nathan Young
Mar 02, 2024
LW - Increasing IQ is trivial by George3d6
Mar 02, 2024
LW - Notes on Dwarkesh Patel's Podcast with Demis Hassabis by Zvi
Mar 02, 2024
LW - Elon files grave charges against OpenAI by mako yass
Mar 01, 2024
LW - Locating My Eyes (Part 3 of "The Sense of Physical Necessity") by LoganStrohl
Mar 01, 2024
LW - The Parable Of The Fallen Pendulum - Part 1 by johnswentworth
Mar 01, 2024
LW - AI #53: One More Leap by Zvi
Feb 29, 2024
LW - Bengio's Alignment Proposal: "Towards a Cautious Scientist AI with Convergent Safety Bounds" by mattmacdermott
Feb 29, 2024
LW - Timaeus's First Four Months by Jesse Hoogland
Feb 28, 2024
LW - How I internalized my achievements to better deal with negative feelings by Raymond Koopmanschap
Feb 28, 2024
LW - The Gemini Incident Continues by Zvi
Feb 28, 2024
LW - Announcing 'The LeastWrong' and review winner post pages by kave
Feb 28, 2024
LW - Examining Language Model Performance with Reconstructed Activations using Sparse Autoencoders by Evan Anders
Feb 27, 2024
LW - How I build and run behavioral interviews by benkuhn
Feb 27, 2024
LW - Can an AI do our alignment homework for us? by Chris Leong
Feb 26, 2024
LW - Ideological Bayesians by Kevin Dorst
Feb 26, 2024
LW - China-AI forecasts by NathanBarnard
Feb 26, 2024
LW - "In-Context" "Learning" by Arjun Panickssery
Feb 25, 2024
LW - A starting point for making sense of task structure (in machine learning) by Kaarel
Feb 25, 2024
LW - We Need Major, But Not Radical, FDA Reform by Maxwell Tabarrok
Feb 25, 2024
LW - Deep and obvious points in the gap between your thoughts and your pictures of thought by KatjaGrace
Feb 25, 2024
LW - How well do truth probes generalise? by mishajw
Feb 25, 2024
LW - Choosing My Quest (Part 2 of "The Sense Of Physical Necessity") by LoganStrohl
Feb 24, 2024
LW - Rationality Research Report: Towards 10x OODA Looping? by Raemon
Feb 24, 2024
LW - Balancing Games by jefftk
Feb 24, 2024
LW - The Sense Of Physical Necessity: A Naturalism Demo (Introduction) by LoganStrohl
Feb 24, 2024
LW - The Shutdown Problem: Incomplete Preferences as a Solution by EJT
Feb 23, 2024
LW - The Byronic Hero Always Loses by Cole Wyeth
Feb 23, 2024
LW - Everything Wrong with Roko's Claims about an Engineered Pandemic by EZ97
Feb 23, 2024
LW - Gemini Has a Problem by Zvi
Feb 23, 2024
LW - AI #52: Oops by Zvi
Feb 23, 2024
LW - Research Post: Tasks That Language Models Don't Learn by Bruce W. Lee
Feb 23, 2024
LW - Sora What by Zvi
Feb 23, 2024
LW - Contra Ngo et al. "Every 'Every Bay Area House Party' Bay Area House Party" by Ricki Heicklen
Feb 23, 2024
LW - Do sparse autoencoders find "true features"? by Demian Till
Feb 22, 2024
LW - Job Listing: Managing Editor / Writer by Gretta Duleba
Feb 22, 2024
LW - The One and a Half Gemini by Zvi
Feb 22, 2024
LW - The Pareto Best and the Curse of Doom by Screwtape
Feb 22, 2024
LW - Dual Wielding Kindle Scribes by mesaoptimizer
Feb 21, 2024
LW - Why does generalization work? by Martín Soto
Feb 21, 2024
LW - Less Wrong automated systems are inadvertently Censoring me by Roko
Feb 21, 2024
LW - AI #51: Altman's Ambition by Zvi
Feb 21, 2024
LW - I'd also take $7 trillion by bhauth
Feb 19, 2024
LW - Retirement Accounts and Short Timelines by jefftk
Feb 19, 2024
LW - On coincidences and Bayesian reasoning, as applied to the origins of COVID-19 by viking math
Feb 19, 2024
LW - Things I've Grieved by Raemon
Feb 18, 2024
LW - Self-Awareness: Taxonomy and eval suite proposal by Daniel Kokotajlo
Feb 18, 2024
LW - The Pointer Resolution Problem by Jozdien
Feb 17, 2024
LW - 2023 Survey Results by Screwtape
Feb 16, 2024
LW - Fixing Feature Suppression in SAEs by Benjamin Wright
Feb 16, 2024
LW - OpenAI's Sora is an agent by CBiddulph
Feb 16, 2024
LW - Every "Every Bay Area House Party" Bay Area House Party by Richard Ngo
Feb 16, 2024
LW - Raising children on the eve of AI by juliawise
Feb 15, 2024
LW - On the Proposed California SB 1047 by Zvi
Feb 14, 2024
LW - CFAR Takeaways: Andrew Critch by Raemon
Feb 14, 2024
LW - Masterpiece by Richard Ngo
Feb 14, 2024
LW - Where is the Town Square? by Gretta Duleba
Feb 13, 2024
LW - Tort Law Can Play an Important Role in Mitigating AI Risk by Gabriel Weil
Feb 13, 2024
LW - Lsusr's Rationality Dojo by lsusr
Feb 13, 2024
LW - AI #50: The Most Dangerous Thing by Zvi
Feb 08, 2024
LW - A Chess-GPT Linear Emergent World Representation by karvonenadam
Feb 08, 2024
LW - Believing In by AnnaSalamon
Feb 08, 2024
LW - Conditional prediction markets are evidential, not causal by philh
Feb 08, 2024
LW - More Hyphenation by Arjun Panickssery
Feb 07, 2024
LW - story-based decision-making by bhauth
Feb 07, 2024
LW - Why I think it's net harmful to do technical safety research at AGI labs by Remmelt
Feb 07, 2024
LW - what does davidad want from "boundaries"? by Chipmonk
Feb 07, 2024
LW - My guess at Conjecture's vision: triggering a narrative bifurcation by Alexandre Variengien
Feb 06, 2024
LW - Fluent dreaming for language models (AI interpretability method) by tbenthompson
Feb 06, 2024
LW - Preventing model exfiltration with upload limits by ryan greenblatt
Feb 06, 2024
LW - Things You're Allowed to Do: University Edition by Saul Munn
Feb 06, 2024
LW - Implementing activation steering by Annah
Feb 05, 2024
LW - Noticing Panic by Cole Wyeth
Feb 05, 2024
LW - On Dwarkesh's 3rd Podcast With Tyler Cowen by Zvi
Feb 04, 2024
LW - Theories of Applied Rationality by Camille Berger
Feb 04, 2024
LW - Why I no longer identify as transhumanist by Kaj Sotala
Feb 04, 2024
LW - Brute Force Manufactured Consensus is Hiding the Crime of the Century by Roko
Feb 03, 2024
LW - Announcing the London Initiative for Safe AI (LISA) by James Fox
Feb 03, 2024
LW - Survey for alignment researchers: help us build better field-level models by Cameron Berg
Feb 03, 2024
LW - Soft Prompts for Evaluation: Measuring Conditional Distance of Capabilities by porby
Feb 03, 2024
LW - Most experts believe COVID-19 was probably not a lab leak by DanielFilan
Feb 02, 2024
LW - On Not Requiring Vaccination by jefftk
Feb 02, 2024
LW - Open Source Sparse Autoencoders for all Residual Stream Layers of GPT2-Small by Joseph Bloom
Feb 02, 2024
LW - Wrong answer bias by lukehmiles
Feb 02, 2024
LW - Davidad's Provably Safe AI Architecture - ARIA's Programme Thesis by simeon c
Feb 02, 2024
LW - Ten Modes of Culture War Discourse by jchan
Feb 01, 2024
LW - Simple distribution approximation: When sampled 100 times, can language models yield 80% A and 20% B? by Teun van der Weij
Feb 01, 2024
LW - Per protocol analysis as medical malpractice by braces
Feb 01, 2024
LW - Leading The Parade by johnswentworth
Jan 31, 2024
LW - Without Fundamental Advances, Rebellion and Coup d'État are the Inevitable Outcomes of Dictators and Monarchs Trying to Control Large, Capable Countries by Roko
Jan 31, 2024
LW - Explaining Impact Markets by Saul Munn
Jan 31, 2024
LW - Childhood and Education Roundup #4 by Zvi
Jan 31, 2024
LW - on neodymium magnets by bhauth
Jan 31, 2024
LW - Win Friends and Influence People Ch. 2: The Bombshell by gull
Jan 30, 2024
LW - Things You're Allowed to Do: At the Dentist by rbinnn
Jan 30, 2024
LW - Processor clock speeds are not how fast AIs think by Ege Erdil
Jan 29, 2024
LW - Why I take short timelines seriously by NicholasKees
Jan 29, 2024
LW - Palworld development blog post by bhauth
Jan 28, 2024
LW - Epistemic Hell by rogersbacon
Jan 28, 2024
LW - Don't sleep on Coordination Takeoffs by trevor
Jan 28, 2024
LW - Aligned AI is dual use technology by lc
Jan 27, 2024
LW - The Good Balsamic Vinegar by jenn
Jan 27, 2024
LW - Surgery Works Well Without The FDA by Maxwell Tabarrok
Jan 27, 2024
LW - Making every researcher seek grants is a broken model by jasoncrawford
Jan 26, 2024
LW - "Does your paradigm beget new, good, paradigms?" by Raemon
Jan 26, 2024
LW - AI #48: The Talk of Davos by Zvi
Jan 26, 2024
LW - Is a random box of gas predictable after 20 seconds? by Thomas Kwa
Jan 26, 2024
LW - [Repost] The Copenhagen Interpretation of Ethics by mesaoptimizer
Jan 26, 2024
LW - RAND report finds no effect of current LLMs on viability of bioterrorism attacks by StellaAthena
Jan 25, 2024
LW - Will quantum randomness affect the 2028 election? by Thomas Kwa
Jan 25, 2024
LW - Humans aren't fleeb. by Charlie Steiner
Jan 24, 2024
LW - This might be the last AI Safety Camp by Remmelt
Jan 24, 2024
LW - the subreddit size threshold by bhauth
Jan 24, 2024
LW - Making a Secular Solstice Songbook by jefftk
Jan 24, 2024
LW - Loneliness and suicide mitigation for students using GPT3-enabled chatbots (survey of Replika users in Nature) by Kaj Sotala
Jan 24, 2024
LW - legged robot scaling laws by bhauth
Jan 22, 2024
LW - On "Geeks, MOPs, and Sociopaths" by alkjash
Jan 22, 2024
LW - Book review: Cuisine and Empire by eukaryote
Jan 22, 2024
LW - When Does Altruism Strengthen Altruism? by jefftk
Jan 22, 2024
LW - A quick investigation of AI pro-AI bias by Fabien Roger
Jan 20, 2024
LW - What rationality failure modes are there? by Ulisse Mini
Jan 19, 2024
LW - There is way too much serendipity by Malmesbury
Jan 19, 2024
LW - Logical Line-Of-Sight Makes Games Sequential or Loopy by StrivingForLegibility
Jan 19, 2024
LW - Does literacy remove your ability to be a bard as good as Homer? by Adrià Garriga-alonso
Jan 19, 2024
LW - The True Story of How GPT-2 Became Maximally Lewd by Writer
Jan 19, 2024
LW - On the abolition of man by Joe Carlsmith
Jan 18, 2024
LW - On Anthropic's Sleeper Agents Paper by Zvi
Jan 17, 2024
LW - AlphaGeometry: An Olympiad-level AI system for geometry by alyssavance
Jan 17, 2024
LW - An Introduction To The Mandelbrot Set That Doesn't Mention Complex Numbers by Yitz
Jan 17, 2024
LW - Medical Roundup #1 by Zvi
Jan 17, 2024
LW - Why wasn't preservation with the goal of potential future revival started earlier in history? by Andy McKenzie
Jan 17, 2024
LW - Being nicer than Clippy by Joe Carlsmith
Jan 17, 2024
LW - The impossible problem of due process by mingyuan
Jan 16, 2024
LW - The case for training frontier AIs on Sumerian-only corpus by Alexandre Variengien
Jan 15, 2024
LW - DandD.Sci(-fi): Colonizing the SuperHyperSphere by abstractapplic
Jan 15, 2024
LW - Gender Exploration by sapphire
Jan 14, 2024
LW - Notice When People Are Directionally Correct by Chris Leong
Jan 14, 2024
LW - Against most AI risk analogies by Matthew Barnett
Jan 14, 2024
LW - Land Reclamation is in the 9th Circle of Stagnation Hell by Maxwell Tabarrok
Jan 13, 2024
LW - What good is G-factor if you're dumped in the woods? A field report from a camp counselor. by Hastings
Jan 12, 2024
LW - An Actually Intuitive Explanation of the Oberth Effect by Isaac King
Jan 12, 2024
LW - Introduce a Speed Maximum by jefftk
Jan 12, 2024
LW - An even deeper atheism by Joe Carlsmith
Jan 11, 2024
LW - The Perceptron Controversy by Yuxi Liu
Jan 11, 2024
LW - Universal Love Integration Test: Hitler by Raemon
Jan 11, 2024
LW - The Aspiring Rationalist Congregation by maia
Jan 11, 2024
LW - Does AI risk "other" the AIs? by Joe Carlsmith
Jan 10, 2024
LW - Saving the world sucks by Defective Altruism
Jan 10, 2024
LW - On the Contrary, Steelmanning Is Normal; ITT-Passing Is Niche by Zack M Davis
Jan 10, 2024
LW - Goodbye, Shoggoth: The Stage, its Animatronics, and the Puppeteer - a New Metaphor by RogerDearnaley
Jan 10, 2024
LW - Learning Math in Time for Alignment by NicholasKross
Jan 09, 2024
LW - A model of research skill by L Rudolf L
Jan 09, 2024
LW - When "yang" goes wrong by Joe Carlsmith
Jan 08, 2024
LW - 2023 Prediction Evaluations by Zvi
Jan 08, 2024
LW - Bayesians Commit the Gambler's Fallacy by Kevin Dorst
Jan 07, 2024
LW - Defending against hypothetical moon life during Apollo 11 by eukaryote
Jan 07, 2024
LW - AI Risk and the US Presidential Candidates by Zane
Jan 07, 2024
LW - The Next ChatGPT Moment: AI Avatars by kolmplex
Jan 06, 2024
LW - Survey of 2,778 AI authors: six parts in pictures by KatjaGrace
Jan 06, 2024
LW - Almost everyone I've met would be well-served thinking more about what to focus on by Henrik Karlsson
Jan 06, 2024
LW - AI #45: To Be Determined by Zvi
Jan 05, 2024
LW - MIRI 2024 Mission and Strategy Update by Malo
Jan 05, 2024
LW - Deep atheism and AI risk by Joe Carlsmith
Jan 04, 2024
LW - Copyright Confrontation #1 by Zvi
Jan 04, 2024
LW - Some Vacation Photos by johnswentworth
Jan 04, 2024
LW - Safety First: safety before full alignment. The deontic sufficiency hypothesis. by Chipmonk
Jan 04, 2024
LW - Trading off Lives by jefftk
Jan 03, 2024
LW - AI Is Not Software by Davidmanheim
Jan 02, 2024
LW - Stop talking about p(doom) by Isaac King
Jan 02, 2024
LW - Gentleness and the artificial Other by Joe Carlsmith
Jan 02, 2024
LW - Apologizing is a Core Rationalist Skill by johnswentworth
Jan 02, 2024
LW - Boston Solstice 2023 Retrospective by jefftk
Jan 02, 2024
LW - Bayesian updating in real life is mostly about understanding your hypotheses by Max H
Jan 01, 2024
LW - 2023 in AI predictions by jessicata
Jan 01, 2024
LW - Planning to build a cryptographic box with perfect secrecy by Lysandre Terrisse
Jan 01, 2024
LW - Dark Skies Book Review by PeterMcCluskey
Dec 31, 2023
LW - shoes with springs by bhauth
Dec 31, 2023
LW - Taking responsibility and partial derivatives by Ruby
Dec 31, 2023
LW - The proper response to mistakes that have harmed others? by Ruby
Dec 31, 2023
LW - The Plan - 2023 Version by johnswentworth
Dec 30, 2023
LW - Will 2024 be very hot? Should we be worried? by A.H.
Dec 29, 2023
LW - NYT is suing OpenAIandMicrosoft for alleged copyright infringement; some quick thoughts by Mikhail Samin
Dec 28, 2023
LW - In Defense of Epistemic Empathy by Kevin Dorst
Dec 28, 2023
LW - How Emergency Medicine Solves the Alignment Problem by StrivingForLegibility
Dec 27, 2023
LW - Environmental allergies are curable? (Sublingual immunotherapy) by Chipmonk
Dec 27, 2023
LW - AI's impact on biology research: Part I, today by octopocta
Dec 27, 2023
LW - METR is hiring! by Beth Barnes
Dec 27, 2023
LW - How "Pause AI" advocacy could be net harmful by Tamsin Leake
Dec 26, 2023
LW - Flagging Potentially Unfair Parenting by jefftk
Dec 26, 2023
LW - A Crisper Explanation of Simulacrum Levels by Thane Ruthenis
Dec 24, 2023
LW - AI Girlfriends Won't Matter Much by Maxwell Tabarrok
Dec 23, 2023
LW - The problem with infohazards as a concept [Linkpost] by Noosphere89
Dec 22, 2023
LW - Pseudonymity and Accusations by jefftk
Dec 22, 2023
LW - The LessWrong 2022 Review: Review Phase by RobertM
Dec 22, 2023
LW - AI Safety Chatbot by markov
Dec 22, 2023
LW - On OpenAI's Preparedness Framework by Zvi
Dec 21, 2023
LW - Most People Don't Realize We Have No Idea How Our AIs Work by Thane Ruthenis
Dec 21, 2023
LW - Prediction Markets aren't Magic by SimonM
Dec 21, 2023
LW - Legalize butanol? by bhauth
Dec 21, 2023
LW - Matrix completion prize results by paulfchristiano
Dec 20, 2023
LW - Goal-Completeness is like Turing-Completeness for AGI by Liron
Dec 20, 2023
LW - Monthly Roundup #13: December 2023 by Zvi
Dec 20, 2023
LW - The Dark Arts by lsusr
Dec 19, 2023
LW - A Universal Emergent Decomposition of Retrieval Tasks in Language Models by Alexandre Variengien
Dec 19, 2023
LW - Constellations are Younger than Continents by Jeffrey Heninger
Dec 19, 2023
LW - OpenAI: Preparedness framework by Zach Stein-Perlman
Dec 18, 2023
LW - The 'Neglected Approaches' Approach: AE Studio's Alignment Agenda by Cameron Berg
Dec 18, 2023
LW - What makes teaching math special by Viliam
Dec 18, 2023
LW - Talk: "AI Would Be A Lot Less Alarming If We Understood Agents" by johnswentworth
Dec 18, 2023
LW - Scale Was All We Needed, At First by Gabriel Mukobi
Dec 18, 2023
LW - The Serendipity of Density by jefftk
Dec 17, 2023
LW - cold aluminum for medicine by bhauth
Dec 17, 2023
LW - 2022 (and All Time) Posts by Pingback Count by Raemon
Dec 17, 2023
LW - "Humanity vs. AGI" Will Never Look Like "Humanity vs. AGI" to Humanity by Thane Ruthenis
Dec 16, 2023
LW - Talking With People Who Speak to Congressional Staffers about AI risk by Eneasz
Dec 16, 2023
LW - Contra Scott on Abolishing the FDA by Maxwell Tabarrok
Dec 15, 2023
LW - "AI Alignment" is a Dangerously Overloaded Term by Roko
Dec 15, 2023
LW - EU policymakers reach an agreement on the AI Act by tlevin
Dec 15, 2023
LW - Some for-profit AI alignment org ideas by Eric Ho
Dec 14, 2023
LW - Love, Reverence, and Life by Elizabeth
Dec 14, 2023
LW - Bayesian Injustice by Kevin Dorst
Dec 14, 2023
LW - Update on Chinese IQ-related gene panels by Lao Mein
Dec 14, 2023
LW - How bad is chlorinated water? by bhauth
Dec 14, 2023
LW - Are There Examples of Overhang for Other Technologies? by Jeffrey Heninger
Dec 14, 2023
LW - Is being sexy for your homies? by Valentine
Dec 13, 2023
LW - The Best of Don't Worry About the Vase by Zvi
Dec 13, 2023
LW - AI Views Snapshots by Rob Bensinger
Dec 13, 2023
LW - Enhancing intelligence by banging your head on the wall by Bezzi
Dec 13, 2023
LW - [Valence series] 3. Valence and Beliefs by Steven Byrnes
Dec 13, 2023
LW - Balsa Update and General Thank You by Zvi
Dec 13, 2023
LW - Funding case: AI Safety Camp by Remmelt
Dec 12, 2023
LW - OpenAI: Leaks Confirm the Story by Zvi
Dec 12, 2023
LW - What is the next level of rationality? by lsusr
Dec 12, 2023
LW - Secondary Risk Markets by Vaniver
Dec 12, 2023
LW - The Consciousness Box by GradualImprovement
Dec 12, 2023
LW - The likely first longevity drug is based on sketchy science. This is bad for science and bad for longevity. by BobBurgers
Dec 12, 2023
LW - On plans for a functional society by kave
Dec 12, 2023
LW - re: Yudkowsky on biological materials by bhauth
Dec 11, 2023
LW - Principles For Product Liability (With Application To AI) by johnswentworth
Dec 10, 2023
LW - Unpicking Extinction by ukc10014
Dec 10, 2023
LW - Some negative steganography results by Fabien Roger
Dec 09, 2023
LW - The Offense-Defense Balance Rarely Changes by Maxwell Tabarrok
Dec 09, 2023
LW - "Model UN Solutions" by Arjun Panickssery
Dec 09, 2023
LW - Refusal mechanisms: initial experiments with Llama-2-7b-chat by andyrdt
Dec 08, 2023
LW - What I Would Do If I Were Working On AI Governance by johnswentworth
Dec 08, 2023
LW - [Valence series] 2. Valence and Normativity by Steven Byrnes
Dec 08, 2023
LW - Is AlphaGo actually a consequentialist utility maximizer? by faul sname
Dec 08, 2023
LW - Meetup Tip: Heartbeat Messages by Screwtape
Dec 07, 2023
LW - Gemini 1.0 by Zvi
Dec 07, 2023
LW - On Trust by johnswentworth
Dec 07, 2023
LW - Anthropical Paradoxes are Paradoxes of Probability Theory by Ape in the coat
Dec 07, 2023
LW - Originality vs. Correctness by alkjash
Dec 06, 2023
LW - Based Beff Jezos and the Accelerationists by Zvi
Dec 06, 2023
LW - A Socratic dialogue with my student by lsusr
Dec 06, 2023
LW - On 'Responsible Scaling Policies' (RSPs) by Zvi
Dec 06, 2023
LW - How do you feel about LessWrong these days? [Open feedback thread] by jacobjacob
Dec 05, 2023
LW - We're all in this together by Tamsin Leake
Dec 05, 2023
LW - The LessWrong 2022 Review by habryka
Dec 05, 2023
LW - Speaking to Congressional staffers about AI risk by Akash
Dec 05, 2023
LW - n of m ring signatures by DanielFilan
Dec 05, 2023
LW - [Valence series] 1. Introduction by Steven Byrnes
Dec 04, 2023
LW - Book Review: 1948 by Benny Morris by Yair Halberstadt
Dec 04, 2023
LW - Meditations on Mot by Richard Ngo
Dec 04, 2023
LW - List of strategies for mitigating deceptive alignment by joshc
Dec 04, 2023
LW - Nietzsche's Morality in Plain English by Arjun Panickssery
Dec 04, 2023
LW - the micro-fulfillment cambrian explosion by bhauth
Dec 04, 2023
LW - The Witness by Richard Ngo
Dec 04, 2023
LW - Quick takes on "AI is easy to control" by So8res
Dec 02, 2023
LW - Out-of-distribution Bioattacks by jefftk
Dec 02, 2023
LW - 2023 Unofficial LessWrong Census/Survey by Screwtape
Dec 02, 2023
LW - Complex systems research as a field (and its relevance to AI Alignment) by Nora Ammann
Dec 02, 2023
LW - MATS Summer 2023 Postmortem by Rocket
Dec 02, 2023
LW - Queuing theory: Benefits of operating at 70% capacity by ampdot
Dec 02, 2023
LW - How useful is mechanistic interpretability? by ryan greenblatt
Dec 01, 2023
LW - What's next for the field of Agent Foundations? by Nora Ammann
Nov 30, 2023
LW - Scaling laws for dominant assurance contracts by jessicata
Nov 30, 2023
LW - OpenAI: Altman Returns by Zvi
Nov 30, 2023
LW - Stupid Question: Why am I getting consistently downvoted? by MadHatter
Nov 30, 2023
LW - Lying Alignment Chart by Zack M Davis
Nov 29, 2023
LW - How to Control an LLM's Behavior (why my P(DOOM) went down) by RogerDearnaley
Nov 29, 2023
LW - Black Box Biology by GeneSmith
Nov 29, 2023
LW - The 101 Space You Will Always Have With You by Screwtape
Nov 29, 2023
LW - I'm confused about innate smell neuroanatomy by Steven Byrnes
Nov 29, 2023
LW - AISC 2024 - Project Summaries by NickyP
Nov 29, 2023
LW - Update #2 to "Dominant Assurance Contract Platform": EnsureDone by moyamo
Nov 28, 2023
LW - [Linkpost] George Mack's Razors by trevor
Nov 28, 2023
LW - Apply to the Conceptual Boundaries Workshop for AI Safety by Chipmonk
Nov 28, 2023
LW - My techno-optimism [By Vitalik Buterin] by habryka
Nov 28, 2023
LW - "Epistemic range of motion" and LessWrong moderation by habryka
Nov 28, 2023
LW - Apocalypse insurance, and the hardline libertarian take on AI risk by So8res
Nov 28, 2023
LW - Paper: "FDT in an evolutionary environment" by the gears to ascension
Nov 27, 2023
LW - why did OpenAI employees sign by bhauth
Nov 27, 2023
LW - Spaced repetition for teaching two-year olds how to read (Interview) by Chipmonk
Nov 27, 2023
LW - Moral Reality Check (a short story) by jessicata
Nov 26, 2023
LW - What are the results of more parental supervision and less outdoor play? by juliawise
Nov 25, 2023
LW - Progress links digest, 2023-11-24: Bottlenecks of aging, Starship launches, and much more by jasoncrawford
Nov 25, 2023
LW - Prepsgiving, A Convergently Instrumental Human Practice by JenniferRM
Nov 25, 2023
LW - What did you change your mind about in the last year? by mike hawke
Nov 24, 2023
LW - Never Drop A Ball by Screwtape
Nov 24, 2023
LW - AI #39: The Week of OpenAI by Zvi
Nov 23, 2023
LW - Possible OpenAI's Q* breakthrough and DeepMind's AlphaGo-type systems plus LLMs by Burny
Nov 23, 2023
LW - so you want to save the world? an account in paladinhood by Tamsin Leake
Nov 22, 2023
LW - OpenAI: The Battle of the Board by Zvi
Nov 22, 2023
LW - Atlantis: Berkeley event venue available for rent by Jonas Vollmer
Nov 22, 2023
LW - Userscript to always show LW comments in context vs at the top by Vlad Sitalo
Nov 21, 2023
LW - Dialogue on the Claim: "OpenAI's Firing of Sam Altman (And Shortly-Subsequent Events) On Net Reduced Existential Risk From AGI" by johnswentworth
Nov 21, 2023
LW - Why not electric trains and excavators? by bhauth
Nov 21, 2023
LW - Navigating emotions in an uncertain and confusing world by Akash
Nov 21, 2023
LW - For Civilization and Against Niceness by Gabriel Alfour
Nov 21, 2023
LW - Vote on worthwhile OpenAI topics to discuss by Ben Pace
Nov 21, 2023
LW - Agent Boundaries Aren't Markov Blankets. [no longer endorsed] by abramdemski
Nov 20, 2023
LW - OpenAI Staff (including Sutskever) Threaten to Quit Unless Board Resigns by Seth Herd
Nov 20, 2023
LW - OpenAI: Facts from a Weekend by Zvi
Nov 20, 2023
LW - Sam Altman, Greg Brockman and others from OpenAI join Microsoft by Ozyrus
Nov 20, 2023
LW - New paper shows truthfulness and instruction-following don't generalize by default by joshc
Nov 19, 2023
LW - "Why can't you just turn it off?" by Roko
Nov 19, 2023
LW - Spaciousness In Partner Dance: A Naturalism Demo by LoganStrohl
Nov 19, 2023
LW - Altman firing retaliation incoming? by trevor
Nov 19, 2023
LW - Sam Altman fired from OpenAI by LawrenceC
Nov 17, 2023
LW - On the lethality of biased human reward ratings by Eli Tyre
Nov 17, 2023
LW - On Lies and Liars by Gabriel Alfour
Nov 17, 2023
LW - On Tapping Out by Screwtape
Nov 17, 2023
LW - A to Z of things by KatjaGrace
Nov 17, 2023
LW - Forecasting AI (Overview) by jsteinhardt
Nov 17, 2023
LW - New LessWrong feature: Dialogue Matching by jacobjacob
Nov 16, 2023
LW - Learning coefficient estimation: the details by Zach Furman
Nov 16, 2023
LW - Social Dark Matter by [DEACTIVATED] Duncan Sabien
Nov 16, 2023
LW - In Defense of Parselmouths by Screwtape
Nov 16, 2023
LW - 'Theories of Values' and 'Theories of Agents': confusions, musings and desiderata by Mateusz Bagiński
Nov 16, 2023
LW - Extrapolating from Five Words by Gordon Seidoh Worley
Nov 16, 2023
LW - With or without a scratchpad, Large Language Models can Strategically Deceive their Users when Put Under Pressure. Results of an autonomous stock trading agent in a realistic, simulated environment. by ReaderM
Nov 15, 2023
LW - Testbed evals: evaluating AI safety even when it can't be directly measured by joshc
Nov 15, 2023
LW - Reinforcement Via Giving People Cookies by Screwtape
Nov 15, 2023
LW - Monthly Roundup #12: November 2023 by Zvi
Nov 15, 2023
LW - Raemon's Deliberate ("Purposeful?") Practice Club by Raemon
Nov 15, 2023
LW - Kids or No kids by Kids or no kids
Nov 14, 2023
LW - A framing for interpretability by Nina Rimsky
Nov 14, 2023
LW - What is wisdom? by TsviBT
Nov 14, 2023
LW - When did Eliezer Yudkowsky change his mind about neural networks? by Yarrow Bouchard
Nov 14, 2023
LW - They are made of repeating patterns by quetzal rainbow
Nov 14, 2023
LW - Loudly Give Up, Don't Quietly Fade by Screwtape
Nov 14, 2023
LW - Redirecting one's own taxes as an effective altruism method by David Gross
Nov 13, 2023
LW - Bostrom Goes Unheard by Zvi
Nov 13, 2023
LW - The Fundamental Theorem for measurable factor spaces by Matthias G. Mayer
Nov 13, 2023
LW - You can just spontaneously call people you haven't met in years by lc
Nov 13, 2023
LW - Zvi's Manifold Markets House Rules by Zvi
Nov 13, 2023
LW - Don't Donate A Kidney To A Stranger by George3d6
Nov 12, 2023
LW - It's OK to be biased towards humans by dr s
Nov 12, 2023
LW - Palisade is hiring Research Engineers by Charlie Rogers-Smith
Nov 11, 2023
LW - GPT-2030 and Catastrophic Drives: Four Vignettes by jsteinhardt
Nov 11, 2023
LW - Picking Mentors For Research Programmes by Raymond D
Nov 10, 2023
LW - Text Posts from the Kids Group: 2021 by jefftk
Nov 10, 2023
LW - Making Bad Decisions On Purpose by Screwtape
Nov 10, 2023
LW - Polysemantic Attention Head in a 4-Layer Transformer by Jett
Nov 09, 2023
LW - On OpenAI Dev Day by Zvi
Nov 09, 2023
LW - A free to enter, 240 character, open-source iterated prisoner's dilemma tournament by Isaac King
Nov 09, 2023
LW - Concrete positive visions for a future without AGI by Max H
Nov 09, 2023
LW - Five projects from AI Safety Hub Labs 2023 by charlie griffin
Nov 08, 2023
LW - Thinking By The Clock by Screwtape
Nov 08, 2023
LW - The Stochastic Parrot Hypothesis is debatable for the last generation of LLMs by Quentin FEUILLADE--MONTIXI
Nov 08, 2023
LW - The Perils of Professionalism by Screwtape
Nov 08, 2023
LW - Vote on Interesting Disagreements by Ben Pace
Nov 07, 2023
LW - Announcing Athena - Women in AI Alignment Research by Claire Short
Nov 07, 2023
LW - AMA: Earning to Give by jefftk
Nov 07, 2023
LW - On the UK Summit by Zvi
Nov 07, 2023
LW - Job listing: Communications Generalist / Project Manager by Gretta Duleba
Nov 07, 2023
LW - Are language models good at making predictions? by dynomight
Nov 06, 2023
LW - The Assumed Intent Bias by silentbob
Nov 06, 2023
LW - On Overhangs and Technological Change by Roko
Nov 06, 2023
LW - Being good at the basics by dominicq
Nov 06, 2023
LW - Pivotal Acts might Not be what You Think they are by Johannes C. Mayer
Nov 05, 2023
LW - Stuxnet, not Skynet: Humanity's disempowerment by AI by Roko
Nov 04, 2023
LW - We are already in a persuasion-transformed world and must take precautions by trevor
Nov 04, 2023
LW - The Soul Key by Richard Ngo
Nov 04, 2023
LW - The 6D effect: When companies take risks, one email can be very powerful. by scasper
Nov 04, 2023
LW - Deception Chess: Game #1 by Zane
Nov 03, 2023
LW - 8 examples informing my pessimism on uploading without reverse engineering by Steven Byrnes
Nov 03, 2023
LW - Integrity in AI Governance and Advocacy by habryka
Nov 03, 2023
LW - One Day Sooner by Screwtape
Nov 03, 2023
LW - The other side of the tidal wave by KatjaGrace
Nov 03, 2023
LW - Saying the quiet part out loud: trading off x-risk for personal immortality by disturbance
Nov 02, 2023
LW - Propaganda or Science: A Look at Open Source AI and Bioterrorism Risk by 1a3orn
Nov 02, 2023
LW - Estimating effective dimensionality of MNIST models by Arjun Panickssery
Nov 02, 2023
LW - Snapshot of narratives and frames against regulating AI by Jan Kulveit
Nov 02, 2023
LW - Public Weights? by jefftk
Nov 02, 2023
LW - Chinese scientists acknowledge xrisk and call for international regulatory body [Linkpost] by Akash
Nov 02, 2023
LW - Reactions to the Executive Order by Zvi
Nov 01, 2023
LW - 2023 LessWrong Community Census, Request for Comments by Screwtape
Nov 01, 2023
LW - On the Executive Order by Zvi
Nov 01, 2023
LW - Mission Impossible: Dead Reckoning Part 1 AI Takeaways by Zvi
Nov 01, 2023
LW - Urging an International AI Treaty: An Open Letter by Loppukilpailija
Nov 01, 2023
LW - Linkpost: A Post Mortem on the Gino Case by Linch
Oct 24, 2023
LW - What is an "anti-Occamian prior"? by Zane
Oct 23, 2023
LW - Sam Altman's sister, Annie Altman, claims Sam has severely abused her by pl5015
Oct 07, 2023
LW - Monthly Roundup #11: October 2023 by Zvi
Oct 04, 2023
LW - When to Get the Booster? by jefftk
Oct 04, 2023
LW - OpenAI-Microsoft partnership by Zach Stein-Perlman
Oct 04, 2023
LW - energy landscapes of experts by bhauth
Oct 03, 2023
LW - Linkpost: They Studied Dishonesty. Was Their Work a Lie? by Linch
Oct 02, 2023
LW - Thomas Kwa's MIRI research experience by Thomas Kwa
Oct 02, 2023
LW - Conditionals All The Way Down by lunatic at large
Oct 02, 2023
LW - The 99% principle for personal problems by Kaj Sotala
Oct 02, 2023
LW - Fifty Flips by abstractapplic
Oct 02, 2023
LW - My Effortless Weightloss Story: A Quick Runthrough by CuoreDiVetro
Oct 01, 2023
LW - Competitive, Cooperative, and Cohabitive by Screwtape
Oct 01, 2023
LW - The Lighthaven Campus is open for bookings by habryka
Sep 29, 2023
LW - Announcing FAR Labs, an AI safety coworking space by bgold
Sep 29, 2023
LW - Bids To Defer On Value Judgements by johnswentworth
Sep 29, 2023
LW - What's your standard for good work performance? by Chi Nguyen
Sep 29, 2023
LW - High-level interpretability: detecting an AI's objectives by Paul Colognese
Sep 29, 2023
LW - Peacewagers so Far by mako yass
Sep 29, 2023
LW - The point of a game is not to win, and you shouldn't even pretend that it is by mako yass
Sep 29, 2023
LW - EA Vegan Advocacy is not truthseeking, and it's everyone's problem by Elizabeth
Sep 28, 2023
LW - AI #31: It Can Do What Now? by Zvi
Sep 28, 2023
LW - The Hidden Complexity of Wishes - The Animation by Writer
Sep 28, 2023
LW - Petrov Day Retrospective, 2023 (re: the most important virtue of Petrov Day and unilaterally promoting it) by Ruby
Sep 28, 2023
LW - Jacob on the Precipice by Richard Ngo
Sep 27, 2023
LW - GPT-4 for personal productivity: online distraction blocker by Sergii
Sep 27, 2023
LW - Making AIs less likely to be spiteful by Nicolas Macé
Sep 26, 2023
LW - The King and the Golem by Richard Ngo
Sep 26, 2023
LW - "X distracts from Y" as a thinly-disguised fight over group status / politics by Steven Byrnes
Sep 25, 2023
LW - Inside Views, Impostor Syndrome, and the Great LARP by johnswentworth
Sep 25, 2023
LW - Interpreting OpenAI's Whisper by EllenaR
Sep 24, 2023
LW - I designed an AI safety course (for a philosophy department) by Eleni Angelou
Sep 24, 2023
LW - Paper: LLMs trained on "A is B" fail to learn "B is A" by lberglund
Sep 23, 2023
LW - Luck based medicine: inositol for anxiety and brain fog by Elizabeth
Sep 23, 2023
LW - Update to "Dominant Assurance Contract Platform" by moyamo
Sep 23, 2023
LW - Fund Transit With Development by jefftk
Sep 22, 2023
LW - Let's talk about Impostor syndrome in AI safety by Igor Ivanov
Sep 22, 2023
LW - Neel Nanda on the Mechanistic Interpretability Researcher Mindset by Michaël Trazzi
Sep 22, 2023
LW - If influence functions are not approximating leave-one-out, how are they supposed to help? by Fabien Roger
Sep 22, 2023
LW - Immortality or death by AGI by ImmortalityOrDeathByAGI
Sep 22, 2023
LW - Atoms to Agents Proto-Lectures by johnswentworth
Sep 22, 2023
LW - Would You Work Harder In The Least Convenient Possible World? by Firinn
Sep 22, 2023
LW - AI #30: Dalle-3 and GPT-3.5-Instruct-Turbo by Zvi
Sep 21, 2023
LW - Interpretability Externalities Case Study - Hungry Hungry Hippos by Magdalena Wache
Sep 20, 2023
LW - [Review] Move First, Think Later: Sense and Nonsense in Improving Your Chess by Arjun Panickssery
Sep 19, 2023
LW - Luck based medicine: angry eldritch sugar gods edition by Elizabeth
Sep 19, 2023
LW - Anthropic's Responsible Scaling Policy and Long-Term Benefit Trust by Zac Hatfield-Dodds
Sep 19, 2023
LW - Some reasons why I frequently prefer communicating via text by Adam Zerner
Sep 19, 2023
LW - Technical AI Safety Research Landscape [Slides] by Magdalena Wache
Sep 18, 2023
LW - Microdooms averted by working on AI Safety by nikola
Sep 18, 2023
LW - Show LW: Get a phone call if prediction markets predict nuclear war by Lorenzo
Sep 18, 2023
LW - The Talk: a brief explanation of sexual dimorphism by Malmesbury
Sep 18, 2023
LW - Three ways interpretability could be impactful by Arthur Conmy
Sep 18, 2023
LW - Eugenics Performed By A Blind, Idiot God by omnizoid
Sep 18, 2023
LW - Actually, "personal attacks after object-level arguments" is a pretty good rule of epistemic conduct by Max H
Sep 18, 2023
LW - Memory bandwidth constraints imply economies of scale in AI inference by Ege Erdil
Sep 17, 2023
LW - I compiled a ebook of 'Project Lawful' for eBook readers by OrwellGoesShopping
Sep 16, 2023
LW - Navigating an ecosystem that might or might not be bad for the world by habryka
Sep 16, 2023
LW - Deconfusing Regret by Alex Hollow
Sep 15, 2023
LW - A Theory of Laughter - Follow-Up by Steven Byrnes
Sep 15, 2023
LW - "Did you lock it?" by ymeskhout
Sep 15, 2023
LW - Can I take ducks home from the park? by dynomight
Sep 14, 2023
LW - Uncovering Latent Human Wellbeing in LLM Embeddings by ChengCheng
Sep 14, 2023
LW - Padding the Corner by jefftk
Sep 14, 2023
LW - AI #29: Take a Deep Breath by Zvi
Sep 14, 2023
LW - Instrumental Convergence Bounty by Logan Zoellner
Sep 14, 2023
LW - Highlights: Wentworth, Shah, and Murphy on "Retargeting the Search" by RobertM
Sep 14, 2023
LW - Linkpost for Jan Leike on Self-Exfiltration by Daniel Kokotajlo
Sep 14, 2023
LW - Is AI Safety dropping the ball on privacy? by markov
Sep 13, 2023
LW - Contra Yudkowsky on Epistemic Conduct for Author Criticism by Zack M Davis
Sep 13, 2023
LW - UDT shows that decision theory is more puzzling than ever by Wei Dai
Sep 13, 2023
LW - PSA: The community is in Berkeley/Oakland, not "the Bay Area" by maia
Sep 11, 2023
LW - Logical Share Splitting by DaemonicSigil
Sep 11, 2023
LW - Focus on the Hardest Part First by Johannes C. Mayer
Sep 11, 2023
LW - US presidents discuss AI alignment agendas by TurnTrout
Sep 09, 2023
LW - Have Attention Spans Been Declining? by niplav
Sep 09, 2023
LW - What is the optimal frontier for due diligence? by RobertM
Sep 09, 2023
LW - Sum-threshold attacks by TsviBT
Sep 08, 2023
LW - AI#28: Watching and Waiting by Zvi
Sep 08, 2023
LW - Recreating the caring drive by Catnee
Sep 08, 2023
LW - A quick update from Nonlinear by KatWoods
Sep 07, 2023
LW - Feedback-loops, Deliberate Practice, and Transfer Learning by jacobjacob
Sep 07, 2023
LW - My First Post by Jaivardhan Nawani
Sep 07, 2023
LW - Sharing Information About Nonlinear by Ben Pace
Sep 07, 2023
LW - Find Hot French Food Near Me: A Follow-up by aphyer
Sep 06, 2023
LW - Who Has the Best Food? by Zvi
Sep 05, 2023
LW - Text Posts from the Kids Group: 2023 I by jefftk
Sep 05, 2023
LW - a rant on politician-engineer coalitional conflict by bhauth
Sep 04, 2023
LW - Defunding My Mistake by ymeskhout
Sep 04, 2023
LW - Tensor Trust: An online game to uncover prompt injection vulnerabilities by Luke Bailey
Sep 04, 2023
LW - The goal of physics by Jim Pivarski
Sep 03, 2023
LW - The smallest possible button by Neil
Sep 02, 2023
LW - One Minute Every Moment by abramdemski
Sep 01, 2023
LW - Reproducing ARC Evals' recent report on language model agents by Thomas Broadley
Sep 01, 2023
LW - A Golden Age of Building? Excerpts and lessons from Empire State, Pentagon, Skunk Works and SpaceX by jacobjacob
Sep 01, 2023
LW - AI #27: Portents of Gemini by Zvi
Aug 31, 2023
LW - Responses to apparent rationalist confusions about game / decision theory by Anthony DiGiovanni
Aug 31, 2023
LW - Report on Frontier Model Training by YafahEdelman
Aug 31, 2023
LW - Biosecurity Culture, Computer Security Culture by jefftk
Aug 30, 2023
LW - Open Call for Research Assistants in Developmental Interpretability by Jesse Hoogland
Aug 30, 2023
LW - The Economics of the Asteroid Deflection Problem by moyamo
Aug 30, 2023
LW - Trying a Wet Suit by jefftk
Aug 29, 2023
LW - Broken Benchmark: MMLU by awg
Aug 29, 2023
LW - Dating Roundup #1: This is Why You're Single by Zvi
Aug 29, 2023
LW - Information warfare historically revolved around human conduits by trevor
Aug 29, 2023
LW - ACX Meetups Everywhere 2023: Times and Places by Scott Alexander
Aug 29, 2023
LW - Introducing the Center for AI Policy (and we're hiring!) by Thomas Larsen
Aug 28, 2023
LW - AI pause/governance advocacy might be net-negative, especially without focus on explaining the x-risk by Mikhail Samin
Aug 28, 2023
LW - Dear Self; we need to talk about ambition by Elizabeth
Aug 28, 2023
LW - Aumann-agreement is common by tailcalled
Aug 27, 2023
LW - Digital brains beat biological ones because diffusion is too slow by GeneSmith
Aug 26, 2023
LW - Red-teaming language models via activation engineering by Nina Rimsky
Aug 26, 2023
LW - When Omnipotence is Not Enough by lsusr
Aug 26, 2023
LW - Assume Bad Faith by Zack M Davis
Aug 25, 2023
LW - Apply for the 2023 Developmental Interpretability Conference! by Stan van Wingerden
Aug 25, 2023
LW - AI #26: Fine Tuning Time by Zvi
Aug 25, 2023
LW - AI Regulation May Be More Important Than AI Alignment For Existential Safety by otto.barten
Aug 24, 2023
LW - The lost millennium by Ege Erdil
Aug 24, 2023
LW - Assessment of intelligence agency functionality is difficult yet important by trevor
Aug 24, 2023
LW - The God of Humanity, and the God of the Robot Utilitarians by Raemon
Aug 24, 2023
LW - Seth Explains Consciousness by Jacob Falkovich
Aug 24, 2023
LW - The Low-Hanging Fruit Prior and sloped valleys in the loss landscape by Dmitry Vaintrob
Aug 24, 2023
LW - Diet Experiment Preregistration: Long-term water fasting + seed oil removal by lc
Aug 24, 2023
LW - Why Is No One Trying To Align Profit Incentives With Alignment Research? by Prometheus
Aug 23, 2023
LW - A Theory of Laughter by Steven Byrnes
Aug 23, 2023
LW - Walk while you talk: don't balk at "no chalk" by dkl9
Aug 23, 2023
LW - State of Generally Available Self-Driving by jefftk
Aug 22, 2023
LW - Large Language Models will be Great for Censorship by Ethan Edwards
Aug 22, 2023
LW - Which possible AI systems are relatively safe? by Zach Stein-Perlman
Aug 22, 2023
LW - DIY Deliberate Practice by lynettebye
Aug 21, 2023
LW - Ideas for improving epistemics in AI safety outreach by mic
Aug 21, 2023
LW - Efficiency and resource use scaling parity by Ege Erdil
Aug 21, 2023
LW - Ruining an expected-log-money maximizer by philh
Aug 21, 2023
LW - Chess as a case study in hidden capabilities in ChatGPT by AdamYedidia
Aug 21, 2023
LW - Steven Wolfram on AI Alignment by Bill Benzon
Aug 21, 2023
LW - "Dirty concepts" in AI alignment discourses, and some guesses for how to deal with them by Nora Ammann
Aug 20, 2023
LW - AI Forecasting: Two Years In by jsteinhardt
Aug 20, 2023
LW - Is Chinese total factor productivity lower today than it was in 1956? by Ege Erdil
Aug 19, 2023
LW - The U.S. is mildly destabilizing by lc
Aug 18, 2023
LW - 6 non-obvious mental health issues specific to AI safety. by Igor Ivanov
Aug 18, 2023
LW - Announcing Foresight Institute's AI Safety Grants Program by Allison Duettmann
Aug 18, 2023
LW - What does it mean to "trust science"? by jasoncrawford
Aug 18, 2023
LW - AI #25: Inflection Point by Zvi
Aug 17, 2023
LW - Against Almost Every Theory of Impact of Interpretability by Charbel-Raphaël
Aug 17, 2023
LW - Reflections on "Making the Atomic Bomb" by boazbarak
Aug 17, 2023
LW - The Dunbar Playbook: A CRM system for your friends by Severin T. Seehrich
Aug 17, 2023
LW - If we had known the atmosphere would ignite by Jeffs
Aug 17, 2023
LW - Book Launch: "The Carving of Reality," Best of LessWrong vol. III by Raemon
Aug 17, 2023
LW - Summary of and Thoughts on the Hotz/Yudkowsky Debate by Zvi
Aug 16, 2023
LW - Understanding and visualizing sycophancy datasets by Nina Rimsky
Aug 16, 2023
LW - Ten Thousand Years of Solitude by agp
Aug 16, 2023
LW - Optical Illusions are Out of Distribution Errors by vitaliya
Aug 15, 2023
LW - My checklist for publishing a blog post by Steven Byrnes
Aug 15, 2023
LW - A short calculation about a Twitter poll by Ege Erdil
Aug 14, 2023
LW - Decomposing independent generalizations in neural networks via Hessian analysis by Dmitry Vaintrob
Aug 14, 2023
LW - Stepping down as moderator on LW by Kaj Sotala
Aug 14, 2023
LW - We Should Prepare for a Larger Representation of Academia in AI Safety by Leon Lang
Aug 13, 2023
LW - [Linkpost] Personal and Psychological Dimensions of AI Researchers Confronting AI Catastrophic Risks by Bogdan Ionut Cirstea
Aug 13, 2023
LW - Simulate the CEO by robotelvis
Aug 13, 2023
LW - Biological Anchors: The Trick that Might or Might Not Work by Scott Alexander
Aug 12, 2023
LW - AI #24: Week of the Podcast by Zvi
Aug 11, 2023
LW - LLMs are (mostly) not helped by filler tokens by Kshitij Sachan
Aug 10, 2023
LW - marine cloud brightening by bhauth
Aug 10, 2023
LW - Modulating sycophancy in an RLHF model via activation steering by NinaR
Aug 09, 2023
LW - Inflection.ai is a major AGI lab by nikola
Aug 09, 2023
LW - What's A "Market"? by johnswentworth
Aug 09, 2023
LW - Perpetually Declining Population? by jefftk
Aug 08, 2023
LW - A plea for more funding shortfall transparency by porby
Aug 08, 2023
LW - Feedbackloop-first Rationality by Raemon
Aug 07, 2023
LW - 'We're changing the clouds.' An unforeseen test of geoengineering is fueling record ocean warmth by Annapurna
Aug 07, 2023
LW - Problems with Robin Hanson's Quillette Article On AI by DaemonicSigil
Aug 07, 2023
LW - Computational Thread Art by TheMcDouglas
Aug 07, 2023
LW - how 2 tell if ur input is out of distribution given only model weights by dkirmani
Aug 06, 2023
LW - Stomach Ulcers and Dental Cavities by Metacelsus
Aug 06, 2023
LW - The Sinews of Sudan's Latest War by Tim Liptrot
Aug 05, 2023
LW - Private notes on LW? by Raemon
Aug 04, 2023
LW - Password-locked models: a stress case for capabilities evaluation by Fabien Roger
Aug 03, 2023
LW - AI #23: Fundamental Problems with RLHF by Zvi
Aug 03, 2023
LW - "Is There Anything That's Worth More" by Zack M Davis
Aug 02, 2023
LW - My current LK99 questions by Eliezer Yudkowsky
Aug 01, 2023
LW - Barbieheimer: Across the Dead Reckoning by Zvi
Aug 01, 2023
LW - Exercise: Solve "Thinking Physics" by Raemon
Aug 01, 2023
LW - The "public debate" about AI is confusing for the general public and for policymakers because it is a three-sided debate by Adam David Long
Aug 01, 2023
LW - A Social History of Truth by Vaniver
Aug 01, 2023
LW - "Building a House" Review by jefftk
Jul 31, 2023
LW - Lack of Social Grace Is an Epistemic Virtue by Zack M Davis
Jul 31, 2023
LW - Is Light Drinking Protective? by jefftk
Jul 31, 2023
LW - Apollo Neuro Results by Elizabeth
Jul 30, 2023
LW - How to make real-money prediction markets on arbitrary topics by yutaka
Jul 30, 2023
LW - Self-driving car bets by paulfchristiano
Jul 29, 2023
LW - Introductory Textbook to Vision Models Interpretability by jeanne
Jul 29, 2023
LW - Why You Should Never Update Your Beliefs by Arjun Panickssery
Jul 29, 2023
LW - When can we trust model evaluations? by evhub
Jul 29, 2023
LW - Mech Interp Puzzle 2: Word2Vec Style Embeddings by Neel Nanda
Jul 28, 2023
LW - Yes, It's Subjective, But Why All The Crabs? by johnswentworth
Jul 28, 2023
LW - Pulling the Rope Sideways: Empirical Test Results by Daniel Kokotajlo
Jul 28, 2023
LW - Visible loss landscape basins don't correspond to distinct algorithms by Mikhail Samin
Jul 28, 2023
LW - Reducing sycophancy and improving honesty via activation steering by NinaR
Jul 28, 2023
LW - SSA rejects anthropic shadow, too by jessicata
Jul 27, 2023
LW - SSA rejects anthropic shadow, too by jessicata
Jul 27, 2023
LW - Cultivating a state of mind where new ideas are born by Henrik Karlsson
Jul 27, 2023
LW - Cultivating a state of mind where new ideas are born by Henrik Karlsson
Jul 27, 2023
LW - The First Room-Temperature Ambient-Pressure Superconductor by Annapurna
Jul 27, 2023
LW - Llama We Doing This Again? by Zvi
Jul 27, 2023
LW - Why no Roman Industrial Revolution? by jasoncrawford
Jul 26, 2023
LW - Neuronpedia - AI Safety Game by hijohnnylin
Jul 26, 2023
LW - Meta-level adversarial evaluation of oversight techniques might allow robust measurement of their adequacy by Buck
Jul 26, 2023
LW - AI Safety Hub Serbia Soft Launch by DusanDNesic
Jul 26, 2023
LW - Underwater Torture Chambers: The Horror Of Fish Farming by omnizoid
Jul 26, 2023
LW - Thoughts on Loss Landscapes and why Deep Learning works by beren
Jul 25, 2023
LW - Anthropic Observations by Zvi
Jul 25, 2023
LW - How LLMs are and are not myopic by janus
Jul 25, 2023
LW - Open problems in activation engineering by TurnTrout
Jul 24, 2023
LW - A reformulation of Finite Factored Sets by Matthias G. Mayer
Jul 24, 2023
LW - Slowing down AI progress is an underexplored alignment strategy by Norman Borlaug
Jul 24, 2023
LW - Brain Efficiency Cannell Prize Contest Award Ceremony by Alexander Gietelink Oldenziel
Jul 24, 2023
LW - Cryonics and Regret by MvB
Jul 24, 2023
LW - Rationality !== winning by Raemon
Jul 24, 2023
LW - Which rationality posts are begging for further practical development? by LoganStrohl
Jul 24, 2023
LW - QAPR 5: grokking is maybe not that big a deal? by Quintin Pope
Jul 23, 2023
LW - Autogynephilia discourse is so absurdly bad on all sides by tailcalled
Jul 23, 2023
LW - "Justice, Cherryl." by Zack M Davis
Jul 23, 2023
LW - The UAP Disclosure Act of 2023 and its implications by andeslodes
Jul 23, 2023
LW - BCIs and the ecosystem of modular minds by beren
Jul 22, 2023
LW - All AGI Safety questions welcome (especially basic ones) [July 2023] by smallsilo
Jul 22, 2023
LW - I'm consistently overwhelmed by basic obligations. Are there any paradigm shifts or other rationality-based tips that would be helpful? by Benjamin Hendricks
Jul 21, 2023
LW - News : Biden-⁠Harris Administration Secures Voluntary Commitments from Leading Artificial Intelligence Companies to Manage the Risks Posed by AI by Jonathan Claybrough
Jul 21, 2023
LW - Priorities for the UK Foundation Models Taskforce by Andrea Miotti
Jul 21, 2023
LW - Boundary Placement Rebellion by tailcalled
Jul 21, 2023
LW - Announcement: AI Narrations Available for All New LessWrong Posts by Solenoid Entity
Jul 21, 2023
LW - Even Superhuman Go AIs Have Surprising Failures Modes by AdamGleave
Jul 20, 2023
LW - Does Circuit Analysis Interpretability Scale? Evidence from Multiple Choice Capabilities in Chinchilla by Neel Nanda
Jul 20, 2023
LW - Still no Lie Detector for LLMs by Whispermute
Jul 19, 2023
LW - Alignment Grantmaking is Funding-Limited Right Now by johnswentworth
Jul 19, 2023
LW - A brief history of computers by Adam Zerner
Jul 19, 2023
LW - Tiny Mech Interp Projects: Emergent Positional Embeddings of Words by Neel Nanda
Jul 19, 2023
LW - Meta announces Llama 2; "open sources" it for commercial use by LawrenceC
Jul 19, 2023
LW - Measuring and Improving the Faithfulness of Model-Generated Reasoning by Ansh Radhakrishnan
Jul 18, 2023
LW - Proof of posteriority: a defense against AI-generated misinformation by jchan
Jul 18, 2023
LW - Existential Risk Persuasion Tournament by PeterMcCluskey
Jul 18, 2023
LW - Predictive history classes by dkl9
Jul 18, 2023
LW - AutoInterpretation Finds Sparse Coding Beats Alternatives by Hoagy
Jul 17, 2023
LW - Sapient Algorithms by Valentine
Jul 17, 2023
LW - Thoughts on "Process-Based Supervision" by Steven Byrnes
Jul 17, 2023
LW - An upcoming US Supreme Court case may impede AI governance efforts by NickGabs
Jul 17, 2023
LW - Mech Interp Puzzle 1: Suspiciously Similar Embeddings in GPT-Neo by Neel Nanda
Jul 17, 2023
LW - Activation adding experiments with llama-7b by NinaR
Jul 16, 2023
LW - Robustness of Model-Graded Evaluations and Automated Interpretability by Simon Lermen
Jul 16, 2023
LW - Why was the AI Alignment community so unprepared for this moment? by Ras1513
Jul 15, 2023
LW - When Someone Tells You They're Lying, Believe Them by ymeskhout
Jul 14, 2023
LW - The Goddess of Everything Else - The Animation by Writer
Jul 13, 2023
LW - Jailbreaking GPT-4's code interpreter by nikolaisalreadytaken
Jul 13, 2023
LW - How can I get help becoming a better rationalist? by TeaTieAndHat
Jul 13, 2023
LW - Winners of AI Alignment Awards Research Contest by Akash
Jul 13, 2023
LW - Are there any good, easy-to-understand examples of cases where statistical causal network discovery worked well in practice? by tailcalled
Jul 13, 2023
LW - Accidentally Load Bearing by jefftk
Jul 13, 2023
LW - Elon Musk announces xAI by Jan Kulveit
Jul 13, 2023
LW - What does the launch of x.ai mean for AI Safety? by Chris Leong
Jul 13, 2023
LW - Alignment Megaprojects: You're Not Even Trying to Have Ideas by NicholasKross
Jul 13, 2023
LW - Report on modeling evidential cooperation in large worlds by Johannes Treutlein
Jul 12, 2023
LW - A review of Principia Qualia by jessicata
Jul 12, 2023
LW - Towards Developmental Interpretability by Jesse Hoogland
Jul 12, 2023
LW - A transcript of the TED talk by Eliezer Yudkowsky by Mikhail Samin
Jul 12, 2023
LW - My Weirdest Experience by Bridgett Kay
Jul 12, 2023
LW - Why it's necessary to shoot yourself in the foot by g-w1
Jul 12, 2023
LW - Aging and the geroscience hypothesis by DirectedEvolution
Jul 12, 2023
LW - Introducing Fatebook: the fastest way to make and track predictions by Adam B
Jul 11, 2023
LW - OpenAI Launches Superalignment Taskforce by Zvi
Jul 11, 2023
LW - [UPDATE: deadline extended to July 24!] New wind in rationality’s sails: Applications for Epistea Residency 2023 are now open by Jana Meixnerová
Jul 11, 2023
LW - Drawn Out: a story by Richard Ngo
Jul 11, 2023
LW - Open-minded updatelessness by Nicolas Macé
Jul 11, 2023
LW - I think Michael Bailey's dismissal of my autogynephilia questions for Scott Alexander and Aella makes very little sense by tailcalled
Jul 10, 2023
LW - “Reframing Superintelligence” + LLMs + 4 years by Eric Drexler
Jul 10, 2023
LW - Consider Joining the UK Foundation Model Taskforce by Zvi
Jul 10, 2023
LW - Consciousness as a conflationary alliance term by Andrew Critch
Jul 10, 2023
LW - Some reasons to not say "Doomer" by Ruby
Jul 10, 2023
LW - The Seeker’s Game - Vignettes from the Bay by Yulia
Jul 09, 2023
LW - Really Strong Features Found in Residual Stream by Logan Riggs
Jul 09, 2023
LW - Taboo Truth by Tomás B.
Jul 09, 2023
LW - Fixed Point: a love story by Richard Ngo
Jul 08, 2023
LW - Views on when AGI comes and on strategy to reduce existential risk by TsviBT
Jul 08, 2023
LW - What Does LessWrong/EA Think of Human Intelligence Augmentation as of mid-2023? by marc/er
Jul 08, 2023
LW - Passing the ideological Turing test? Arguments against existential risk from AI. by NinaR
Jul 07, 2023
LW - Apparently, of the 195 Million the DoD allocated in University Research Funding Awards in 2022, more than half of them concerned AI or compute hardware research by mako yass
Jul 07, 2023
LW - Introducing bayescalc.io by Adele Lopez
Jul 07, 2023
LW - What are the best non-LW places to read on alignment progress? by Raemon
Jul 07, 2023
LW - Jesse Hoogland on Developmental Interpretability and Singular Learning Theory by Michaël Trazzi
Jul 06, 2023
LW - Agency begets agency (the world is malleable) by Richard Ngo
Jul 06, 2023
LW - AI #19: Hofstadter, Sutskever, Leike by Zvi
Jul 06, 2023
LW - AI labs' statements on governance by Zach Stein-Perlman
Jul 06, 2023
LW - Optimized for Something other than Winning or: How Cricket Resists Moloch and Goodhart's Law by A.H.
Jul 06, 2023
LW - (tentatively) Found 600+ Monosemantic Features in a Small LM Using Sparse Autoencoders by Logan Riggs
Jul 05, 2023
LW - [Linkpost] Introducing Superalignment by beren
Jul 05, 2023
LW - Dominant Assurance Contract Experiment #2: Berkeley House Dinners by Arjun Panickssery
Jul 05, 2023
LW - Twitter Twitches by Zvi
Jul 05, 2023
LW - Ways I Expect AI Regulation To Increase Extinction Risk by 1a3orn
Jul 04, 2023
LW - Two Percolation Puzzles by Adam Scherlis
Jul 04, 2023
LW - The literature on aluminum adjuvants is very suspicious. Small IQ tax is plausible - can any experts help me estimate it? by mikes
Jul 04, 2023
LW - Ten Levels of AI Alignment Difficulty by Sammy Martin
Jul 04, 2023
LW - Monthly Roundup #8: July 2023 by Zvi
Jul 04, 2023
LW - Frames in context by Richard Ngo
Jul 03, 2023
LW - Meta-rationality and frames by Richard Ngo
Jul 03, 2023
LW - Douglas Hofstadter changes his mind on Deep Learning and AI risk (June 2023)? by gwern
Jul 03, 2023
LW - Going Crazy and Getting Better Again by Evenstar
Jul 02, 2023
LW - Why it's so hard to talk about Consciousness by Rafael Harth
Jul 02, 2023
LW - Consider giving money to people, not projects or organizations by NinaR
Jul 02, 2023
LW - Micro Habits that Improve One’s Day by silentbob
Jul 02, 2023
LW - Forum Karma: view stats and find highly-rated comments for any LW user by Max H
Jul 02, 2023
LW - Alpha by Erich Grunewald
Jul 02, 2023
LW - Grant applications and grand narratives by Elizabeth
Jul 02, 2023
LW - My "trauma" frame by Raemon
Jul 01, 2023
LW - On household dust by NinaR
Jun 30, 2023
LW - Introducing EffiSciences’ AI Safety Unit by WCargo
Jun 30, 2023
LW - I Think Eliezer Should Go on Glenn Beck by Lao Mein
Jun 30, 2023
LW - The Case for Overconfidence is Overstated by Kevin Dorst
Jun 30, 2023
LW - AI #18: The Great Debate Debate by Zvi
Jun 29, 2023
LW - Anthropically Blind: the anthropic shadow is reflectively inconsistent by Christopher King
Jun 29, 2023
LW - A "weak" AGI may attempt an unlikely-to-succeed takeover by RobertM
Jun 29, 2023
LW - What money-pumps exist, if any, for deontologists? by Daniel Kokotajlo
Jun 29, 2023
LW - My research agenda in agent foundations by Alex Altair
Jun 28, 2023
LW - When do "brains beat brawn" in Chess? An experiment by titotal
Jun 28, 2023
LW - Nature: "Stop talking about tomorrow’s AI doomsday when AI poses risks today" by Ben Smith
Jun 28, 2023
LW - Carl Shulman on The Lunar Society (7 hour, two-part podcast) by ESRogs
Jun 28, 2023
LW - Self-Blinded Caffeine RCT by niplav
Jun 28, 2023
LW - On the Cost of Thriving Index by Zvi
Jun 27, 2023
LW - Munk AI debate: confusions and possible cruxes by Steven Byrnes
Jun 27, 2023
LW - Man in the Arena by Richard Ngo
Jun 27, 2023
LW - AI-Plans.com - a contributable compendium by Iknownothing
Jun 27, 2023
LW - 60+ Possible Futures by Stuckwork
Jun 27, 2023
LW - "Safety Culture for AI" is important, but isn't going to be easy by Davidmanheim
Jun 26, 2023
LW - Model, Care, Execution by Ricki Heicklen
Jun 26, 2023
LW - Why am I Me? by dadadarren
Jun 26, 2023
LW - Another medical miracle by Dentin
Jun 25, 2023
LW - Crystal Healing — or the Origins of Expected Utility Maximizers by Alexander Gietelink Oldenziel
Jun 25, 2023
LW - Did Bengio and Tegmark lose a debate about AI x-risk against LeCun and Mitchell? by Karl von Wendt
Jun 25, 2023
LW - Correctly Calibrated Trust by habryka
Jun 25, 2023
LW - Will the growing deer prion epidemic spread to humans? Why not? by eukaryote
Jun 25, 2023
LW - How tall is the Shard, really? by philh
Jun 24, 2023
LW - Automatic Rate Limiting on LessWrong by Raemon
Jun 23, 2023
LW - When is correlation transitive? by Ege Erdil
Jun 23, 2023
LW - Catastrophic Risks from AI #1: Summary by Dan H
Jun 23, 2023
LW - Catastrophic Risks from AI #2: Malicious Use by Dan H
Jun 23, 2023
LW - "textbooks are all you need" by bhauth
Jun 22, 2023
LW - The Hubinger lectures on AGI safety: an introductory lecture series by evhub
Jun 22, 2023
LW - Which personality traits are real? Stress-testing the lexical hypothesis by tailcalled
Jun 22, 2023
LW - Short timelines and slow, continuous takeoff as the safest path to AGI by rosehadshar
Jun 21, 2023
LW - My side of an argument with Jacob Cannell about chip interconnect losses by Steven Byrnes
Jun 21, 2023
LW - My tentative best guess on how EAs and Rationalists sometimes turn crazy by habryka
Jun 21, 2023
LW - Never Fight The Last War by ChristianKl
Jun 21, 2023
LW - Public Transit is not Infinitely Safe by jefftk
Jun 20, 2023
LW - Causality: A Brief Introduction by tom4everitt
Jun 20, 2023
LW - A Friendly Face (Another Failure Story) by Karl von Wendt
Jun 20, 2023
LW - Experiments in Evaluating Steering Vectors by Gytis Daujotas
Jun 20, 2023
LW - Ban development of unpredictable powerful models? by TurnTrout
Jun 20, 2023
LW - 10 quick takes about AGI by Max H
Jun 20, 2023
LW - resolving some neural network mysteries by bhauth
Jun 20, 2023
LW - Mode collapse in RL may be fueled by the update equation by TurnTrout
Jun 20, 2023
LW - Lessons On How To Get Things Right On The First Try by johnswentworth
Jun 20, 2023
LW - My impression of singular learning theory by Ege Erdil
Jun 19, 2023
LW - Guide to rationalist interior decorating by mingyuan
Jun 19, 2023
LW - Solomonoff induction still works if the universe is uncomputable, and its usefulness doesn't require knowing Occam's razor by Christopher King
Jun 18, 2023
LW - A summary of current work in AI governance by constructive
Jun 18, 2023
LW - UK Foundation Model Task Force - Expression of Interest by ojorgensen
Jun 18, 2023
LW - Cryonics Career Survey (more jobs than you think) by Mati Roy
Jun 18, 2023
LW - Adventist Health Study-2 supports pescetarianism more than veganism by Elizabeth
Jun 18, 2023
LW - [Replication] Conjecture's Sparse Coding in Small Transformers by Hoagy
Jun 17, 2023
LW - Are Bayesian methods guaranteed to overfit? by Ege Erdil
Jun 17, 2023
LW - The ones who endure by Richard Ngo
Jun 17, 2023
LW - Updating Drexler's CAIS model by Matthew Barnett
Jun 17, 2023
LW - LLMs Sometimes Generate Purely Negatively-Reinforced Text by Fabien Roger
Jun 16, 2023
LW - Distilling Singular Learning Theory by Liam Carroll
Jun 16, 2023
LW - Rational Animations is looking for an AI Safety scriptwriter, a lead community manager, and other roles. by Writer
Jun 16, 2023
LW - Leveling Up Or Leveling Off? Understanding The Science Behind Skill Plateaus by lynettebye
Jun 16, 2023
LW - Conjecture: A standing offer for public debates on AI by Andrea Miotti
Jun 16, 2023
LW - Looking Back On Ads by jefftk
Jun 15, 2023
LW - I still think it's very unlikely we're observing alien aircraft by dynomight
Jun 15, 2023
LW - Why libertarians are advocating for regulation on AI by RobertM
Jun 15, 2023
LW - On the Apple Vision Pro by Zvi
Jun 15, 2023
LW - Anthropic | Charting a Path to AI Accountability by Gabriel Mukobi
Jun 14, 2023
LW - Contingency: A Conceptual Tool from Evolutionary Biology for Alignment by clem acs
Jun 14, 2023
LW - Multiple stages of fallacy - justifications and non-justifications for the multiple stage fallacy by AronT
Jun 14, 2023
LW - Lightcone Infrastructure is looking for funding by habryka
Jun 14, 2023
LW - MetaAI: less is less for alignment. by Cleo Nardo
Jun 14, 2023
LW - My guess for why I was wrong about US housing by romeostevensit
Jun 14, 2023
LW - <$750k grants for General Purpose AI Assurance/Safety Research by Phosphorous
Jun 13, 2023
LW - TASRA: A Taxonomy and Analysis of Societal-Scale Risks from AI by Andrew Critch
Jun 13, 2023
LW - The Dial of Progress by Zvi
Jun 13, 2023
LW - UFO Betting: Put Up or Shut Up by RatsWrongAboutUAP
Jun 13, 2023
LW - If you are too stressed, walk away from the front lines by Neil Warren
Jun 13, 2023
LW - Introduction to Towards Causal Foundations of Safe AGI by tom4everitt
Jun 13, 2023
LW - UK PM: $125M for AI safety by Hauke Hillebrandt
Jun 12, 2023
LW - Manifold Predicted the AI Extinction Statement and CAIS Wanted it Deleted by David Chee
Jun 12, 2023
LW - ARC is hiring theoretical researchers by paulfchristiano
Jun 12, 2023
LW - Critiques of prominent AI safety labs: Conjecture by Omega.
Jun 12, 2023
LW - [Linkpost] Large Language Models Converge on Brain-Like Word Representations by Bogdan Ionut Cirstea
Jun 12, 2023
LW - I can see how I am Dumb by Johannes C. Mayer
Jun 11, 2023
LW - Inference-Time Intervention: Eliciting Truthful Answers from a Language Model by likenneth
Jun 11, 2023
LW - The Dictatorship Problem by alyssavance
Jun 11, 2023
LW - Ethodynamics of Omelas by dr s
Jun 11, 2023
LW - How biosafety could inform AI standards by Olivia Jimenez
Jun 09, 2023
LW - A plea for solutionism on AI safety by jasoncrawford
Jun 09, 2023
LW - InternLM - China's Best (Unverified) by Lao Mein
Jun 09, 2023
LW - Updates and Reflections on Optimal Exercise after Nearly a Decade by romeostevensit
Jun 09, 2023
LW - Takeaways from the Mechanistic Interpretability Challenges by scasper
Jun 08, 2023
LW - A moral backlash against AI will probably slow down AGI development by geoffreymiller
Jun 08, 2023
LW - AI #15: The Principle of Charity by Zvi
Jun 08, 2023
LW - An Exercise to Build Intuitions on AGI Risk by Lauro Langosco
Jun 08, 2023
LW - LEAst-squares Concept Erasure (LEACE) by tricky labyrinth
Jun 08, 2023
LW - What will GPT-2030 look like? by jsteinhardt
Jun 08, 2023
LW - Elon talked with senior Chinese leadership about AI X-risk by ChristianKl
Jun 07, 2023
LW - Why I am not a longtermist (May 2022) by boazbarak
Jun 07, 2023
LW - Transformative AGI by 2043 is <1% likely by Ted Sanders
Jun 07, 2023
LW - Cultivate an obsession with the object level (the world is fascinating) by Richard Ngo
Jun 07, 2023
LW - The Sharp Right Turn: sudden deceptive alignment as a convergent goal by avturchin
Jun 07, 2023
LW - Launching Lightspeed Grants (Apply by July 6th) by habryka
Jun 07, 2023
LW - Society Library seeking contributions for canonical AI Safety debate map by Jarred Filmer
Jun 07, 2023
LW - The Base Rate Times, news through prediction markets by vandemonian
Jun 06, 2023
LW - Andrew Ng wants to have a conversation about extinction risk from AI by Leon Lang
Jun 06, 2023
LW - A Playbook for AI Risk Reduction (focused on misaligned AI) by HoldenKarnofsky
Jun 06, 2023
LW - Intelligence Officials Say U.S. Has Retrieved Craft of Non-Human Origin by lc
Jun 06, 2023
LW - One implementation of regulatory GPU restrictions by porby
Jun 06, 2023
LW - How to Think About Activation Patching by Neel Nanda
Jun 06, 2023
LW - The (local) unit of intelligence is FLOPs by boazbarak
Jun 06, 2023
LW - AISafety.info "How can I help?" FAQ by steven0461
Jun 06, 2023
LW - Algorithmic Improvement Is Probably Faster Than Scaling Now by johnswentworth
Jun 06, 2023
LW - We Are Less Wrong than E. T. Jaynes on Loss Functions in Human Society by Zack M Davis
Jun 05, 2023
LW - Meta-conversation shouldn't be taboo by Adam Zerner
Jun 05, 2023
LW - Wildfire of strategicness by TsviBT
Jun 05, 2023
LW - Ages Survey: Results by jefftk
Jun 05, 2023
LW - The ants and the grasshopper by Richard Ngo
Jun 04, 2023
LW - Announcing AISafety.info's Write-a-thon (June 16-18) and Second Distillation Fellowship (July 3-October 2) by steven0461
Jun 04, 2023
LW - The Control Problem: Unsolved or Unsolvable? by Remmelt
Jun 04, 2023
LW - A Double-Feature on The Extropians by Maxwell Tabarrok
Jun 03, 2023
LW - The AGI Race Between the US and China Doesn’t Exist. by Eva B
Jun 03, 2023
LW - A mind needn't be curious to reap the benefits of curiosity by So8res
Jun 02, 2023
LW - Co-found an incubator for independent AI Safety researchers by Alexandra Bos
Jun 02, 2023
LW - Dreams of "Mathopedia" by NicholasKross
Jun 02, 2023
LW - Think carefully before calling RL policies "agents" by TurnTrout
Jun 02, 2023
LW - Uncertainty about the future does not imply that AGI will go well by Lauro Langosco
Jun 02, 2023
LW - Yudkowsky vs Hanson on FOOM: Whose Predictions Were Better? by 1a3orn
Jun 02, 2023
LW - Things I Learned by Spending Five Thousand Hours In Non-EA Charities by jenn
Jun 01, 2023
LW - Change my mind: Veganism entails trade-offs, and health is one of the axes by Elizabeth
Jun 01, 2023
LW - Short Remark on the (subjective) mathematical 'naturalness' of the Nanda--Lieberum addition modulo 113 algorithm by Spencer Becker-Kahn
Jun 01, 2023
LW - Work dumber not smarter by lukehmiles
Jun 01, 2023
LW - The challenge of articulating tacit knowledge by NinaR
Jun 01, 2023
LW - The Crux List by Zvi
May 31, 2023
LW - To Predict What Happens, Ask What Happens by Zvi
May 31, 2023
LW - Cosmopolitan values don't come free by So8res
May 31, 2023
LW - Contrast Pairs Drive the Empirical Performance of Contrast Consistent Search (CCS) by Scott Emmons
May 31, 2023
LW - Product Endorsement: Food for sleep interruptions by Elizabeth
May 31, 2023
LW - The case for removing alignment and ML research from the training dataset by beren
May 31, 2023
LW - PaLM-2 and GPT-4 in "Extrapolating GPT-N performance" by Lukas Finnveden
May 30, 2023
LW - Announcing Apollo Research by Marius Hobbhahn
May 30, 2023
LW - Reply to a fertility doctor concerning polygenic embryo screening by GeneSmith
May 30, 2023
LW - Sentience matters by So8res
May 29, 2023
LW - Wikipedia as an introduction to the alignment problem by SoerenMind
May 29, 2023
LW - Gemini will bring the next big timeline update by p.b.
May 29, 2023
LW - Kelly betting vs expectation maximization by MorgneticField
May 29, 2023
LW - Morality is Accidental and Self-Congratulatory by ymeskhout
May 29, 2023
LW - TinyStories: Small Language Models That Still Speak Coherent English by Ulisse Mini
May 29, 2023
LW - DandD.Sci 5E: Return of the League of Defenders by aphyer
May 28, 2023
LW - You now can enable reacts on your own posts! (and other react info) by Ruby
May 28, 2023
LW - Self-administered EMDR without a therapist is very useful for a lot of things! by Anton Rodenhauser
May 27, 2023
LW - Should Rational Animations invite viewers to read content on LessWrong? by Writer
May 27, 2023
LW - Why I'm Not (Yet) A Full-Time Technical Alignment Researcher by NicholasKross
May 26, 2023
LW - Request: stop advancing AI capabilities by So8res
May 26, 2023
LW - Book Review: How Minds Change by bc4026bd4aaa5b7fe
May 26, 2023
LW - Mob and Bailey by Screwtape
May 26, 2023
LW - [Linkpost] Interpretability Dreams by DanielFilan
May 25, 2023
LW - Look At What's In Front Of You (Conclusion to The Nuts and Bolts of Naturalism) by LoganStrohl
May 25, 2023
LW - Solving the Mechanistic Interpretability challenges: EIS VII Challenge 2 by StefanHex
May 25, 2023
LW - DeepMind: Model evaluation for extreme risks by Zach Stein-Perlman
May 25, 2023
LW - Adumbrations on AGI from an outsider by nicholashalden
May 25, 2023