Listen to a podcast, please open Podcast Republic app. Available on Google Play Store and Apple App Store.
Episode | Date |
---|---|
LW - Prompts for Big-Picture Planning by Raemon
|
Apr 14, 2024 |
LW - What convincing warning shot could help prevent extinction from AI? by Charbel-Raphaël
|
Apr 13, 2024 |
LW - Things Solenoid Narrates by Solenoid Entity
|
Apr 13, 2024 |
LW - Carl Sagan, nuking the moon, and not nuking the moon by eukaryote
|
Apr 13, 2024 |
LW - MIRI's April 2024 Newsletter by Harlan
|
Apr 13, 2024 |
LW - UDT1.01: Plannable and Unplanned Observations (3/10) by Diffractor
|
Apr 12, 2024 |
LW - Generalized Stat Mech: The Boltzmann Approach by David Lorell
|
Apr 12, 2024 |
LW - A DandD.Sci Dodecalogue by abstractapplic
|
Apr 12, 2024 |
LW - Announcing Atlas Computing by miyazono
|
Apr 12, 2024 |
LW - DandD.Sci: The Mad Tyrant's Pet Turtles [Evaluation and Ruleset] by abstractapplic
|
Apr 10, 2024 |
LW - RTFB: On the New Proposed CAIP AI Bill by Zvi
|
Apr 10, 2024 |
LW - Ophiology (or, how the Mamba architecture works) by Danielle Ensign
|
Apr 09, 2024 |
LW - Conflict in Posthuman Literature by Martín Soto
|
Apr 09, 2024 |
LW - Medical Roundup #2 by Zvi
|
Apr 09, 2024 |
LW - Math-to-English Cheat Sheet by nahoj
|
Apr 09, 2024 |
LW - Gated Attention Blocks: Preliminary Progress toward Removing Attention Head Superposition by cmathw
|
Apr 09, 2024 |
LW - on the dollar-yen exchange rate by bhauth
|
Apr 08, 2024 |
LW - How We Picture Bayesian Agents by johnswentworth
|
Apr 08, 2024 |
LW - A Dozen Ways to Get More Dakka by Davidmanheim
|
Apr 08, 2024 |
LW - My intellectual journey to (dis)solve the hard problem of consciousness by Charbel-Raphaël
|
Apr 07, 2024 |
LW - "Fractal Strategy" workshop report by Raemon
|
Apr 07, 2024 |
LW - The 2nd Demographic Transition by Maxwell Tabarrok
|
Apr 07, 2024 |
LW - On Complexity Science by Garrett Baker
|
Apr 05, 2024 |
LW - Partial value takeover without world takeover by KatjaGrace
|
Apr 05, 2024 |
LW - New report: A review of the empirical evidence for existential risk from AI via misaligned power-seeking by Harlan
|
Apr 05, 2024 |
LW - AI #58: Stargate AGI by Zvi
|
Apr 05, 2024 |
LW - LLMs for Alignment Research: a safety priority? by abramdemski
|
Apr 04, 2024 |
LW - A gentle introduction to mechanistic anomaly detection by Erik Jenner
|
Apr 04, 2024 |
LW - What's with all the bans recently? by Gerald Monroe
|
Apr 04, 2024 |
LW - Best in Class Life Improvement by sapphire
|
Apr 04, 2024 |
LW - Notes on Dwarkesh Patel's Podcast with Sholto Douglas and Trenton Bricken by Zvi
|
Apr 02, 2024 |
LW - LessWrong: After Dark, a new side of LessWrong by So8res
|
Apr 02, 2024 |
LW - Gradient Descent on the Human Brain by Jozdien
|
Apr 02, 2024 |
LW - Coherence of Caches and Agents by johnswentworth
|
Apr 02, 2024 |
LW - Announcing Suffering For Good by Garrett Baker
|
Apr 01, 2024 |
LW - OMMC Announces RIP by Adam Scholl
|
Apr 01, 2024 |
LW - The Evolution of Humans Was Net-Negative for Human Values by Zack M Davis
|
Apr 01, 2024 |
LW - So You Created a Sociopath - New Book Announcement! by Garrett Baker
|
Apr 01, 2024 |
LW - A Selection of Randomly Selected SAE Features by CallumMcDougall
|
Apr 01, 2024 |
LW - Apply to be a Safety Engineer at Lockheed Martin! by yanni
|
Apr 01, 2024 |
LW - The Story of "I Have Been A Good Bing" by habryka
|
Apr 01, 2024 |
LW - The Best Tacit Knowledge Videos on Every Subject by Parker Conley
|
Mar 31, 2024 |
LW - SAE-VIS: Announcement Post by CallumMcDougall
|
Mar 31, 2024 |
LW - My simple AGI investment and insurance strategy by lc
|
Mar 31, 2024 |
LW - Back to Basics: Truth is Unitary by lsusr
|
Mar 30, 2024 |
LW - DandD.Sci: The Mad Tyrant's Pet Turtles by abstractapplic
|
Mar 30, 2024 |
LW - SAE reconstruction errors are (empirically) pathological by wesg
|
Mar 29, 2024 |
LW - How to safely use an optimizer by Simon Fischer
|
Mar 29, 2024 |
LW - [Linkpost] Practically-A-Book Review: Rootclaim $100,000 Lab Leak Debate by trevor
|
Mar 28, 2024 |
LW - Was Releasing Claude-3 Net-Negative? by Logan Riggs
|
Mar 28, 2024 |
LW - The Cognitive-Theoretic Model of the Universe: A Partial Summary and Review by jessicata
|
Mar 27, 2024 |
LW - Daniel Kahneman has died by DanielFilan
|
Mar 27, 2024 |
LW - AE Studio @ SXSW: We need more AI consciousness research (and further resources) by AE Studio
|
Mar 27, 2024 |
LW - Failures in Kindness by silentbob
|
Mar 27, 2024 |
LW - Modern Transformers are AGI, and Human-Level by abramdemski
|
Mar 26, 2024 |
LW - My Interview With Cade Metz on His Reporting About Slate Star Codex by Zack M Davis
|
Mar 26, 2024 |
LW - Should rationalists be spiritual / Spirituality as overcoming delusion by Kaj Sotala
|
Mar 26, 2024 |
LW - LessOnline (May 31 - June 2, Berkeley, CA) by Ben Pace
|
Mar 26, 2024 |
LW - On attunement by Joe Carlsmith
|
Mar 25, 2024 |
LW - My Detailed Notes and Commentary from Secular Solstice by Jeffrey Heninger
|
Mar 25, 2024 |
LW - On Lex Fridman's Second Podcast with Altman by Zvi
|
Mar 25, 2024 |
LW - Do not delete your misaligned AGI. by mako yass
|
Mar 25, 2024 |
LW - All About Concave and Convex Agents by mako yass
|
Mar 25, 2024 |
LW - Vipassana Meditation and Active Inference: A Framework for Understanding Suffering and its Cessation by Benjamin Sturgeon
|
Mar 24, 2024 |
LW - General Thoughts on Secular Solstice by Jeffrey Heninger
|
Mar 24, 2024 |
LW - A Teacher vs. Everyone Else by ronak69
|
Mar 23, 2024 |
LW - AI #56: Blackwell That Ends Well by Zvi
|
Mar 23, 2024 |
LW - ChatGPT can learn indirect control by Raymond D
|
Mar 22, 2024 |
LW - Vernor Vinge, who coined the term "Technological Singularity", dies at 79 by Kaj Sotala
|
Mar 21, 2024 |
LW - On green by Joe Carlsmith
|
Mar 21, 2024 |
LW - "Deep Learning" Is Function Approximation by Zack M Davis
|
Mar 21, 2024 |
LW - DeepMind: Evaluating Frontier Models for Dangerous Capabilities by Zach Stein-Perlman
|
Mar 21, 2024 |
LW - On the Gladstone Report by Zvi
|
Mar 21, 2024 |
LW - Stagewise Development in Neural Networks by Jesse Hoogland
|
Mar 20, 2024 |
LW - Monthly Roundup #16: March 2024 by Zvi
|
Mar 20, 2024 |
LW - Natural Latents: The Concepts by johnswentworth
|
Mar 20, 2024 |
LW - New report: Safety Cases for AI by joshc
|
Mar 20, 2024 |
LW - Increasing IQ by 10 Points is Possible by George3d6
|
Mar 19, 2024 |
LW - Inferring the model dimension of API-protected LLMs by Ege Erdil
|
Mar 19, 2024 |
LW - Experimentation (Part 7 of "The Sense Of Physical Necessity") by LoganStrohl
|
Mar 19, 2024 |
LW - Neuroscience and Alignment by Garrett Baker
|
Mar 19, 2024 |
LW - Toki pona FAQ by dkl9
|
Mar 19, 2024 |
LW - 5 Physics Problems by DaemonicSigil
|
Mar 18, 2024 |
LW - Measuring Coherence of Policies in Toy Environments by dx26
|
Mar 18, 2024 |
LW - Community Notes by X by NicholasKees
|
Mar 18, 2024 |
LW - On Devin by Zvi
|
Mar 18, 2024 |
LW - The Worst Form Of Government (Except For Everything Else We've Tried) by johnswentworth
|
Mar 17, 2024 |
LW - Anxiety vs. Depression by Sable
|
Mar 17, 2024 |
LW - My PhD thesis: Algorithmic Bayesian Epistemology by Eric Neyman
|
Mar 17, 2024 |
LW - Rational Animations offers animation production and writing services! by Writer
|
Mar 16, 2024 |
LW - Introducing METR's Autonomy Evaluation Resources by Megan Kinniment
|
Mar 16, 2024 |
LW - Constructive Cauchy sequences vs. Dedekind cuts by jessicata
|
Mar 15, 2024 |
LW - Conditional on Getting to Trade, Your Trade Wasn't All That Great by Ricki Heicklen
|
Mar 14, 2024 |
LW - Highlights from Lex Fridman's interview of Yann LeCun by Joel Burget
|
Mar 14, 2024 |
LW - How useful is "AI Control" as a framing on AI X-Risk? by habryka
|
Mar 14, 2024 |
LW - AI #55: Keep Clauding Along by Zvi
|
Mar 14, 2024 |
LW - Jobs, Relationships, and Other Cults by Ruby
|
Mar 14, 2024 |
LW - 'Empiricism!' as Anti-Epistemology by Eliezer Yudkowsky
|
Mar 14, 2024 |
LW - On the Latest TikTok Bill by Zvi
|
Mar 13, 2024 |
LW - What could a policy banning AGI look like? by TsviBT
|
Mar 13, 2024 |
LW - The Parable Of The Fallen Pendulum - Part 2 by johnswentworth
|
Mar 13, 2024 |
LW - Superforecasting the Origins of the Covid-19 Pandemic by DanielFilan
|
Mar 13, 2024 |
LW - OpenAI: The Board Expands by Zvi
|
Mar 12, 2024 |
LW - Be More Katja by Nathan Young
|
Mar 12, 2024 |
LW - "Artificial General Intelligence": an extremely brief FAQ by Steven Byrnes
|
Mar 11, 2024 |
LW - Some (problematic) aesthetics of what constitutes good work in academia by Steven Byrnes
|
Mar 11, 2024 |
LW - What do we know about the AI knowledge and views, especially about existential risk, of the new OpenAI board members? by Zvi
|
Mar 11, 2024 |
LW - Twelve Lawsuits against OpenAI by Remmelt
|
Mar 11, 2024 |
LW - "How could I have thought that faster?" by mesaoptimizer
|
Mar 11, 2024 |
LW - Simple versus Short: Higher-order degeneracy and error-correction by Daniel Murfet
|
Mar 11, 2024 |
LW - One-shot strategy games? by Raemon
|
Mar 11, 2024 |
LW - Notes from a Prompt Factory by Richard Ngo
|
Mar 10, 2024 |
LW - Closeness To the Issue (Part 5 of "The Sense Of Physical Necessity") by LoganStrohl
|
Mar 09, 2024 |
LW - Lies and disrespect from the EA Infrastructure Fund by Igor Ivanov
|
Mar 08, 2024 |
LW - Woods' new preprint on object permanence by Steven Byrnes
|
Mar 08, 2024 |
LW - AI #54: Clauding Along by Zvi
|
Mar 08, 2024 |
LW - MATS AI Safety Strategy Curriculum by Ryan Kidd
|
Mar 08, 2024 |
LW - Simple Kelly betting in prediction markets by jessicata
|
Mar 07, 2024 |
LW - Mud and Despair (Part 4 of "The Sense Of Physical Necessity") by LoganStrohl
|
Mar 07, 2024 |
LW - Social status part 1/2: negotiations over object-level preferences by Steven Byrnes
|
Mar 07, 2024 |
LW - Movie posters by KatjaGrace
|
Mar 07, 2024 |
LW - On Claude 3.0 by Zvi
|
Mar 06, 2024 |
LW - Vote on Anthropic Topics to Discuss by Ben Pace
|
Mar 06, 2024 |
LW - Using axis lines for good or evil by dynomight
|
Mar 06, 2024 |
LW - My Clients, The Liars by ymeskhout
|
Mar 05, 2024 |
LW - Research Report: Sparse Autoencoders find only 9/180 board state features in OthelloGPT by Robert AIZI
|
Mar 05, 2024 |
LW - Read the Roon by Zvi
|
Mar 05, 2024 |
LW - Claude 3 claims it's conscious, doesn't want to die or be modified by Mikhail Samin
|
Mar 05, 2024 |
LW - Housing Roundup #7 by Zvi
|
Mar 05, 2024 |
LW - The Broken Screwdriver and other parables by bhauth
|
Mar 04, 2024 |
LW - Are we so good to simulate? by KatjaGrace
|
Mar 04, 2024 |
LW - Self-Resolving Prediction Markets by PeterMcCluskey
|
Mar 04, 2024 |
LW - Grief is a fire sale by Nathan Young
|
Mar 04, 2024 |
LW - Agreeing With Stalin in Ways That Exhibit Generally Rationalist Principles by Zack M Davis
|
Mar 03, 2024 |
LW - The Defence production act and AI policy by NathanBarnard
|
Mar 02, 2024 |
LW - If you weren't such an idiot... by kave
|
Mar 02, 2024 |
LW - The World in 2029 by Nathan Young
|
Mar 02, 2024 |
LW - Increasing IQ is trivial by George3d6
|
Mar 02, 2024 |
LW - Notes on Dwarkesh Patel's Podcast with Demis Hassabis by Zvi
|
Mar 02, 2024 |
LW - Elon files grave charges against OpenAI by mako yass
|
Mar 01, 2024 |
LW - Locating My Eyes (Part 3 of "The Sense of Physical Necessity") by LoganStrohl
|
Mar 01, 2024 |
LW - The Parable Of The Fallen Pendulum - Part 1 by johnswentworth
|
Mar 01, 2024 |
LW - AI #53: One More Leap by Zvi
|
Feb 29, 2024 |
LW - Bengio's Alignment Proposal: "Towards a Cautious Scientist AI with Convergent Safety Bounds" by mattmacdermott
|
Feb 29, 2024 |
LW - Timaeus's First Four Months by Jesse Hoogland
|
Feb 28, 2024 |
LW - How I internalized my achievements to better deal with negative feelings by Raymond Koopmanschap
|
Feb 28, 2024 |
LW - The Gemini Incident Continues by Zvi
|
Feb 28, 2024 |
LW - Announcing 'The LeastWrong' and review winner post pages by kave
|
Feb 28, 2024 |
LW - Examining Language Model Performance with Reconstructed Activations using Sparse Autoencoders by Evan Anders
|
Feb 27, 2024 |
LW - How I build and run behavioral interviews by benkuhn
|
Feb 27, 2024 |
LW - Can an AI do our alignment homework for us? by Chris Leong
|
Feb 26, 2024 |
LW - Ideological Bayesians by Kevin Dorst
|
Feb 26, 2024 |
LW - China-AI forecasts by NathanBarnard
|
Feb 26, 2024 |
LW - "In-Context" "Learning" by Arjun Panickssery
|
Feb 25, 2024 |
LW - A starting point for making sense of task structure (in machine learning) by Kaarel
|
Feb 25, 2024 |
LW - We Need Major, But Not Radical, FDA Reform by Maxwell Tabarrok
|
Feb 25, 2024 |
LW - Deep and obvious points in the gap between your thoughts and your pictures of thought by KatjaGrace
|
Feb 25, 2024 |
LW - How well do truth probes generalise? by mishajw
|
Feb 25, 2024 |
LW - Choosing My Quest (Part 2 of "The Sense Of Physical Necessity") by LoganStrohl
|
Feb 24, 2024 |
LW - Rationality Research Report: Towards 10x OODA Looping? by Raemon
|
Feb 24, 2024 |
LW - Balancing Games by jefftk
|
Feb 24, 2024 |
LW - The Sense Of Physical Necessity: A Naturalism Demo (Introduction) by LoganStrohl
|
Feb 24, 2024 |
LW - The Shutdown Problem: Incomplete Preferences as a Solution by EJT
|
Feb 23, 2024 |
LW - The Byronic Hero Always Loses by Cole Wyeth
|
Feb 23, 2024 |
LW - Everything Wrong with Roko's Claims about an Engineered Pandemic by EZ97
|
Feb 23, 2024 |
LW - Gemini Has a Problem by Zvi
|
Feb 23, 2024 |
LW - AI #52: Oops by Zvi
|
Feb 23, 2024 |
LW - Research Post: Tasks That Language Models Don't Learn by Bruce W. Lee
|
Feb 23, 2024 |
LW - Sora What by Zvi
|
Feb 23, 2024 |
LW - Contra Ngo et al. "Every 'Every Bay Area House Party' Bay Area House Party" by Ricki Heicklen
|
Feb 23, 2024 |
LW - Do sparse autoencoders find "true features"? by Demian Till
|
Feb 22, 2024 |
LW - Job Listing: Managing Editor / Writer by Gretta Duleba
|
Feb 22, 2024 |
LW - The One and a Half Gemini by Zvi
|
Feb 22, 2024 |
LW - The Pareto Best and the Curse of Doom by Screwtape
|
Feb 22, 2024 |
LW - Dual Wielding Kindle Scribes by mesaoptimizer
|
Feb 21, 2024 |
LW - Why does generalization work? by Martín Soto
|
Feb 21, 2024 |
LW - Less Wrong automated systems are inadvertently Censoring me by Roko
|
Feb 21, 2024 |
LW - AI #51: Altman's Ambition by Zvi
|
Feb 21, 2024 |
LW - I'd also take $7 trillion by bhauth
|
Feb 19, 2024 |
LW - Retirement Accounts and Short Timelines by jefftk
|
Feb 19, 2024 |
LW - On coincidences and Bayesian reasoning, as applied to the origins of COVID-19 by viking math
|
Feb 19, 2024 |
LW - Things I've Grieved by Raemon
|
Feb 18, 2024 |
LW - Self-Awareness: Taxonomy and eval suite proposal by Daniel Kokotajlo
|
Feb 18, 2024 |
LW - The Pointer Resolution Problem by Jozdien
|
Feb 17, 2024 |
LW - 2023 Survey Results by Screwtape
|
Feb 16, 2024 |
LW - Fixing Feature Suppression in SAEs by Benjamin Wright
|
Feb 16, 2024 |
LW - OpenAI's Sora is an agent by CBiddulph
|
Feb 16, 2024 |
LW - Every "Every Bay Area House Party" Bay Area House Party by Richard Ngo
|
Feb 16, 2024 |
LW - Raising children on the eve of AI by juliawise
|
Feb 15, 2024 |
LW - On the Proposed California SB 1047 by Zvi
|
Feb 14, 2024 |
LW - CFAR Takeaways: Andrew Critch by Raemon
|
Feb 14, 2024 |
LW - Masterpiece by Richard Ngo
|
Feb 14, 2024 |
LW - Where is the Town Square? by Gretta Duleba
|
Feb 13, 2024 |
LW - Tort Law Can Play an Important Role in Mitigating AI Risk by Gabriel Weil
|
Feb 13, 2024 |
LW - Lsusr's Rationality Dojo by lsusr
|
Feb 13, 2024 |
LW - AI #50: The Most Dangerous Thing by Zvi
|
Feb 08, 2024 |
LW - A Chess-GPT Linear Emergent World Representation by karvonenadam
|
Feb 08, 2024 |
LW - Believing In by AnnaSalamon
|
Feb 08, 2024 |
LW - Conditional prediction markets are evidential, not causal by philh
|
Feb 08, 2024 |
LW - More Hyphenation by Arjun Panickssery
|
Feb 07, 2024 |
LW - story-based decision-making by bhauth
|
Feb 07, 2024 |
LW - Why I think it's net harmful to do technical safety research at AGI labs by Remmelt
|
Feb 07, 2024 |
LW - what does davidad want from "boundaries"? by Chipmonk
|
Feb 07, 2024 |
LW - My guess at Conjecture's vision: triggering a narrative bifurcation by Alexandre Variengien
|
Feb 06, 2024 |
LW - Fluent dreaming for language models (AI interpretability method) by tbenthompson
|
Feb 06, 2024 |
LW - Preventing model exfiltration with upload limits by ryan greenblatt
|
Feb 06, 2024 |
LW - Things You're Allowed to Do: University Edition by Saul Munn
|
Feb 06, 2024 |
LW - Implementing activation steering by Annah
|
Feb 05, 2024 |
LW - Noticing Panic by Cole Wyeth
|
Feb 05, 2024 |
LW - On Dwarkesh's 3rd Podcast With Tyler Cowen by Zvi
|
Feb 04, 2024 |
LW - Theories of Applied Rationality by Camille Berger
|
Feb 04, 2024 |
LW - Why I no longer identify as transhumanist by Kaj Sotala
|
Feb 04, 2024 |
LW - Brute Force Manufactured Consensus is Hiding the Crime of the Century by Roko
|
Feb 03, 2024 |
LW - Announcing the London Initiative for Safe AI (LISA) by James Fox
|
Feb 03, 2024 |
LW - Survey for alignment researchers: help us build better field-level models by Cameron Berg
|
Feb 03, 2024 |
LW - Soft Prompts for Evaluation: Measuring Conditional Distance of Capabilities by porby
|
Feb 03, 2024 |
LW - Most experts believe COVID-19 was probably not a lab leak by DanielFilan
|
Feb 02, 2024 |
LW - On Not Requiring Vaccination by jefftk
|
Feb 02, 2024 |
LW - Open Source Sparse Autoencoders for all Residual Stream Layers of GPT2-Small by Joseph Bloom
|
Feb 02, 2024 |
LW - Wrong answer bias by lukehmiles
|
Feb 02, 2024 |
LW - Davidad's Provably Safe AI Architecture - ARIA's Programme Thesis by simeon c
|
Feb 02, 2024 |
LW - Ten Modes of Culture War Discourse by jchan
|
Feb 01, 2024 |
LW - Simple distribution approximation: When sampled 100 times, can language models yield 80% A and 20% B? by Teun van der Weij
|
Feb 01, 2024 |
LW - Per protocol analysis as medical malpractice by braces
|
Feb 01, 2024 |
LW - Leading The Parade by johnswentworth
|
Jan 31, 2024 |
LW - Without Fundamental Advances, Rebellion and Coup d'État are the Inevitable Outcomes of Dictators and Monarchs Trying to Control Large, Capable Countries by Roko
|
Jan 31, 2024 |
LW - Explaining Impact Markets by Saul Munn
|
Jan 31, 2024 |
LW - Childhood and Education Roundup #4 by Zvi
|
Jan 31, 2024 |
LW - on neodymium magnets by bhauth
|
Jan 31, 2024 |
LW - Win Friends and Influence People Ch. 2: The Bombshell by gull
|
Jan 30, 2024 |
LW - Things You're Allowed to Do: At the Dentist by rbinnn
|
Jan 30, 2024 |
LW - Processor clock speeds are not how fast AIs think by Ege Erdil
|
Jan 29, 2024 |
LW - Why I take short timelines seriously by NicholasKees
|
Jan 29, 2024 |
LW - Palworld development blog post by bhauth
|
Jan 28, 2024 |
LW - Epistemic Hell by rogersbacon
|
Jan 28, 2024 |
LW - Don't sleep on Coordination Takeoffs by trevor
|
Jan 28, 2024 |
LW - Aligned AI is dual use technology by lc
|
Jan 27, 2024 |
LW - The Good Balsamic Vinegar by jenn
|
Jan 27, 2024 |
LW - Surgery Works Well Without The FDA by Maxwell Tabarrok
|
Jan 27, 2024 |
LW - Making every researcher seek grants is a broken model by jasoncrawford
|
Jan 26, 2024 |
LW - "Does your paradigm beget new, good, paradigms?" by Raemon
|
Jan 26, 2024 |
LW - AI #48: The Talk of Davos by Zvi
|
Jan 26, 2024 |
LW - Is a random box of gas predictable after 20 seconds? by Thomas Kwa
|
Jan 26, 2024 |
LW - [Repost] The Copenhagen Interpretation of Ethics by mesaoptimizer
|
Jan 26, 2024 |
LW - RAND report finds no effect of current LLMs on viability of bioterrorism attacks by StellaAthena
|
Jan 25, 2024 |
LW - Will quantum randomness affect the 2028 election? by Thomas Kwa
|
Jan 25, 2024 |
LW - Humans aren't fleeb. by Charlie Steiner
|
Jan 24, 2024 |
LW - This might be the last AI Safety Camp by Remmelt
|
Jan 24, 2024 |
LW - the subreddit size threshold by bhauth
|
Jan 24, 2024 |
LW - Making a Secular Solstice Songbook by jefftk
|
Jan 24, 2024 |
LW - Loneliness and suicide mitigation for students using GPT3-enabled chatbots (survey of Replika users in Nature) by Kaj Sotala
|
Jan 24, 2024 |
LW - legged robot scaling laws by bhauth
|
Jan 22, 2024 |
LW - On "Geeks, MOPs, and Sociopaths" by alkjash
|
Jan 22, 2024 |
LW - Book review: Cuisine and Empire by eukaryote
|
Jan 22, 2024 |
LW - When Does Altruism Strengthen Altruism? by jefftk
|
Jan 22, 2024 |
LW - A quick investigation of AI pro-AI bias by Fabien Roger
|
Jan 20, 2024 |
LW - What rationality failure modes are there? by Ulisse Mini
|
Jan 19, 2024 |
LW - There is way too much serendipity by Malmesbury
|
Jan 19, 2024 |
LW - Logical Line-Of-Sight Makes Games Sequential or Loopy by StrivingForLegibility
|
Jan 19, 2024 |
LW - Does literacy remove your ability to be a bard as good as Homer? by Adrià Garriga-alonso
|
Jan 19, 2024 |
LW - The True Story of How GPT-2 Became Maximally Lewd by Writer
|
Jan 19, 2024 |
LW - On the abolition of man by Joe Carlsmith
|
Jan 18, 2024 |
LW - On Anthropic's Sleeper Agents Paper by Zvi
|
Jan 17, 2024 |
LW - AlphaGeometry: An Olympiad-level AI system for geometry by alyssavance
|
Jan 17, 2024 |
LW - An Introduction To The Mandelbrot Set That Doesn't Mention Complex Numbers by Yitz
|
Jan 17, 2024 |
LW - Medical Roundup #1 by Zvi
|
Jan 17, 2024 |
LW - Why wasn't preservation with the goal of potential future revival started earlier in history? by Andy McKenzie
|
Jan 17, 2024 |
LW - Being nicer than Clippy by Joe Carlsmith
|
Jan 17, 2024 |
LW - The impossible problem of due process by mingyuan
|
Jan 16, 2024 |
LW - The case for training frontier AIs on Sumerian-only corpus by Alexandre Variengien
|
Jan 15, 2024 |
LW - DandD.Sci(-fi): Colonizing the SuperHyperSphere by abstractapplic
|
Jan 15, 2024 |
LW - Gender Exploration by sapphire
|
Jan 14, 2024 |
LW - Notice When People Are Directionally Correct by Chris Leong
|
Jan 14, 2024 |
LW - Against most AI risk analogies by Matthew Barnett
|
Jan 14, 2024 |
LW - Land Reclamation is in the 9th Circle of Stagnation Hell by Maxwell Tabarrok
|
Jan 13, 2024 |
LW - What good is G-factor if you're dumped in the woods? A field report from a camp counselor. by Hastings
|
Jan 12, 2024 |
LW - An Actually Intuitive Explanation of the Oberth Effect by Isaac King
|
Jan 12, 2024 |
LW - Introduce a Speed Maximum by jefftk
|
Jan 12, 2024 |
LW - An even deeper atheism by Joe Carlsmith
|
Jan 11, 2024 |
LW - The Perceptron Controversy by Yuxi Liu
|
Jan 11, 2024 |
LW - Universal Love Integration Test: Hitler by Raemon
|
Jan 11, 2024 |
LW - The Aspiring Rationalist Congregation by maia
|
Jan 11, 2024 |
LW - Does AI risk "other" the AIs? by Joe Carlsmith
|
Jan 10, 2024 |
LW - Saving the world sucks by Defective Altruism
|
Jan 10, 2024 |
LW - On the Contrary, Steelmanning Is Normal; ITT-Passing Is Niche by Zack M Davis
|
Jan 10, 2024 |
LW - Goodbye, Shoggoth: The Stage, its Animatronics, and the Puppeteer - a New Metaphor by RogerDearnaley
|
Jan 10, 2024 |
LW - Learning Math in Time for Alignment by NicholasKross
|
Jan 09, 2024 |
LW - A model of research skill by L Rudolf L
|
Jan 09, 2024 |
LW - When "yang" goes wrong by Joe Carlsmith
|
Jan 08, 2024 |
LW - 2023 Prediction Evaluations by Zvi
|
Jan 08, 2024 |
LW - Bayesians Commit the Gambler's Fallacy by Kevin Dorst
|
Jan 07, 2024 |
LW - Defending against hypothetical moon life during Apollo 11 by eukaryote
|
Jan 07, 2024 |
LW - AI Risk and the US Presidential Candidates by Zane
|
Jan 07, 2024 |
LW - The Next ChatGPT Moment: AI Avatars by kolmplex
|
Jan 06, 2024 |
LW - Survey of 2,778 AI authors: six parts in pictures by KatjaGrace
|
Jan 06, 2024 |
LW - Almost everyone I've met would be well-served thinking more about what to focus on by Henrik Karlsson
|
Jan 06, 2024 |
LW - AI #45: To Be Determined by Zvi
|
Jan 05, 2024 |
LW - MIRI 2024 Mission and Strategy Update by Malo
|
Jan 05, 2024 |
LW - Deep atheism and AI risk by Joe Carlsmith
|
Jan 04, 2024 |
LW - Copyright Confrontation #1 by Zvi
|
Jan 04, 2024 |
LW - Some Vacation Photos by johnswentworth
|
Jan 04, 2024 |
LW - Safety First: safety before full alignment. The deontic sufficiency hypothesis. by Chipmonk
|
Jan 04, 2024 |
LW - Trading off Lives by jefftk
|
Jan 03, 2024 |
LW - AI Is Not Software by Davidmanheim
|
Jan 02, 2024 |
LW - Stop talking about p(doom) by Isaac King
|
Jan 02, 2024 |
LW - Gentleness and the artificial Other by Joe Carlsmith
|
Jan 02, 2024 |
LW - Apologizing is a Core Rationalist Skill by johnswentworth
|
Jan 02, 2024 |
LW - Boston Solstice 2023 Retrospective by jefftk
|
Jan 02, 2024 |
LW - Bayesian updating in real life is mostly about understanding your hypotheses by Max H
|
Jan 01, 2024 |
LW - 2023 in AI predictions by jessicata
|
Jan 01, 2024 |
LW - Planning to build a cryptographic box with perfect secrecy by Lysandre Terrisse
|
Jan 01, 2024 |
LW - Dark Skies Book Review by PeterMcCluskey
|
Dec 31, 2023 |
LW - shoes with springs by bhauth
|
Dec 31, 2023 |
LW - Taking responsibility and partial derivatives by Ruby
|
Dec 31, 2023 |
LW - The proper response to mistakes that have harmed others? by Ruby
|
Dec 31, 2023 |
LW - The Plan - 2023 Version by johnswentworth
|
Dec 30, 2023 |
LW - Will 2024 be very hot? Should we be worried? by A.H.
|
Dec 29, 2023 |
LW - NYT is suing OpenAIandMicrosoft for alleged copyright infringement; some quick thoughts by Mikhail Samin
|
Dec 28, 2023 |
LW - In Defense of Epistemic Empathy by Kevin Dorst
|
Dec 28, 2023 |
LW - How Emergency Medicine Solves the Alignment Problem by StrivingForLegibility
|
Dec 27, 2023 |
LW - Environmental allergies are curable? (Sublingual immunotherapy) by Chipmonk
|
Dec 27, 2023 |
LW - AI's impact on biology research: Part I, today by octopocta
|
Dec 27, 2023 |
LW - METR is hiring! by Beth Barnes
|
Dec 27, 2023 |
LW - How "Pause AI" advocacy could be net harmful by Tamsin Leake
|
Dec 26, 2023 |
LW - Flagging Potentially Unfair Parenting by jefftk
|
Dec 26, 2023 |
LW - A Crisper Explanation of Simulacrum Levels by Thane Ruthenis
|
Dec 24, 2023 |
LW - AI Girlfriends Won't Matter Much by Maxwell Tabarrok
|
Dec 23, 2023 |
LW - The problem with infohazards as a concept [Linkpost] by Noosphere89
|
Dec 22, 2023 |
LW - Pseudonymity and Accusations by jefftk
|
Dec 22, 2023 |
LW - The LessWrong 2022 Review: Review Phase by RobertM
|
Dec 22, 2023 |
LW - AI Safety Chatbot by markov
|
Dec 22, 2023 |
LW - On OpenAI's Preparedness Framework by Zvi
|
Dec 21, 2023 |
LW - Most People Don't Realize We Have No Idea How Our AIs Work by Thane Ruthenis
|
Dec 21, 2023 |
LW - Prediction Markets aren't Magic by SimonM
|
Dec 21, 2023 |
LW - Legalize butanol? by bhauth
|
Dec 21, 2023 |
LW - Matrix completion prize results by paulfchristiano
|
Dec 20, 2023 |
LW - Goal-Completeness is like Turing-Completeness for AGI by Liron
|
Dec 20, 2023 |
LW - Monthly Roundup #13: December 2023 by Zvi
|
Dec 20, 2023 |
LW - The Dark Arts by lsusr
|
Dec 19, 2023 |
LW - A Universal Emergent Decomposition of Retrieval Tasks in Language Models by Alexandre Variengien
|
Dec 19, 2023 |
LW - Constellations are Younger than Continents by Jeffrey Heninger
|
Dec 19, 2023 |
LW - OpenAI: Preparedness framework by Zach Stein-Perlman
|
Dec 18, 2023 |
LW - The 'Neglected Approaches' Approach: AE Studio's Alignment Agenda by Cameron Berg
|
Dec 18, 2023 |
LW - What makes teaching math special by Viliam
|
Dec 18, 2023 |
LW - Talk: "AI Would Be A Lot Less Alarming If We Understood Agents" by johnswentworth
|
Dec 18, 2023 |
LW - Scale Was All We Needed, At First by Gabriel Mukobi
|
Dec 18, 2023 |
LW - The Serendipity of Density by jefftk
|
Dec 17, 2023 |
LW - cold aluminum for medicine by bhauth
|
Dec 17, 2023 |
LW - 2022 (and All Time) Posts by Pingback Count by Raemon
|
Dec 17, 2023 |
LW - "Humanity vs. AGI" Will Never Look Like "Humanity vs. AGI" to Humanity by Thane Ruthenis
|
Dec 16, 2023 |
LW - Talking With People Who Speak to Congressional Staffers about AI risk by Eneasz
|
Dec 16, 2023 |
LW - Contra Scott on Abolishing the FDA by Maxwell Tabarrok
|
Dec 15, 2023 |
LW - "AI Alignment" is a Dangerously Overloaded Term by Roko
|
Dec 15, 2023 |
LW - EU policymakers reach an agreement on the AI Act by tlevin
|
Dec 15, 2023 |
LW - Some for-profit AI alignment org ideas by Eric Ho
|
Dec 14, 2023 |
LW - Love, Reverence, and Life by Elizabeth
|
Dec 14, 2023 |
LW - Bayesian Injustice by Kevin Dorst
|
Dec 14, 2023 |
LW - Update on Chinese IQ-related gene panels by Lao Mein
|
Dec 14, 2023 |
LW - How bad is chlorinated water? by bhauth
|
Dec 14, 2023 |
LW - Are There Examples of Overhang for Other Technologies? by Jeffrey Heninger
|
Dec 14, 2023 |
LW - Is being sexy for your homies? by Valentine
|
Dec 13, 2023 |
LW - The Best of Don't Worry About the Vase by Zvi
|
Dec 13, 2023 |
LW - AI Views Snapshots by Rob Bensinger
|
Dec 13, 2023 |
LW - Enhancing intelligence by banging your head on the wall by Bezzi
|
Dec 13, 2023 |
LW - [Valence series] 3. Valence and Beliefs by Steven Byrnes
|
Dec 13, 2023 |
LW - Balsa Update and General Thank You by Zvi
|
Dec 13, 2023 |
LW - Funding case: AI Safety Camp by Remmelt
|
Dec 12, 2023 |
LW - OpenAI: Leaks Confirm the Story by Zvi
|
Dec 12, 2023 |
LW - What is the next level of rationality? by lsusr
|
Dec 12, 2023 |
LW - Secondary Risk Markets by Vaniver
|
Dec 12, 2023 |
LW - The Consciousness Box by GradualImprovement
|
Dec 12, 2023 |
LW - The likely first longevity drug is based on sketchy science. This is bad for science and bad for longevity. by BobBurgers
|
Dec 12, 2023 |
LW - On plans for a functional society by kave
|
Dec 12, 2023 |
LW - re: Yudkowsky on biological materials by bhauth
|
Dec 11, 2023 |
LW - Principles For Product Liability (With Application To AI) by johnswentworth
|
Dec 10, 2023 |
LW - Unpicking Extinction by ukc10014
|
Dec 10, 2023 |
LW - Some negative steganography results by Fabien Roger
|
Dec 09, 2023 |
LW - The Offense-Defense Balance Rarely Changes by Maxwell Tabarrok
|
Dec 09, 2023 |
LW - "Model UN Solutions" by Arjun Panickssery
|
Dec 09, 2023 |
LW - Refusal mechanisms: initial experiments with Llama-2-7b-chat by andyrdt
|
Dec 08, 2023 |
LW - What I Would Do If I Were Working On AI Governance by johnswentworth
|
Dec 08, 2023 |
LW - [Valence series] 2. Valence and Normativity by Steven Byrnes
|
Dec 08, 2023 |
LW - Is AlphaGo actually a consequentialist utility maximizer? by faul sname
|
Dec 08, 2023 |
LW - Meetup Tip: Heartbeat Messages by Screwtape
|
Dec 07, 2023 |
LW - Gemini 1.0 by Zvi
|
Dec 07, 2023 |
LW - On Trust by johnswentworth
|
Dec 07, 2023 |
LW - Anthropical Paradoxes are Paradoxes of Probability Theory by Ape in the coat
|
Dec 07, 2023 |
LW - Originality vs. Correctness by alkjash
|
Dec 06, 2023 |
LW - Based Beff Jezos and the Accelerationists by Zvi
|
Dec 06, 2023 |
LW - A Socratic dialogue with my student by lsusr
|
Dec 06, 2023 |
LW - On 'Responsible Scaling Policies' (RSPs) by Zvi
|
Dec 06, 2023 |
LW - How do you feel about LessWrong these days? [Open feedback thread] by jacobjacob
|
Dec 05, 2023 |
LW - We're all in this together by Tamsin Leake
|
Dec 05, 2023 |
LW - The LessWrong 2022 Review by habryka
|
Dec 05, 2023 |
LW - Speaking to Congressional staffers about AI risk by Akash
|
Dec 05, 2023 |
LW - n of m ring signatures by DanielFilan
|
Dec 05, 2023 |
LW - [Valence series] 1. Introduction by Steven Byrnes
|
Dec 04, 2023 |
LW - Book Review: 1948 by Benny Morris by Yair Halberstadt
|
Dec 04, 2023 |
LW - Meditations on Mot by Richard Ngo
|
Dec 04, 2023 |
LW - List of strategies for mitigating deceptive alignment by joshc
|
Dec 04, 2023 |
LW - Nietzsche's Morality in Plain English by Arjun Panickssery
|
Dec 04, 2023 |
LW - the micro-fulfillment cambrian explosion by bhauth
|
Dec 04, 2023 |
LW - The Witness by Richard Ngo
|
Dec 04, 2023 |
LW - Quick takes on "AI is easy to control" by So8res
|
Dec 02, 2023 |
LW - Out-of-distribution Bioattacks by jefftk
|
Dec 02, 2023 |
LW - 2023 Unofficial LessWrong Census/Survey by Screwtape
|
Dec 02, 2023 |
LW - Complex systems research as a field (and its relevance to AI Alignment) by Nora Ammann
|
Dec 02, 2023 |
LW - MATS Summer 2023 Postmortem by Rocket
|
Dec 02, 2023 |
LW - Queuing theory: Benefits of operating at 70% capacity by ampdot
|
Dec 02, 2023 |
LW - How useful is mechanistic interpretability? by ryan greenblatt
|
Dec 01, 2023 |
LW - What's next for the field of Agent Foundations? by Nora Ammann
|
Nov 30, 2023 |
LW - Scaling laws for dominant assurance contracts by jessicata
|
Nov 30, 2023 |
LW - OpenAI: Altman Returns by Zvi
|
Nov 30, 2023 |
LW - Stupid Question: Why am I getting consistently downvoted? by MadHatter
|
Nov 30, 2023 |
LW - Lying Alignment Chart by Zack M Davis
|
Nov 29, 2023 |
LW - How to Control an LLM's Behavior (why my P(DOOM) went down) by RogerDearnaley
|
Nov 29, 2023 |
LW - Black Box Biology by GeneSmith
|
Nov 29, 2023 |
LW - The 101 Space You Will Always Have With You by Screwtape
|
Nov 29, 2023 |
LW - I'm confused about innate smell neuroanatomy by Steven Byrnes
|
Nov 29, 2023 |
LW - AISC 2024 - Project Summaries by NickyP
|
Nov 29, 2023 |
LW - Update #2 to "Dominant Assurance Contract Platform": EnsureDone by moyamo
|
Nov 28, 2023 |
LW - [Linkpost] George Mack's Razors by trevor
|
Nov 28, 2023 |
LW - Apply to the Conceptual Boundaries Workshop for AI Safety by Chipmonk
|
Nov 28, 2023 |
LW - My techno-optimism [By Vitalik Buterin] by habryka
|
Nov 28, 2023 |
LW - "Epistemic range of motion" and LessWrong moderation by habryka
|
Nov 28, 2023 |
LW - Apocalypse insurance, and the hardline libertarian take on AI risk by So8res
|
Nov 28, 2023 |
LW - Paper: "FDT in an evolutionary environment" by the gears to ascension
|
Nov 27, 2023 |
LW - why did OpenAI employees sign by bhauth
|
Nov 27, 2023 |
LW - Spaced repetition for teaching two-year olds how to read (Interview) by Chipmonk
|
Nov 27, 2023 |
LW - Moral Reality Check (a short story) by jessicata
|
Nov 26, 2023 |
LW - What are the results of more parental supervision and less outdoor play? by juliawise
|
Nov 25, 2023 |
LW - Progress links digest, 2023-11-24: Bottlenecks of aging, Starship launches, and much more by jasoncrawford
|
Nov 25, 2023 |
LW - Prepsgiving, A Convergently Instrumental Human Practice by JenniferRM
|
Nov 25, 2023 |
LW - What did you change your mind about in the last year? by mike hawke
|
Nov 24, 2023 |
LW - Never Drop A Ball by Screwtape
|
Nov 24, 2023 |
LW - AI #39: The Week of OpenAI by Zvi
|
Nov 23, 2023 |
LW - Possible OpenAI's Q* breakthrough and DeepMind's AlphaGo-type systems plus LLMs by Burny
|
Nov 23, 2023 |
LW - so you want to save the world? an account in paladinhood by Tamsin Leake
|
Nov 22, 2023 |
LW - OpenAI: The Battle of the Board by Zvi
|
Nov 22, 2023 |
LW - Atlantis: Berkeley event venue available for rent by Jonas Vollmer
|
Nov 22, 2023 |
LW - Userscript to always show LW comments in context vs at the top by Vlad Sitalo
|
Nov 21, 2023 |
LW - Dialogue on the Claim: "OpenAI's Firing of Sam Altman (And Shortly-Subsequent Events) On Net Reduced Existential Risk From AGI" by johnswentworth
|
Nov 21, 2023 |
LW - Why not electric trains and excavators? by bhauth
|
Nov 21, 2023 |
LW - Navigating emotions in an uncertain and confusing world by Akash
|
Nov 21, 2023 |
LW - For Civilization and Against Niceness by Gabriel Alfour
|
Nov 21, 2023 |
LW - Vote on worthwhile OpenAI topics to discuss by Ben Pace
|
Nov 21, 2023 |
LW - Agent Boundaries Aren't Markov Blankets. [no longer endorsed] by abramdemski
|
Nov 20, 2023 |
LW - OpenAI Staff (including Sutskever) Threaten to Quit Unless Board Resigns by Seth Herd
|
Nov 20, 2023 |
LW - OpenAI: Facts from a Weekend by Zvi
|
Nov 20, 2023 |
LW - Sam Altman, Greg Brockman and others from OpenAI join Microsoft by Ozyrus
|
Nov 20, 2023 |
LW - New paper shows truthfulness and instruction-following don't generalize by default by joshc
|
Nov 19, 2023 |
LW - "Why can't you just turn it off?" by Roko
|
Nov 19, 2023 |
LW - Spaciousness In Partner Dance: A Naturalism Demo by LoganStrohl
|
Nov 19, 2023 |
LW - Altman firing retaliation incoming? by trevor
|
Nov 19, 2023 |
LW - Sam Altman fired from OpenAI by LawrenceC
|
Nov 17, 2023 |
LW - On the lethality of biased human reward ratings by Eli Tyre
|
Nov 17, 2023 |
LW - On Lies and Liars by Gabriel Alfour
|
Nov 17, 2023 |
LW - On Tapping Out by Screwtape
|
Nov 17, 2023 |
LW - A to Z of things by KatjaGrace
|
Nov 17, 2023 |
LW - Forecasting AI (Overview) by jsteinhardt
|
Nov 17, 2023 |
LW - New LessWrong feature: Dialogue Matching by jacobjacob
|
Nov 16, 2023 |
LW - Learning coefficient estimation: the details by Zach Furman
|
Nov 16, 2023 |
LW - Social Dark Matter by [DEACTIVATED] Duncan Sabien
|
Nov 16, 2023 |
LW - In Defense of Parselmouths by Screwtape
|
Nov 16, 2023 |
LW - 'Theories of Values' and 'Theories of Agents': confusions, musings and desiderata by Mateusz Bagiński
|
Nov 16, 2023 |
LW - Extrapolating from Five Words by Gordon Seidoh Worley
|
Nov 16, 2023 |
LW - With or without a scratchpad, Large Language Models can Strategically Deceive their Users when Put Under Pressure. Results of an autonomous stock trading agent in a realistic, simulated environment. by ReaderM
|
Nov 15, 2023 |
LW - Testbed evals: evaluating AI safety even when it can't be directly measured by joshc
|
Nov 15, 2023 |
LW - Reinforcement Via Giving People Cookies by Screwtape
|
Nov 15, 2023 |
LW - Monthly Roundup #12: November 2023 by Zvi
|
Nov 15, 2023 |
LW - Raemon's Deliberate ("Purposeful?") Practice Club by Raemon
|
Nov 15, 2023 |
LW - Kids or No kids by Kids or no kids
|
Nov 14, 2023 |
LW - A framing for interpretability by Nina Rimsky
|
Nov 14, 2023 |
LW - What is wisdom? by TsviBT
|
Nov 14, 2023 |
LW - When did Eliezer Yudkowsky change his mind about neural networks? by Yarrow Bouchard
|
Nov 14, 2023 |
LW - They are made of repeating patterns by quetzal rainbow
|
Nov 14, 2023 |
LW - Loudly Give Up, Don't Quietly Fade by Screwtape
|
Nov 14, 2023 |
LW - Redirecting one's own taxes as an effective altruism method by David Gross
|
Nov 13, 2023 |
LW - Bostrom Goes Unheard by Zvi
|
Nov 13, 2023 |
LW - The Fundamental Theorem for measurable factor spaces by Matthias G. Mayer
|
Nov 13, 2023 |
LW - You can just spontaneously call people you haven't met in years by lc
|
Nov 13, 2023 |
LW - Zvi's Manifold Markets House Rules by Zvi
|
Nov 13, 2023 |
LW - Don't Donate A Kidney To A Stranger by George3d6
|
Nov 12, 2023 |
LW - It's OK to be biased towards humans by dr s
|
Nov 12, 2023 |
LW - Palisade is hiring Research Engineers by Charlie Rogers-Smith
|
Nov 11, 2023 |
LW - GPT-2030 and Catastrophic Drives: Four Vignettes by jsteinhardt
|
Nov 11, 2023 |
LW - Picking Mentors For Research Programmes by Raymond D
|
Nov 10, 2023 |
LW - Text Posts from the Kids Group: 2021 by jefftk
|
Nov 10, 2023 |
LW - Making Bad Decisions On Purpose by Screwtape
|
Nov 10, 2023 |
LW - Polysemantic Attention Head in a 4-Layer Transformer by Jett
|
Nov 09, 2023 |
LW - On OpenAI Dev Day by Zvi
|
Nov 09, 2023 |
LW - A free to enter, 240 character, open-source iterated prisoner's dilemma tournament by Isaac King
|
Nov 09, 2023 |
LW - Concrete positive visions for a future without AGI by Max H
|
Nov 09, 2023 |
LW - Five projects from AI Safety Hub Labs 2023 by charlie griffin
|
Nov 08, 2023 |
LW - Thinking By The Clock by Screwtape
|
Nov 08, 2023 |
LW - The Stochastic Parrot Hypothesis is debatable for the last generation of LLMs by Quentin FEUILLADE--MONTIXI
|
Nov 08, 2023 |
LW - The Perils of Professionalism by Screwtape
|
Nov 08, 2023 |
LW - Vote on Interesting Disagreements by Ben Pace
|
Nov 07, 2023 |
LW - Announcing Athena - Women in AI Alignment Research by Claire Short
|
Nov 07, 2023 |
LW - AMA: Earning to Give by jefftk
|
Nov 07, 2023 |
LW - On the UK Summit by Zvi
|
Nov 07, 2023 |
LW - Job listing: Communications Generalist / Project Manager by Gretta Duleba
|
Nov 07, 2023 |
LW - Are language models good at making predictions? by dynomight
|
Nov 06, 2023 |
LW - The Assumed Intent Bias by silentbob
|
Nov 06, 2023 |
LW - On Overhangs and Technological Change by Roko
|
Nov 06, 2023 |
LW - Being good at the basics by dominicq
|
Nov 06, 2023 |
LW - Pivotal Acts might Not be what You Think they are by Johannes C. Mayer
|
Nov 05, 2023 |
LW - Stuxnet, not Skynet: Humanity's disempowerment by AI by Roko
|
Nov 04, 2023 |
LW - We are already in a persuasion-transformed world and must take precautions by trevor
|
Nov 04, 2023 |
LW - The Soul Key by Richard Ngo
|
Nov 04, 2023 |
LW - The 6D effect: When companies take risks, one email can be very powerful. by scasper
|
Nov 04, 2023 |
LW - Deception Chess: Game #1 by Zane
|
Nov 03, 2023 |
LW - 8 examples informing my pessimism on uploading without reverse engineering by Steven Byrnes
|
Nov 03, 2023 |
LW - Integrity in AI Governance and Advocacy by habryka
|
Nov 03, 2023 |
LW - One Day Sooner by Screwtape
|
Nov 03, 2023 |
LW - The other side of the tidal wave by KatjaGrace
|
Nov 03, 2023 |
LW - Saying the quiet part out loud: trading off x-risk for personal immortality by disturbance
|
Nov 02, 2023 |
LW - Propaganda or Science: A Look at Open Source AI and Bioterrorism Risk by 1a3orn
|
Nov 02, 2023 |
LW - Estimating effective dimensionality of MNIST models by Arjun Panickssery
|
Nov 02, 2023 |
LW - Snapshot of narratives and frames against regulating AI by Jan Kulveit
|
Nov 02, 2023 |
LW - Public Weights? by jefftk
|
Nov 02, 2023 |
LW - Chinese scientists acknowledge xrisk and call for international regulatory body [Linkpost] by Akash
|
Nov 02, 2023 |
LW - Reactions to the Executive Order by Zvi
|
Nov 01, 2023 |
LW - 2023 LessWrong Community Census, Request for Comments by Screwtape
|
Nov 01, 2023 |
LW - On the Executive Order by Zvi
|
Nov 01, 2023 |
LW - Mission Impossible: Dead Reckoning Part 1 AI Takeaways by Zvi
|
Nov 01, 2023 |
LW - Urging an International AI Treaty: An Open Letter by Loppukilpailija
|
Nov 01, 2023 |
LW - Linkpost: A Post Mortem on the Gino Case by Linch
|
Oct 24, 2023 |
LW - What is an "anti-Occamian prior"? by Zane
|
Oct 23, 2023 |
LW - Sam Altman's sister, Annie Altman, claims Sam has severely abused her by pl5015
|
Oct 07, 2023 |
LW - Monthly Roundup #11: October 2023 by Zvi
|
Oct 04, 2023 |
LW - When to Get the Booster? by jefftk
|
Oct 04, 2023 |
LW - OpenAI-Microsoft partnership by Zach Stein-Perlman
|
Oct 04, 2023 |
LW - energy landscapes of experts by bhauth
|
Oct 03, 2023 |
LW - Linkpost: They Studied Dishonesty. Was Their Work a Lie? by Linch
|
Oct 02, 2023 |
LW - Thomas Kwa's MIRI research experience by Thomas Kwa
|
Oct 02, 2023 |
LW - Conditionals All The Way Down by lunatic at large
|
Oct 02, 2023 |
LW - The 99% principle for personal problems by Kaj Sotala
|
Oct 02, 2023 |
LW - Fifty Flips by abstractapplic
|
Oct 02, 2023 |
LW - My Effortless Weightloss Story: A Quick Runthrough by CuoreDiVetro
|
Oct 01, 2023 |
LW - Competitive, Cooperative, and Cohabitive by Screwtape
|
Oct 01, 2023 |
LW - The Lighthaven Campus is open for bookings by habryka
|
Sep 29, 2023 |
LW - Announcing FAR Labs, an AI safety coworking space by bgold
|
Sep 29, 2023 |
LW - Bids To Defer On Value Judgements by johnswentworth
|
Sep 29, 2023 |
LW - What's your standard for good work performance? by Chi Nguyen
|
Sep 29, 2023 |
LW - High-level interpretability: detecting an AI's objectives by Paul Colognese
|
Sep 29, 2023 |
LW - Peacewagers so Far by mako yass
|
Sep 29, 2023 |
LW - The point of a game is not to win, and you shouldn't even pretend that it is by mako yass
|
Sep 29, 2023 |
LW - EA Vegan Advocacy is not truthseeking, and it's everyone's problem by Elizabeth
|
Sep 28, 2023 |
LW - AI #31: It Can Do What Now? by Zvi
|
Sep 28, 2023 |
LW - The Hidden Complexity of Wishes - The Animation by Writer
|
Sep 28, 2023 |
LW - Petrov Day Retrospective, 2023 (re: the most important virtue of Petrov Day and unilaterally promoting it) by Ruby
|
Sep 28, 2023 |
LW - Jacob on the Precipice by Richard Ngo
|
Sep 27, 2023 |
LW - GPT-4 for personal productivity: online distraction blocker by Sergii
|
Sep 27, 2023 |
LW - Making AIs less likely to be spiteful by Nicolas Macé
|
Sep 26, 2023 |
LW - The King and the Golem by Richard Ngo
|
Sep 26, 2023 |
LW - "X distracts from Y" as a thinly-disguised fight over group status / politics by Steven Byrnes
|
Sep 25, 2023 |
LW - Inside Views, Impostor Syndrome, and the Great LARP by johnswentworth
|
Sep 25, 2023 |
LW - Interpreting OpenAI's Whisper by EllenaR
|
Sep 24, 2023 |
LW - I designed an AI safety course (for a philosophy department) by Eleni Angelou
|
Sep 24, 2023 |
LW - Paper: LLMs trained on "A is B" fail to learn "B is A" by lberglund
|
Sep 23, 2023 |
LW - Luck based medicine: inositol for anxiety and brain fog by Elizabeth
|
Sep 23, 2023 |
LW - Update to "Dominant Assurance Contract Platform" by moyamo
|
Sep 23, 2023 |
LW - Fund Transit With Development by jefftk
|
Sep 22, 2023 |
LW - Let's talk about Impostor syndrome in AI safety by Igor Ivanov
|
Sep 22, 2023 |
LW - Neel Nanda on the Mechanistic Interpretability Researcher Mindset by Michaël Trazzi
|
Sep 22, 2023 |
LW - If influence functions are not approximating leave-one-out, how are they supposed to help? by Fabien Roger
|
Sep 22, 2023 |
LW - Immortality or death by AGI by ImmortalityOrDeathByAGI
|
Sep 22, 2023 |
LW - Atoms to Agents Proto-Lectures by johnswentworth
|
Sep 22, 2023 |
LW - Would You Work Harder In The Least Convenient Possible World? by Firinn
|
Sep 22, 2023 |
LW - AI #30: Dalle-3 and GPT-3.5-Instruct-Turbo by Zvi
|
Sep 21, 2023 |
LW - Interpretability Externalities Case Study - Hungry Hungry Hippos by Magdalena Wache
|
Sep 20, 2023 |
LW - [Review] Move First, Think Later: Sense and Nonsense in Improving Your Chess by Arjun Panickssery
|
Sep 19, 2023 |
LW - Luck based medicine: angry eldritch sugar gods edition by Elizabeth
|
Sep 19, 2023 |
LW - Anthropic's Responsible Scaling Policy and Long-Term Benefit Trust by Zac Hatfield-Dodds
|
Sep 19, 2023 |
LW - Some reasons why I frequently prefer communicating via text by Adam Zerner
|
Sep 19, 2023 |
LW - Technical AI Safety Research Landscape [Slides] by Magdalena Wache
|
Sep 18, 2023 |
LW - Microdooms averted by working on AI Safety by nikola
|
Sep 18, 2023 |
LW - Show LW: Get a phone call if prediction markets predict nuclear war by Lorenzo
|
Sep 18, 2023 |
LW - The Talk: a brief explanation of sexual dimorphism by Malmesbury
|
Sep 18, 2023 |
LW - Three ways interpretability could be impactful by Arthur Conmy
|
Sep 18, 2023 |
LW - Eugenics Performed By A Blind, Idiot God by omnizoid
|
Sep 18, 2023 |
LW - Actually, "personal attacks after object-level arguments" is a pretty good rule of epistemic conduct by Max H
|
Sep 18, 2023 |
LW - Memory bandwidth constraints imply economies of scale in AI inference by Ege Erdil
|
Sep 17, 2023 |
LW - I compiled a ebook of 'Project Lawful' for eBook readers by OrwellGoesShopping
|
Sep 16, 2023 |
LW - Navigating an ecosystem that might or might not be bad for the world by habryka
|
Sep 16, 2023 |
LW - Deconfusing Regret by Alex Hollow
|
Sep 15, 2023 |
LW - A Theory of Laughter - Follow-Up by Steven Byrnes
|
Sep 15, 2023 |
LW - "Did you lock it?" by ymeskhout
|
Sep 15, 2023 |
LW - Can I take ducks home from the park? by dynomight
|
Sep 14, 2023 |
LW - Uncovering Latent Human Wellbeing in LLM Embeddings by ChengCheng
|
Sep 14, 2023 |
LW - Padding the Corner by jefftk
|
Sep 14, 2023 |
LW - AI #29: Take a Deep Breath by Zvi
|
Sep 14, 2023 |
LW - Instrumental Convergence Bounty by Logan Zoellner
|
Sep 14, 2023 |
LW - Highlights: Wentworth, Shah, and Murphy on "Retargeting the Search" by RobertM
|
Sep 14, 2023 |
LW - Linkpost for Jan Leike on Self-Exfiltration by Daniel Kokotajlo
|
Sep 14, 2023 |
LW - Is AI Safety dropping the ball on privacy? by markov
|
Sep 13, 2023 |
LW - Contra Yudkowsky on Epistemic Conduct for Author Criticism by Zack M Davis
|
Sep 13, 2023 |
LW - UDT shows that decision theory is more puzzling than ever by Wei Dai
|
Sep 13, 2023 |
LW - PSA: The community is in Berkeley/Oakland, not "the Bay Area" by maia
|
Sep 11, 2023 |
LW - Logical Share Splitting by DaemonicSigil
|
Sep 11, 2023 |
LW - Focus on the Hardest Part First by Johannes C. Mayer
|
Sep 11, 2023 |
LW - US presidents discuss AI alignment agendas by TurnTrout
|
Sep 09, 2023 |
LW - Have Attention Spans Been Declining? by niplav
|
Sep 09, 2023 |
LW - What is the optimal frontier for due diligence? by RobertM
|
Sep 09, 2023 |
LW - Sum-threshold attacks by TsviBT
|
Sep 08, 2023 |
LW - AI#28: Watching and Waiting by Zvi
|
Sep 08, 2023 |
LW - Recreating the caring drive by Catnee
|
Sep 08, 2023 |
LW - A quick update from Nonlinear by KatWoods
|
Sep 07, 2023 |
LW - Feedback-loops, Deliberate Practice, and Transfer Learning by jacobjacob
|
Sep 07, 2023 |
LW - My First Post by Jaivardhan Nawani
|
Sep 07, 2023 |
LW - Sharing Information About Nonlinear by Ben Pace
|
Sep 07, 2023 |
LW - Find Hot French Food Near Me: A Follow-up by aphyer
|
Sep 06, 2023 |
LW - Who Has the Best Food? by Zvi
|
Sep 05, 2023 |
LW - Text Posts from the Kids Group: 2023 I by jefftk
|
Sep 05, 2023 |
LW - a rant on politician-engineer coalitional conflict by bhauth
|
Sep 04, 2023 |
LW - Defunding My Mistake by ymeskhout
|
Sep 04, 2023 |
LW - Tensor Trust: An online game to uncover prompt injection vulnerabilities by Luke Bailey
|
Sep 04, 2023 |
LW - The goal of physics by Jim Pivarski
|
Sep 03, 2023 |
LW - The smallest possible button by Neil
|
Sep 02, 2023 |
LW - One Minute Every Moment by abramdemski
|
Sep 01, 2023 |
LW - Reproducing ARC Evals' recent report on language model agents by Thomas Broadley
|
Sep 01, 2023 |
LW - A Golden Age of Building? Excerpts and lessons from Empire State, Pentagon, Skunk Works and SpaceX by jacobjacob
|
Sep 01, 2023 |
LW - AI #27: Portents of Gemini by Zvi
|
Aug 31, 2023 |
LW - Responses to apparent rationalist confusions about game / decision theory by Anthony DiGiovanni
|
Aug 31, 2023 |
LW - Report on Frontier Model Training by YafahEdelman
|
Aug 31, 2023 |
LW - Biosecurity Culture, Computer Security Culture by jefftk
|
Aug 30, 2023 |
LW - Open Call for Research Assistants in Developmental Interpretability by Jesse Hoogland
|
Aug 30, 2023 |
LW - The Economics of the Asteroid Deflection Problem by moyamo
|
Aug 30, 2023 |
LW - Trying a Wet Suit by jefftk
|
Aug 29, 2023 |
LW - Broken Benchmark: MMLU by awg
|
Aug 29, 2023 |
LW - Dating Roundup #1: This is Why You're Single by Zvi
|
Aug 29, 2023 |
LW - Information warfare historically revolved around human conduits by trevor
|
Aug 29, 2023 |
LW - ACX Meetups Everywhere 2023: Times and Places by Scott Alexander
|
Aug 29, 2023 |
LW - Introducing the Center for AI Policy (and we're hiring!) by Thomas Larsen
|
Aug 28, 2023 |
LW - AI pause/governance advocacy might be net-negative, especially without focus on explaining the x-risk by Mikhail Samin
|
Aug 28, 2023 |
LW - Dear Self; we need to talk about ambition by Elizabeth
|
Aug 28, 2023 |
LW - Aumann-agreement is common by tailcalled
|
Aug 27, 2023 |
LW - Digital brains beat biological ones because diffusion is too slow by GeneSmith
|
Aug 26, 2023 |
LW - Red-teaming language models via activation engineering by Nina Rimsky
|
Aug 26, 2023 |
LW - When Omnipotence is Not Enough by lsusr
|
Aug 26, 2023 |
LW - Assume Bad Faith by Zack M Davis
|
Aug 25, 2023 |
LW - Apply for the 2023 Developmental Interpretability Conference! by Stan van Wingerden
|
Aug 25, 2023 |
LW - AI #26: Fine Tuning Time by Zvi
|
Aug 25, 2023 |
LW - AI Regulation May Be More Important Than AI Alignment For Existential Safety by otto.barten
|
Aug 24, 2023 |
LW - The lost millennium by Ege Erdil
|
Aug 24, 2023 |
LW - Assessment of intelligence agency functionality is difficult yet important by trevor
|
Aug 24, 2023 |
LW - The God of Humanity, and the God of the Robot Utilitarians by Raemon
|
Aug 24, 2023 |
LW - Seth Explains Consciousness by Jacob Falkovich
|
Aug 24, 2023 |
LW - The Low-Hanging Fruit Prior and sloped valleys in the loss landscape by Dmitry Vaintrob
|
Aug 24, 2023 |
LW - Diet Experiment Preregistration: Long-term water fasting + seed oil removal by lc
|
Aug 24, 2023 |
LW - Why Is No One Trying To Align Profit Incentives With Alignment Research? by Prometheus
|
Aug 23, 2023 |
LW - A Theory of Laughter by Steven Byrnes
|
Aug 23, 2023 |
LW - Walk while you talk: don't balk at "no chalk" by dkl9
|
Aug 23, 2023 |
LW - State of Generally Available Self-Driving by jefftk
|
Aug 22, 2023 |
LW - Large Language Models will be Great for Censorship by Ethan Edwards
|
Aug 22, 2023 |
LW - Which possible AI systems are relatively safe? by Zach Stein-Perlman
|
Aug 22, 2023 |
LW - DIY Deliberate Practice by lynettebye
|
Aug 21, 2023 |
LW - Ideas for improving epistemics in AI safety outreach by mic
|
Aug 21, 2023 |
LW - Efficiency and resource use scaling parity by Ege Erdil
|
Aug 21, 2023 |
LW - Ruining an expected-log-money maximizer by philh
|
Aug 21, 2023 |
LW - Chess as a case study in hidden capabilities in ChatGPT by AdamYedidia
|
Aug 21, 2023 |
LW - Steven Wolfram on AI Alignment by Bill Benzon
|
Aug 21, 2023 |
LW - "Dirty concepts" in AI alignment discourses, and some guesses for how to deal with them by Nora Ammann
|
Aug 20, 2023 |
LW - AI Forecasting: Two Years In by jsteinhardt
|
Aug 20, 2023 |
LW - Is Chinese total factor productivity lower today than it was in 1956? by Ege Erdil
|
Aug 19, 2023 |
LW - The U.S. is mildly destabilizing by lc
|
Aug 18, 2023 |
LW - 6 non-obvious mental health issues specific to AI safety. by Igor Ivanov
|
Aug 18, 2023 |
LW - Announcing Foresight Institute's AI Safety Grants Program by Allison Duettmann
|
Aug 18, 2023 |
LW - What does it mean to "trust science"? by jasoncrawford
|
Aug 18, 2023 |
LW - AI #25: Inflection Point by Zvi
|
Aug 17, 2023 |
LW - Against Almost Every Theory of Impact of Interpretability by Charbel-Raphaël
|
Aug 17, 2023 |
LW - Reflections on "Making the Atomic Bomb" by boazbarak
|
Aug 17, 2023 |
LW - The Dunbar Playbook: A CRM system for your friends by Severin T. Seehrich
|
Aug 17, 2023 |
LW - If we had known the atmosphere would ignite by Jeffs
|
Aug 17, 2023 |
LW - Book Launch: "The Carving of Reality," Best of LessWrong vol. III by Raemon
|
Aug 17, 2023 |
LW - Summary of and Thoughts on the Hotz/Yudkowsky Debate by Zvi
|
Aug 16, 2023 |
LW - Understanding and visualizing sycophancy datasets by Nina Rimsky
|
Aug 16, 2023 |
LW - Ten Thousand Years of Solitude by agp
|
Aug 16, 2023 |
LW - Optical Illusions are Out of Distribution Errors by vitaliya
|
Aug 15, 2023 |
LW - My checklist for publishing a blog post by Steven Byrnes
|
Aug 15, 2023 |
LW - A short calculation about a Twitter poll by Ege Erdil
|
Aug 14, 2023 |
LW - Decomposing independent generalizations in neural networks via Hessian analysis by Dmitry Vaintrob
|
Aug 14, 2023 |
LW - Stepping down as moderator on LW by Kaj Sotala
|
Aug 14, 2023 |
LW - We Should Prepare for a Larger Representation of Academia in AI Safety by Leon Lang
|
Aug 13, 2023 |
LW - [Linkpost] Personal and Psychological Dimensions of AI Researchers Confronting AI Catastrophic Risks by Bogdan Ionut Cirstea
|
Aug 13, 2023 |
LW - Simulate the CEO by robotelvis
|
Aug 13, 2023 |
LW - Biological Anchors: The Trick that Might or Might Not Work by Scott Alexander
|
Aug 12, 2023 |
LW - AI #24: Week of the Podcast by Zvi
|
Aug 11, 2023 |
LW - LLMs are (mostly) not helped by filler tokens by Kshitij Sachan
|
Aug 10, 2023 |
LW - marine cloud brightening by bhauth
|
Aug 10, 2023 |
LW - Modulating sycophancy in an RLHF model via activation steering by NinaR
|
Aug 09, 2023 |
LW - Inflection.ai is a major AGI lab by nikola
|
Aug 09, 2023 |
LW - What's A "Market"? by johnswentworth
|
Aug 09, 2023 |
LW - Perpetually Declining Population? by jefftk
|
Aug 08, 2023 |
LW - A plea for more funding shortfall transparency by porby
|
Aug 08, 2023 |
LW - Feedbackloop-first Rationality by Raemon
|
Aug 07, 2023 |
LW - 'We're changing the clouds.' An unforeseen test of geoengineering is fueling record ocean warmth by Annapurna
|
Aug 07, 2023 |
LW - Problems with Robin Hanson's Quillette Article On AI by DaemonicSigil
|
Aug 07, 2023 |
LW - Computational Thread Art by TheMcDouglas
|
Aug 07, 2023 |
LW - how 2 tell if ur input is out of distribution given only model weights by dkirmani
|
Aug 06, 2023 |
LW - Stomach Ulcers and Dental Cavities by Metacelsus
|
Aug 06, 2023 |
LW - The Sinews of Sudan's Latest War by Tim Liptrot
|
Aug 05, 2023 |
LW - Private notes on LW? by Raemon
|
Aug 04, 2023 |
LW - Password-locked models: a stress case for capabilities evaluation by Fabien Roger
|
Aug 03, 2023 |
LW - AI #23: Fundamental Problems with RLHF by Zvi
|
Aug 03, 2023 |
LW - "Is There Anything That's Worth More" by Zack M Davis
|
Aug 02, 2023 |
LW - My current LK99 questions by Eliezer Yudkowsky
|
Aug 01, 2023 |
LW - Barbieheimer: Across the Dead Reckoning by Zvi
|
Aug 01, 2023 |
LW - Exercise: Solve "Thinking Physics" by Raemon
|
Aug 01, 2023 |
LW - The "public debate" about AI is confusing for the general public and for policymakers because it is a three-sided debate by Adam David Long
|
Aug 01, 2023 |
LW - A Social History of Truth by Vaniver
|
Aug 01, 2023 |
LW - "Building a House" Review by jefftk
|
Jul 31, 2023 |
LW - Lack of Social Grace Is an Epistemic Virtue by Zack M Davis
|
Jul 31, 2023 |
LW - Is Light Drinking Protective? by jefftk
|
Jul 31, 2023 |
LW - Apollo Neuro Results by Elizabeth
|
Jul 30, 2023 |
LW - How to make real-money prediction markets on arbitrary topics by yutaka
|
Jul 30, 2023 |
LW - Self-driving car bets by paulfchristiano
|
Jul 29, 2023 |
LW - Introductory Textbook to Vision Models Interpretability by jeanne
|
Jul 29, 2023 |
LW - Why You Should Never Update Your Beliefs by Arjun Panickssery
|
Jul 29, 2023 |
LW - When can we trust model evaluations? by evhub
|
Jul 29, 2023 |
LW - Mech Interp Puzzle 2: Word2Vec Style Embeddings by Neel Nanda
|
Jul 28, 2023 |
LW - Yes, It's Subjective, But Why All The Crabs? by johnswentworth
|
Jul 28, 2023 |
LW - Pulling the Rope Sideways: Empirical Test Results by Daniel Kokotajlo
|
Jul 28, 2023 |
LW - Visible loss landscape basins don't correspond to distinct algorithms by Mikhail Samin
|
Jul 28, 2023 |
LW - Reducing sycophancy and improving honesty via activation steering by NinaR
|
Jul 28, 2023 |
LW - SSA rejects anthropic shadow, too by jessicata
|
Jul 27, 2023 |
LW - SSA rejects anthropic shadow, too by jessicata
|
Jul 27, 2023 |
LW - Cultivating a state of mind where new ideas are born by Henrik Karlsson
|
Jul 27, 2023 |
LW - Cultivating a state of mind where new ideas are born by Henrik Karlsson
|
Jul 27, 2023 |
LW - The First Room-Temperature Ambient-Pressure Superconductor by Annapurna
|
Jul 27, 2023 |
LW - Llama We Doing This Again? by Zvi
|
Jul 27, 2023 |
LW - Why no Roman Industrial Revolution? by jasoncrawford
|
Jul 26, 2023 |
LW - Neuronpedia - AI Safety Game by hijohnnylin
|
Jul 26, 2023 |
LW - Meta-level adversarial evaluation of oversight techniques might allow robust measurement of their adequacy by Buck
|
Jul 26, 2023 |
LW - AI Safety Hub Serbia Soft Launch by DusanDNesic
|
Jul 26, 2023 |
LW - Underwater Torture Chambers: The Horror Of Fish Farming by omnizoid
|
Jul 26, 2023 |
LW - Thoughts on Loss Landscapes and why Deep Learning works by beren
|
Jul 25, 2023 |
LW - Anthropic Observations by Zvi
|
Jul 25, 2023 |
LW - How LLMs are and are not myopic by janus
|
Jul 25, 2023 |
LW - Open problems in activation engineering by TurnTrout
|
Jul 24, 2023 |
LW - A reformulation of Finite Factored Sets by Matthias G. Mayer
|
Jul 24, 2023 |
LW - Slowing down AI progress is an underexplored alignment strategy by Norman Borlaug
|
Jul 24, 2023 |
LW - Brain Efficiency Cannell Prize Contest Award Ceremony by Alexander Gietelink Oldenziel
|
Jul 24, 2023 |
LW - Cryonics and Regret by MvB
|
Jul 24, 2023 |
LW - Rationality !== winning by Raemon
|
Jul 24, 2023 |
LW - Which rationality posts are begging for further practical development? by LoganStrohl
|
Jul 24, 2023 |
LW - QAPR 5: grokking is maybe not that big a deal? by Quintin Pope
|
Jul 23, 2023 |
LW - Autogynephilia discourse is so absurdly bad on all sides by tailcalled
|
Jul 23, 2023 |
LW - "Justice, Cherryl." by Zack M Davis
|
Jul 23, 2023 |
LW - The UAP Disclosure Act of 2023 and its implications by andeslodes
|
Jul 23, 2023 |
LW - BCIs and the ecosystem of modular minds by beren
|
Jul 22, 2023 |
LW - All AGI Safety questions welcome (especially basic ones) [July 2023] by smallsilo
|
Jul 22, 2023 |
LW - I'm consistently overwhelmed by basic obligations. Are there any paradigm shifts or other rationality-based tips that would be helpful? by Benjamin Hendricks
|
Jul 21, 2023 |
LW - News : Biden-Harris Administration Secures Voluntary Commitments from Leading Artificial Intelligence Companies to Manage the Risks Posed by AI by Jonathan Claybrough
|
Jul 21, 2023 |
LW - Priorities for the UK Foundation Models Taskforce by Andrea Miotti
|
Jul 21, 2023 |
LW - Boundary Placement Rebellion by tailcalled
|
Jul 21, 2023 |
LW - Announcement: AI Narrations Available for All New LessWrong Posts by Solenoid Entity
|
Jul 21, 2023 |
LW - Even Superhuman Go AIs Have Surprising Failures Modes by AdamGleave
|
Jul 20, 2023 |
LW - Does Circuit Analysis Interpretability Scale? Evidence from Multiple Choice Capabilities in Chinchilla by Neel Nanda
|
Jul 20, 2023 |
LW - Still no Lie Detector for LLMs by Whispermute
|
Jul 19, 2023 |
LW - Alignment Grantmaking is Funding-Limited Right Now by johnswentworth
|
Jul 19, 2023 |
LW - A brief history of computers by Adam Zerner
|
Jul 19, 2023 |
LW - Tiny Mech Interp Projects: Emergent Positional Embeddings of Words by Neel Nanda
|
Jul 19, 2023 |
LW - Meta announces Llama 2; "open sources" it for commercial use by LawrenceC
|
Jul 19, 2023 |
LW - Measuring and Improving the Faithfulness of Model-Generated Reasoning by Ansh Radhakrishnan
|
Jul 18, 2023 |
LW - Proof of posteriority: a defense against AI-generated misinformation by jchan
|
Jul 18, 2023 |
LW - Existential Risk Persuasion Tournament by PeterMcCluskey
|
Jul 18, 2023 |
LW - Predictive history classes by dkl9
|
Jul 18, 2023 |
LW - AutoInterpretation Finds Sparse Coding Beats Alternatives by Hoagy
|
Jul 17, 2023 |
LW - Sapient Algorithms by Valentine
|
Jul 17, 2023 |
LW - Thoughts on "Process-Based Supervision" by Steven Byrnes
|
Jul 17, 2023 |
LW - An upcoming US Supreme Court case may impede AI governance efforts by NickGabs
|
Jul 17, 2023 |
LW - Mech Interp Puzzle 1: Suspiciously Similar Embeddings in GPT-Neo by Neel Nanda
|
Jul 17, 2023 |
LW - Activation adding experiments with llama-7b by NinaR
|
Jul 16, 2023 |
LW - Robustness of Model-Graded Evaluations and Automated Interpretability by Simon Lermen
|
Jul 16, 2023 |
LW - Why was the AI Alignment community so unprepared for this moment? by Ras1513
|
Jul 15, 2023 |
LW - When Someone Tells You They're Lying, Believe Them by ymeskhout
|
Jul 14, 2023 |
LW - The Goddess of Everything Else - The Animation by Writer
|
Jul 13, 2023 |
LW - Jailbreaking GPT-4's code interpreter by nikolaisalreadytaken
|
Jul 13, 2023 |
LW - How can I get help becoming a better rationalist? by TeaTieAndHat
|
Jul 13, 2023 |
LW - Winners of AI Alignment Awards Research Contest by Akash
|
Jul 13, 2023 |
LW - Are there any good, easy-to-understand examples of cases where statistical causal network discovery worked well in practice? by tailcalled
|
Jul 13, 2023 |
LW - Accidentally Load Bearing by jefftk
|
Jul 13, 2023 |
LW - Elon Musk announces xAI by Jan Kulveit
|
Jul 13, 2023 |
LW - What does the launch of x.ai mean for AI Safety? by Chris Leong
|
Jul 13, 2023 |
LW - Alignment Megaprojects: You're Not Even Trying to Have Ideas by NicholasKross
|
Jul 13, 2023 |
LW - Report on modeling evidential cooperation in large worlds by Johannes Treutlein
|
Jul 12, 2023 |
LW - A review of Principia Qualia by jessicata
|
Jul 12, 2023 |
LW - Towards Developmental Interpretability by Jesse Hoogland
|
Jul 12, 2023 |
LW - A transcript of the TED talk by Eliezer Yudkowsky by Mikhail Samin
|
Jul 12, 2023 |
LW - My Weirdest Experience by Bridgett Kay
|
Jul 12, 2023 |
LW - Why it's necessary to shoot yourself in the foot by g-w1
|
Jul 12, 2023 |
LW - Aging and the geroscience hypothesis by DirectedEvolution
|
Jul 12, 2023 |
LW - Introducing Fatebook: the fastest way to make and track predictions by Adam B
|
Jul 11, 2023 |
LW - OpenAI Launches Superalignment Taskforce by Zvi
|
Jul 11, 2023 |
LW - [UPDATE: deadline extended to July 24!] New wind in rationality’s sails: Applications for Epistea Residency 2023 are now open by Jana Meixnerová
|
Jul 11, 2023 |
LW - Drawn Out: a story by Richard Ngo
|
Jul 11, 2023 |
LW - Open-minded updatelessness by Nicolas Macé
|
Jul 11, 2023 |
LW - I think Michael Bailey's dismissal of my autogynephilia questions for Scott Alexander and Aella makes very little sense by tailcalled
|
Jul 10, 2023 |
LW - “Reframing Superintelligence” + LLMs + 4 years by Eric Drexler
|
Jul 10, 2023 |
LW - Consider Joining the UK Foundation Model Taskforce by Zvi
|
Jul 10, 2023 |
LW - Consciousness as a conflationary alliance term by Andrew Critch
|
Jul 10, 2023 |
LW - Some reasons to not say "Doomer" by Ruby
|
Jul 10, 2023 |
LW - The Seeker’s Game - Vignettes from the Bay by Yulia
|
Jul 09, 2023 |
LW - Really Strong Features Found in Residual Stream by Logan Riggs
|
Jul 09, 2023 |
LW - Taboo Truth by Tomás B.
|
Jul 09, 2023 |
LW - Fixed Point: a love story by Richard Ngo
|
Jul 08, 2023 |
LW - Views on when AGI comes and on strategy to reduce existential risk by TsviBT
|
Jul 08, 2023 |
LW - What Does LessWrong/EA Think of Human Intelligence Augmentation as of mid-2023? by marc/er
|
Jul 08, 2023 |
LW - Passing the ideological Turing test? Arguments against existential risk from AI. by NinaR
|
Jul 07, 2023 |
LW - Apparently, of the 195 Million the DoD allocated in University Research Funding Awards in 2022, more than half of them concerned AI or compute hardware research by mako yass
|
Jul 07, 2023 |
LW - Introducing bayescalc.io by Adele Lopez
|
Jul 07, 2023 |
LW - What are the best non-LW places to read on alignment progress? by Raemon
|
Jul 07, 2023 |
LW - Jesse Hoogland on Developmental Interpretability and Singular Learning Theory by Michaël Trazzi
|
Jul 06, 2023 |
LW - Agency begets agency (the world is malleable) by Richard Ngo
|
Jul 06, 2023 |
LW - AI #19: Hofstadter, Sutskever, Leike by Zvi
|
Jul 06, 2023 |
LW - AI labs' statements on governance by Zach Stein-Perlman
|
Jul 06, 2023 |
LW - Optimized for Something other than Winning or: How Cricket Resists Moloch and Goodhart's Law by A.H.
|
Jul 06, 2023 |
LW - (tentatively) Found 600+ Monosemantic Features in a Small LM Using Sparse Autoencoders by Logan Riggs
|
Jul 05, 2023 |
LW - [Linkpost] Introducing Superalignment by beren
|
Jul 05, 2023 |
LW - Dominant Assurance Contract Experiment #2: Berkeley House Dinners by Arjun Panickssery
|
Jul 05, 2023 |
LW - Twitter Twitches by Zvi
|
Jul 05, 2023 |
LW - Ways I Expect AI Regulation To Increase Extinction Risk by 1a3orn
|
Jul 04, 2023 |
LW - Two Percolation Puzzles by Adam Scherlis
|
Jul 04, 2023 |
LW - The literature on aluminum adjuvants is very suspicious. Small IQ tax is plausible - can any experts help me estimate it? by mikes
|
Jul 04, 2023 |
LW - Ten Levels of AI Alignment Difficulty by Sammy Martin
|
Jul 04, 2023 |
LW - Monthly Roundup #8: July 2023 by Zvi
|
Jul 04, 2023 |
LW - Frames in context by Richard Ngo
|
Jul 03, 2023 |
LW - Meta-rationality and frames by Richard Ngo
|
Jul 03, 2023 |
LW - Douglas Hofstadter changes his mind on Deep Learning and AI risk (June 2023)? by gwern
|
Jul 03, 2023 |
LW - Going Crazy and Getting Better Again by Evenstar
|
Jul 02, 2023 |
LW - Why it's so hard to talk about Consciousness by Rafael Harth
|
Jul 02, 2023 |
LW - Consider giving money to people, not projects or organizations by NinaR
|
Jul 02, 2023 |
LW - Micro Habits that Improve One’s Day by silentbob
|
Jul 02, 2023 |
LW - Forum Karma: view stats and find highly-rated comments for any LW user by Max H
|
Jul 02, 2023 |
LW - Alpha by Erich Grunewald
|
Jul 02, 2023 |
LW - Grant applications and grand narratives by Elizabeth
|
Jul 02, 2023 |
LW - My "trauma" frame by Raemon
|
Jul 01, 2023 |
LW - On household dust by NinaR
|
Jun 30, 2023 |
LW - Introducing EffiSciences’ AI Safety Unit by WCargo
|
Jun 30, 2023 |
LW - I Think Eliezer Should Go on Glenn Beck by Lao Mein
|
Jun 30, 2023 |
LW - The Case for Overconfidence is Overstated by Kevin Dorst
|
Jun 30, 2023 |
LW - AI #18: The Great Debate Debate by Zvi
|
Jun 29, 2023 |
LW - Anthropically Blind: the anthropic shadow is reflectively inconsistent by Christopher King
|
Jun 29, 2023 |
LW - A "weak" AGI may attempt an unlikely-to-succeed takeover by RobertM
|
Jun 29, 2023 |
LW - What money-pumps exist, if any, for deontologists? by Daniel Kokotajlo
|
Jun 29, 2023 |
LW - My research agenda in agent foundations by Alex Altair
|
Jun 28, 2023 |
LW - When do "brains beat brawn" in Chess? An experiment by titotal
|
Jun 28, 2023 |
LW - Nature: "Stop talking about tomorrow’s AI doomsday when AI poses risks today" by Ben Smith
|
Jun 28, 2023 |
LW - Carl Shulman on The Lunar Society (7 hour, two-part podcast) by ESRogs
|
Jun 28, 2023 |
LW - Self-Blinded Caffeine RCT by niplav
|
Jun 28, 2023 |
LW - On the Cost of Thriving Index by Zvi
|
Jun 27, 2023 |
LW - Munk AI debate: confusions and possible cruxes by Steven Byrnes
|
Jun 27, 2023 |
LW - Man in the Arena by Richard Ngo
|
Jun 27, 2023 |
LW - AI-Plans.com - a contributable compendium by Iknownothing
|
Jun 27, 2023 |
LW - 60+ Possible Futures by Stuckwork
|
Jun 27, 2023 |
LW - "Safety Culture for AI" is important, but isn't going to be easy by Davidmanheim
|
Jun 26, 2023 |
LW - Model, Care, Execution by Ricki Heicklen
|
Jun 26, 2023 |
LW - Why am I Me? by dadadarren
|
Jun 26, 2023 |
LW - Another medical miracle by Dentin
|
Jun 25, 2023 |
LW - Crystal Healing — or the Origins of Expected Utility Maximizers by Alexander Gietelink Oldenziel
|
Jun 25, 2023 |
LW - Did Bengio and Tegmark lose a debate about AI x-risk against LeCun and Mitchell? by Karl von Wendt
|
Jun 25, 2023 |
LW - Correctly Calibrated Trust by habryka
|
Jun 25, 2023 |
LW - Will the growing deer prion epidemic spread to humans? Why not? by eukaryote
|
Jun 25, 2023 |
LW - How tall is the Shard, really? by philh
|
Jun 24, 2023 |
LW - Automatic Rate Limiting on LessWrong by Raemon
|
Jun 23, 2023 |
LW - When is correlation transitive? by Ege Erdil
|
Jun 23, 2023 |
LW - Catastrophic Risks from AI #1: Summary by Dan H
|
Jun 23, 2023 |
LW - Catastrophic Risks from AI #2: Malicious Use by Dan H
|
Jun 23, 2023 |
LW - "textbooks are all you need" by bhauth
|
Jun 22, 2023 |
LW - The Hubinger lectures on AGI safety: an introductory lecture series by evhub
|
Jun 22, 2023 |
LW - Which personality traits are real? Stress-testing the lexical hypothesis by tailcalled
|
Jun 22, 2023 |
LW - Short timelines and slow, continuous takeoff as the safest path to AGI by rosehadshar
|
Jun 21, 2023 |
LW - My side of an argument with Jacob Cannell about chip interconnect losses by Steven Byrnes
|
Jun 21, 2023 |
LW - My tentative best guess on how EAs and Rationalists sometimes turn crazy by habryka
|
Jun 21, 2023 |
LW - Never Fight The Last War by ChristianKl
|
Jun 21, 2023 |
LW - Public Transit is not Infinitely Safe by jefftk
|
Jun 20, 2023 |
LW - Causality: A Brief Introduction by tom4everitt
|
Jun 20, 2023 |
LW - A Friendly Face (Another Failure Story) by Karl von Wendt
|
Jun 20, 2023 |
LW - Experiments in Evaluating Steering Vectors by Gytis Daujotas
|
Jun 20, 2023 |
LW - Ban development of unpredictable powerful models? by TurnTrout
|
Jun 20, 2023 |
LW - 10 quick takes about AGI by Max H
|
Jun 20, 2023 |
LW - resolving some neural network mysteries by bhauth
|
Jun 20, 2023 |
LW - Mode collapse in RL may be fueled by the update equation by TurnTrout
|
Jun 20, 2023 |
LW - Lessons On How To Get Things Right On The First Try by johnswentworth
|
Jun 20, 2023 |
LW - My impression of singular learning theory by Ege Erdil
|
Jun 19, 2023 |
LW - Guide to rationalist interior decorating by mingyuan
|
Jun 19, 2023 |
LW - Solomonoff induction still works if the universe is uncomputable, and its usefulness doesn't require knowing Occam's razor by Christopher King
|
Jun 18, 2023 |
LW - A summary of current work in AI governance by constructive
|
Jun 18, 2023 |
LW - UK Foundation Model Task Force - Expression of Interest by ojorgensen
|
Jun 18, 2023 |
LW - Cryonics Career Survey (more jobs than you think) by Mati Roy
|
Jun 18, 2023 |
LW - Adventist Health Study-2 supports pescetarianism more than veganism by Elizabeth
|
Jun 18, 2023 |
LW - [Replication] Conjecture's Sparse Coding in Small Transformers by Hoagy
|
Jun 17, 2023 |
LW - Are Bayesian methods guaranteed to overfit? by Ege Erdil
|
Jun 17, 2023 |
LW - The ones who endure by Richard Ngo
|
Jun 17, 2023 |
LW - Updating Drexler's CAIS model by Matthew Barnett
|
Jun 17, 2023 |
LW - LLMs Sometimes Generate Purely Negatively-Reinforced Text by Fabien Roger
|
Jun 16, 2023 |
LW - Distilling Singular Learning Theory by Liam Carroll
|
Jun 16, 2023 |
LW - Rational Animations is looking for an AI Safety scriptwriter, a lead community manager, and other roles. by Writer
|
Jun 16, 2023 |
LW - Leveling Up Or Leveling Off? Understanding The Science Behind Skill Plateaus by lynettebye
|
Jun 16, 2023 |
LW - Conjecture: A standing offer for public debates on AI by Andrea Miotti
|
Jun 16, 2023 |
LW - Looking Back On Ads by jefftk
|
Jun 15, 2023 |
LW - I still think it's very unlikely we're observing alien aircraft by dynomight
|
Jun 15, 2023 |
LW - Why libertarians are advocating for regulation on AI by RobertM
|
Jun 15, 2023 |
LW - On the Apple Vision Pro by Zvi
|
Jun 15, 2023 |
LW - Anthropic | Charting a Path to AI Accountability by Gabriel Mukobi
|
Jun 14, 2023 |
LW - Contingency: A Conceptual Tool from Evolutionary Biology for Alignment by clem acs
|
Jun 14, 2023 |
LW - Multiple stages of fallacy - justifications and non-justifications for the multiple stage fallacy by AronT
|
Jun 14, 2023 |
LW - Lightcone Infrastructure is looking for funding by habryka
|
Jun 14, 2023 |
LW - MetaAI: less is less for alignment. by Cleo Nardo
|
Jun 14, 2023 |
LW - My guess for why I was wrong about US housing by romeostevensit
|
Jun 14, 2023 |
LW - <$750k grants for General Purpose AI Assurance/Safety Research by Phosphorous
|
Jun 13, 2023 |
LW - TASRA: A Taxonomy and Analysis of Societal-Scale Risks from AI by Andrew Critch
|
Jun 13, 2023 |
LW - The Dial of Progress by Zvi
|
Jun 13, 2023 |
LW - UFO Betting: Put Up or Shut Up by RatsWrongAboutUAP
|
Jun 13, 2023 |
LW - If you are too stressed, walk away from the front lines by Neil Warren
|
Jun 13, 2023 |
LW - Introduction to Towards Causal Foundations of Safe AGI by tom4everitt
|
Jun 13, 2023 |
LW - UK PM: $125M for AI safety by Hauke Hillebrandt
|
Jun 12, 2023 |
LW - Manifold Predicted the AI Extinction Statement and CAIS Wanted it Deleted by David Chee
|
Jun 12, 2023 |
LW - ARC is hiring theoretical researchers by paulfchristiano
|
Jun 12, 2023 |
LW - Critiques of prominent AI safety labs: Conjecture by Omega.
|
Jun 12, 2023 |
LW - [Linkpost] Large Language Models Converge on Brain-Like Word Representations by Bogdan Ionut Cirstea
|
Jun 12, 2023 |
LW - I can see how I am Dumb by Johannes C. Mayer
|
Jun 11, 2023 |
LW - Inference-Time Intervention: Eliciting Truthful Answers from a Language Model by likenneth
|
Jun 11, 2023 |
LW - The Dictatorship Problem by alyssavance
|
Jun 11, 2023 |
LW - Ethodynamics of Omelas by dr s
|
Jun 11, 2023 |
LW - How biosafety could inform AI standards by Olivia Jimenez
|
Jun 09, 2023 |
LW - A plea for solutionism on AI safety by jasoncrawford
|
Jun 09, 2023 |
LW - InternLM - China's Best (Unverified) by Lao Mein
|
Jun 09, 2023 |
LW - Updates and Reflections on Optimal Exercise after Nearly a Decade by romeostevensit
|
Jun 09, 2023 |
LW - Takeaways from the Mechanistic Interpretability Challenges by scasper
|
Jun 08, 2023 |
LW - A moral backlash against AI will probably slow down AGI development by geoffreymiller
|
Jun 08, 2023 |
LW - AI #15: The Principle of Charity by Zvi
|
Jun 08, 2023 |
LW - An Exercise to Build Intuitions on AGI Risk by Lauro Langosco
|
Jun 08, 2023 |
LW - LEAst-squares Concept Erasure (LEACE) by tricky labyrinth
|
Jun 08, 2023 |
LW - What will GPT-2030 look like? by jsteinhardt
|
Jun 08, 2023 |
LW - Elon talked with senior Chinese leadership about AI X-risk by ChristianKl
|
Jun 07, 2023 |
LW - Why I am not a longtermist (May 2022) by boazbarak
|
Jun 07, 2023 |
LW - Transformative AGI by 2043 is <1% likely by Ted Sanders
|
Jun 07, 2023 |
LW - Cultivate an obsession with the object level (the world is fascinating) by Richard Ngo
|
Jun 07, 2023 |
LW - The Sharp Right Turn: sudden deceptive alignment as a convergent goal by avturchin
|
Jun 07, 2023 |
LW - Launching Lightspeed Grants (Apply by July 6th) by habryka
|
Jun 07, 2023 |
LW - Society Library seeking contributions for canonical AI Safety debate map by Jarred Filmer
|
Jun 07, 2023 |
LW - The Base Rate Times, news through prediction markets by vandemonian
|
Jun 06, 2023 |
LW - Andrew Ng wants to have a conversation about extinction risk from AI by Leon Lang
|
Jun 06, 2023 |
LW - A Playbook for AI Risk Reduction (focused on misaligned AI) by HoldenKarnofsky
|
Jun 06, 2023 |
LW - Intelligence Officials Say U.S. Has Retrieved Craft of Non-Human Origin by lc
|
Jun 06, 2023 |
LW - One implementation of regulatory GPU restrictions by porby
|
Jun 06, 2023 |
LW - How to Think About Activation Patching by Neel Nanda
|
Jun 06, 2023 |
LW - The (local) unit of intelligence is FLOPs by boazbarak
|
Jun 06, 2023 |
LW - AISafety.info "How can I help?" FAQ by steven0461
|
Jun 06, 2023 |
LW - Algorithmic Improvement Is Probably Faster Than Scaling Now by johnswentworth
|
Jun 06, 2023 |
LW - We Are Less Wrong than E. T. Jaynes on Loss Functions in Human Society by Zack M Davis
|
Jun 05, 2023 |
LW - Meta-conversation shouldn't be taboo by Adam Zerner
|
Jun 05, 2023 |
LW - Wildfire of strategicness by TsviBT
|
Jun 05, 2023 |
LW - Ages Survey: Results by jefftk
|
Jun 05, 2023 |
LW - The ants and the grasshopper by Richard Ngo
|
Jun 04, 2023 |
LW - Announcing AISafety.info's Write-a-thon (June 16-18) and Second Distillation Fellowship (July 3-October 2) by steven0461
|
Jun 04, 2023 |
LW - The Control Problem: Unsolved or Unsolvable? by Remmelt
|
Jun 04, 2023 |
LW - A Double-Feature on The Extropians by Maxwell Tabarrok
|
Jun 03, 2023 |
LW - The AGI Race Between the US and China Doesn’t Exist. by Eva B
|
Jun 03, 2023 |
LW - A mind needn't be curious to reap the benefits of curiosity by So8res
|
Jun 02, 2023 |
LW - Co-found an incubator for independent AI Safety researchers by Alexandra Bos
|
Jun 02, 2023 |
LW - Dreams of "Mathopedia" by NicholasKross
|
Jun 02, 2023 |
LW - Think carefully before calling RL policies "agents" by TurnTrout
|
Jun 02, 2023 |
LW - Uncertainty about the future does not imply that AGI will go well by Lauro Langosco
|
Jun 02, 2023 |
LW - Yudkowsky vs Hanson on FOOM: Whose Predictions Were Better? by 1a3orn
|
Jun 02, 2023 |
LW - Things I Learned by Spending Five Thousand Hours In Non-EA Charities by jenn
|
Jun 01, 2023 |
LW - Change my mind: Veganism entails trade-offs, and health is one of the axes by Elizabeth
|
Jun 01, 2023 |
LW - Short Remark on the (subjective) mathematical 'naturalness' of the Nanda--Lieberum addition modulo 113 algorithm by Spencer Becker-Kahn
|
Jun 01, 2023 |
LW - Work dumber not smarter by lukehmiles
|
Jun 01, 2023 |
LW - The challenge of articulating tacit knowledge by NinaR
|
Jun 01, 2023 |
LW - The Crux List by Zvi
|
May 31, 2023 |
LW - To Predict What Happens, Ask What Happens by Zvi
|
May 31, 2023 |
LW - Cosmopolitan values don't come free by So8res
|
May 31, 2023 |
LW - Contrast Pairs Drive the Empirical Performance of Contrast Consistent Search (CCS) by Scott Emmons
|
May 31, 2023 |
LW - Product Endorsement: Food for sleep interruptions by Elizabeth
|
May 31, 2023 |
LW - The case for removing alignment and ML research from the training dataset by beren
|
May 31, 2023 |
LW - PaLM-2 and GPT-4 in "Extrapolating GPT-N performance" by Lukas Finnveden
|
May 30, 2023 |
LW - Announcing Apollo Research by Marius Hobbhahn
|
May 30, 2023 |
LW - Reply to a fertility doctor concerning polygenic embryo screening by GeneSmith
|
May 30, 2023 |
LW - Sentience matters by So8res
|
May 29, 2023 |
LW - Wikipedia as an introduction to the alignment problem by SoerenMind
|
May 29, 2023 |
LW - Gemini will bring the next big timeline update by p.b.
|
May 29, 2023 |
LW - Kelly betting vs expectation maximization by MorgneticField
|
May 29, 2023 |
LW - Morality is Accidental and Self-Congratulatory by ymeskhout
|
May 29, 2023 |
LW - TinyStories: Small Language Models That Still Speak Coherent English by Ulisse Mini
|
May 29, 2023 |
LW - DandD.Sci 5E: Return of the League of Defenders by aphyer
|
May 28, 2023 |
LW - You now can enable reacts on your own posts! (and other react info) by Ruby
|
May 28, 2023 |
LW - Self-administered EMDR without a therapist is very useful for a lot of things! by Anton Rodenhauser
|
May 27, 2023 |
LW - Should Rational Animations invite viewers to read content on LessWrong? by Writer
|
May 27, 2023 |
LW - Why I'm Not (Yet) A Full-Time Technical Alignment Researcher by NicholasKross
|
May 26, 2023 |
LW - Request: stop advancing AI capabilities by So8res
|
May 26, 2023 |
LW - Book Review: How Minds Change by bc4026bd4aaa5b7fe
|
May 26, 2023 |
LW - Mob and Bailey by Screwtape
|
May 26, 2023 |
LW - [Linkpost] Interpretability Dreams by DanielFilan
|
May 25, 2023 |
LW - Look At What's In Front Of You (Conclusion to The Nuts and Bolts of Naturalism) by LoganStrohl
|
May 25, 2023 |
LW - Solving the Mechanistic Interpretability challenges: EIS VII Challenge 2 by StefanHex
|
May 25, 2023 |
LW - DeepMind: Model evaluation for extreme risks by Zach Stein-Perlman
|
May 25, 2023 |
LW - Adumbrations on AGI from an outsider by nicholashalden
|
May 25, 2023 |