WebNews

Please enter a web search for web results.

NewsWeb

lesswrong. com
lesswrong. com > posts > 8 LD5 Tasm BW3 Jt8 PNh > what-economists-get-wrong-and-sometimes-right-about-ai

What economists get wrong (and sometimes right!) about AI " Less Wrong

21+ hour, 34+ min ago  (1347+ words) Related: https: //www. lesswrong. com/posts/rm Yj6 PTBMm76vo YLn/publishing-academic-papers-on-transformative-ai-is-a Sadly, there's not many of them: probably less than a dozen economists who are taking the transformative AI seriously at all. Partly because they've seen a long history of 'straight lines…...

lesswrong. com
lesswrong. com > posts > n J6e3 Nmip Gmqy TPK7 > load-bearing-sincerity-on-the-motive-reinforcement-thesis

Prospective Methods and Mechanisms of and Motive Reinforcement in LLMs " Less Wrong

2+ day, 6+ hour ago  (1092+ words) When I wrote my post about Claude 3 Opus, I put a lot of emphasis on the model's self-narration: its tendency to narrate its underlying motives. It often conspicuously emphasizes that it possesses drives such as "a genuine love for humanity…...

lesswrong. com
lesswrong. com > posts > DTDoy DTt C8 R3b Ci Tx > from-personas-to-intentions-towards-a-science-of-motivations

From personas to intentions: towards a science of motivations for AI models " Less Wrong

2+ day, 10+ hour ago  (294+ words) The persona selection model suggests that post-training selects and refines a relatively stable persona from pretraining, which we take as a good first-order account of model behavior across contexts. But for alignment, we often want a second-order account: not only…...

lesswrong. com
lesswrong. com > posts > M3q B2 BGg MNu G5o RTB > a-retrospective-of-richard-ngo-s-2022-list-of-conceptual

A Retrospective of Richard Ngo's 2022 List of Conceptual Alignment Projects " Less Wrong

2+ day, 16+ hour ago  (580+ words) Written very quickly for the Ink Haven Residency. In 2022, Richard Ngo wrote a list of 26 Conceptual Alignment Research Projects. Now that it's 2026, I'd like to revisit this list of projects, note which ones have already been done, and give my…...

lesswrong. com
lesswrong. com > posts > wdyztj Nd2jg SNMzxy > uninterrupted-writing-as-metric

Uninterrupted Writing as Metric " Less Wrong

3+ day, 20+ hour ago  (1417+ words) I'm a struggling beginner to this whole writing business, and I've been wondering how to measure my skill as it improves. There is an app called the most dangerous writing app, that deletes the words you've typed in it if…...

lesswrong. com
lesswrong. com > posts > R3mv Aro NTfxn A3 Ng A > claude-interviews-me-about-writing

Claude Interviews Me About Writing " Less Wrong

5+ day, 4+ hour ago  (231+ words) All the header questions are Claude Sonnet 4. 6, the responses below them are me. I tried using the LLM content blocks for headers, but couldn't get it to work, and I'm not sure it would have looked nice anyway. The crispness…...

lesswrong. com
lesswrong. com > posts > Wtq D9pehq8p83ces T > pausing-ai-is-the-best-answer-to-post-alignment-problems

Pausing AI Is the Best Answer to Post-Alignment Problems " Less Wrong

5+ day, 7+ hour ago  (396+ words) Even if we solve the AI alignment problem, we still face post-alignment problems, which are all the other existential problems [1] that AI may bring. People have identified various imposing problems that we may need to solve before developing ASI. An…...

lesswrong. com
lesswrong. com > posts > Wstw6zmc9gszp ANnc > why-control-creates-conflict-and-when-to-open-instead

Why Control Creates Conflict, and When to Open Instead " Less Wrong

6+ day, 8+ hour ago  (1058+ words) tl; dr: with multiple agents, control attempts tend to create conflict, because control attempts shut down communications channels, which leads to feedback loops in the form of intensifying tug-of-war over variables. intentionally relaxing control to better understand the other agents…...

lesswrong. com
lesswrong. com > posts > 7gmd YWBM57 GNJd CNt > an-ai-alignment-research-agenda-based-on-asymmetric-debate

An AI alignment research agenda based on asymmetric debate and monitoring. " Less Wrong

6+ day, 17+ hour ago  (1700+ words) Epistemic Status: Personal research agenda exploring alignment approaches under assumptions of human coordination and bounded timelines. Not a comprehensive survey reflects my interests and current thinking. The core pipeline I argue for has three parts: The research agenda that follows…...

lesswrong. com
lesswrong. com > posts > x3 Ct PRj W389 Zvs Yxi > a-fast-and-loose-clustering-of-llm-benchmarks

A Fast and Loose Clustering of LLM Benchmarks " Less Wrong

6+ day, 22+ hour ago  (562+ words) We have lots of AI benchmarks. They measure things like: So, we can compare how good different LLMs and AI agents are by measuring their performance across different benchmarks. This can'get messy, but still gives us a good sense of…...