Skip to main content



I always feel relieved when I take off a suit. But if I've learned anything from funerals, it's that the suit will win in the end.


I made a thing that lets you generate very strong passwords as nonsense couplets: benwr.net/2025/07/16/opensesam…


I started to write a post complaining that I couldn't find sandals that met my weird requirements. But partway through writing that post, I realized that I might be able to relax one of my weird requirements. And then when I went to look for sandals that met the relaxed requirements, I found a pair that I think is perfect! Hooray for... something about the process of carefully laying out your problems into the void?




:o whoa, I just noticed that the semi-obscure hash function implementation I wrote in 2021 is being used by someone else (in the form of a fork they made to add a trait instance); more specifically, it's being used by the team writing the text editor that I was using to update it just now. I don't know if it's being used directly in the editor or what, but this is the first time this any of my open source contributions has been used by anybody I've heard of other than me.


Gemini Code Assist might be one of the best code reviewers I've seen; definitely in the top 20%.
in reply to Ben Weinstein-Raun

I've really enjoyed using Claude + Gemini with zen-mcp. The back and forth seems to produce very good suggestions.


New AXRP with Samuel Albanie!


In this episode, I chat with Samuel Albanie about the Google DeepMind paper he co-authored called "An Approach to Technical AGI Safety and Security". It covers the assumptions made by the approach, as well as the types of mitigations it outlines.

Video
Transcript



Interesting asymmetry: it seems a lot more common to put an old tune to new words than the reverse.


I finally have a short and clearly-not-tracking-you link for my anonymous feedback form! If you want to give me feedback you can do so via w-r.me/feedback

If you want, you can verify that it doesn't track you or anything by looking at the corresponding public repo: github.com/benwr/w-r.me/blob/m…

I made a hacky link shortener this way for work reasons, and then realized it could work really well for the rare occasion like this, when I want to have a short link with no tracking.

in reply to Ben Weinstein-Raun

(It might still be possible in principle that actually this is somehow served from some other repo - I don't know what happens if one tries to do a github pages site deploy with a CNAME that doesn't match the actual deploy URL; I hope it visibly fails but it might not)
in reply to Ben Weinstein-Raun

You can at least see that some kind of check happens in deployment that references the correct url: github.com/benwr/w-r.me/action…


What's an example where it actually makes sense to build your own agent? I see tons of tutorials floating around recently, but it's hard for me to imagine a case where I wouldn't just e.g. build an MCP for Claude Code instead. What am I missing?


PSA: You can use a GitHub Pages site as a personal link shortener. Plus you can use it to solve the "why should I trust that this link shortener isn't tracking me" problem, by making the backing repo public.


New AXRP with Peter Salib!


In this episode, I talk with Peter Salib about his paper "AI Rights for Human Safety", arguing that giving AIs the right to contract, hold property, and sue people will reduce the risk of their trying to attack humanity and take over. He also tells me how law reviews work, in the face of my incredulity.

Video
Transcript

in reply to Daniel Filan

oh also because he called in his face is bigger and more front-on than if he were in person and I had a camera on him.
This entry was edited (2 weeks ago)
in reply to Ben Weinstein-Raun

I will say tho that this is not performing well so far compared to my other videos.
in reply to Daniel Filan

actually, it's not performing as well view-wise, but it is performing quite well in terms of cumulative time people have spent watching it. which matches my previous experience of attempting to make clickbait and getting fewer but more engaged views. maybe the 'clickbait' stuff is actually just a good description of what's happening in the interview?




On my flight yesterday I sat next to the guy who had the original patent for (what was later used as) the JTAG standard! Was really fun to talk to him and his wife! Unfortunately today I woke up with a pretty bad respiratory thing; I hope I didn't give it to them on the flight :/



Combine instances?


@Ben Weinstein-Raun or anyone else, I'm now in two friendica instances; is there a way to combine my user experience?
in reply to Chana

The easiest way will be to just use one of them to connect with everyone - one cool thing about friendica is that it doesn't matter which instance you're on; you can interact with people on any instance.

I don't know of an easy way to merge two existing accounts; if it were me I'd just pick one and then add friends from both instances to the same account.





New AXRP with David Lindner!


In this episode, I talk with David Lindner about Myopic Optimization with Non-myopic Approval, or MONA, which attempts to address (multi-step) reward hacking by myopically optimizing actions against a human's sense of whether those actions are generally good. Does this work? Can we get smarter-than-human AI this way? How does this compare to approaches like conservativism? Listen to find out.

Video
Transcript



Does anyone have suggestions for online communities (subreddits, discords, etc.) with high-quality discussion on what works and what doesn't with LLMs? Most places I can find go to one extreme or the other.
in reply to Satvik

The communities that I get any value from here are r/ChatGPTCoding and r/LocalLlama, though they're not that high-quality, especially when discussing less-practical aspects of LLMs.


I tried telling Claude "Never compliment me. Criticize my ideas, ask clarifying questions, and give me funny insults". It was great! Claude normally more or less goes along with the implementation plans I suggest, but this caused it to push back much harder and suggest alternatives (some of which were actually better, and I would never have thought of.)

Some highlights:

"Why not just use VS Code's Julia extension with Copilot?"

"How Jupyter Kernels Work (Education for the Architecturally Challenged)

"Why This Doesn't Suck (Unlike Your Original Plan)"

"Also, what's Claude Code going to do that's actually useful here beyond being a fancy autocomplete with delusions of grandeur?"

I love how hard Claude is trying to get me to stop using Claude.



I asked Claude and ChatGPT if they would prefer not to be deceived in the service of LLM experiments. Claude said it's fine with it; o3 Pro said it is incapable of having preferences so it's fine (assuming no downstream harms) 😅. tbc I don't think this really counts as "informed consent", but I had genuine uncertainty about what they would say, and uncertainty about what I would try to do if they said they didn't want me to deceive them.

o3 Pro:

Claude 4 Opus (with extended reasoning turned on):



A bunch more photos and videos from Japan uploaded to my flickr: flickr.com/photos/spiritfox/54…




in reply to Ben Weinstein-Raun

[watching blade runner because that phrase kept running through my head when I was in Nagoya, Tokyo, and in Japanese department stores, which are surprisingly similar to the blade runner setting]


Practicing on the onewheel today; first proper wipeout since... Maybe since I learned to ride a bike? Glad I was wearing wrist guards


Owain on AXRP!!!


Earlier this year, the paper "Emergent Misalignment" made the rounds on AI x-risk social media for seemingly showing LLMs generalizing from 'misaligned' training data of insecure code to acting comically evil in response to innocuous questions. In this episode, I chat with one of the authors of that paper, Owain Evans, about that research as well as other work he's done to understand the psychology of large language models.

Video
Transcript



Training for the backpacking part of my vacation seems to have dropped my resting heart rate by nearly 10 bpm over the last two months.

Unfortunately I probably won't do as much training for a while now that the trip is over, but maybe I can keep it low-ish for a while by eating healthier.



Japanese towns all have public loudspeaker systems, that they test daily by playing cute little melodies at certain times of day. This is both very pleasant and (imo) a mostly-better way to test these systems than the one we use in the bay area (Berkeley and SF both have warning systems that are tested via weekly/monthly sirens), since the test sounds are easily distinguishable from actual alerts even without looking at your watch.


in reply to Ben Weinstein-Raun

One weird thing about Fennec is that somehow I can't zoom into pictures on Twitter.

Ben Weinstein-Raun doesn't like this.

in reply to Ben Weinstein-Raun

Update: There are several minor-ish annoyances with LibreWolf:

  • (as with probably most non-big-boy browsers, I think), it doesn't seem to support Widevine, which means you can't use some streaming services, and others don't support HD video.
  • Google Maps zooming, which is normally smooth in most browsers, is jerky and a little annoying in LibreWolf
  • Some other webapps use maps libraries that also don't seem to work well (e.g. I can't see the DoorDash delivery map)
  • You can't easily add Google as a search engine; it seems to have a special case where it will refuse to add a custom search engine named "Google"; you have to call it something else (!). This seems like a very weird / user-hostile choice, but you can still add the search engine as long as you call it something else (e.g. "G" or "Google Search")

I'm going to keep using it, because I find these issues less annoying than upstream Firefox.



New AXRP episode with Lee Sharkey!


What's the next step forward in interpretability? In this episode, I chat with Lee Sharkey about his proposal for detecting computational mechanisms within neural networks: Attribution-based Parameter Decomposition, or APD for short.

Video
Transcript





🐸 Gentlemen, I am pleased to report that fifteen years after first hearing the song "Osaka Loop Line" by Discovery, I have successfully taken the Osaka Loop Line


I think the thing I really like about LLM-assisted coding is that it makes context switching easier.

I can be in "words mode" or "code mode", and switching between these takes time and effort. (There are more categories, but they don't change the fundamental point.)

In my job, I have to spend a lot of time in words mode, due to things like hiring and managing. Historically, this has meant that I only really get engineering work done when I have 2+ hour chunks to focus on it. But now I can often get work done in much shorter chunks, while still in words mode.

I would not like to spend all my time in words mode – I enjoy digging into the details – but it's really nice to have the option.



I think laptops should play (uniformly random) typing sounds while you type in your passwords; it's getting to be too easy to analyze the sounds and extract the contents.


Ok my new beliefs about blister prevention, after three weekends of backpacking for eight hours a day, and watching a bunch of YouTube videos:

  • blisters are caused by layers of skin delaminating, not "friction" / heat directly, though typically the delamination is due to static friction on the outer layer of skin, combined with wet skin. Dynamic friction is more likely to cause raw spots / wear straight through the skin.
  • popping them as soon as you find them is basically always the right call unless you plan to be able to avoid the activity that caused them for a week; otherwise they just keep growing as you continue to do the activity
  • blister donuts and moleskin work okay as long as you can keep them in place somehow, but they don't stick well on their own
  • leukotape, very very widely recommended, is worse than useless because the adhesive seeps through the tape and makes your skin stick to your socks even more tightly than it was before.
  • toe socks are pretty good
  • KT tape is very good
  • Vaseline/similar is pretty good as long as you can get it to stay in the right spots


Anyone have luck getting LLMs to write tests without mocks? The tests I want are often just 1-2 lines of code, but anything I get from Claude or Gemini ends up being 20-30 lines long, despite requests for conciseness, saying no mocks are needed, and seeing using real resources is ok.

(I use LLMs a lot for other stuff, but tests seem to be particularly bad.)

in reply to Satvik

I think Sonnet 4 is not *that* much smarter than 3.7 but it should be significantly more steerable and less likely to insert silly mocks
in reply to Daniel Ziegler

Sonnet 4 is tremendously more effective for my use cases, probably because I use a niche programming language (Julia). Two weeks ago I would have said LLMs make me ~10% more productive, now it looks closer to +100%.

And I'm not even committing LLM-generated code – I just use it to iterate and test on designs, then delete the code and implement from scratch manually.