This is a compact account of my current working hypothesis for what's wrong with our culture and what needs to be done. Continue reading
So, some writer named Cathy O’Neil wrote about futurists’ opinions about AI risk. This piece focused on futurists as social groups with different incentives, and didn’t really engage with the content of their arguments. Instead, she points out considerations like this:
First up: the people who believe in the singularity and are not worried about it. […] These futurists are ready and willing to install hardware in their brains because, as they are mostly young or middle-age white men, they have never been oppressed.
She doesn’t engage with the content of their arguments about the future. I used to find this sort of thing inexplicable and annoying. Now I just find it sad but reasonable. Continue reading
I am afraid of the anglerfish. Maybe this is why the comments on my blog tend to be so consistently good.
Recently, a friend was telling me about the marketing strategy for a project of theirs. They favored growth, in a way that I was worried would destroy value. I struggled to articulate my threat model, until I hit upon the metaphor of that old haunter of my dreamscape, the anglerfish. Continue reading
This isn't a carefully structured post - it's just a bunch of pragmatic info, in the hopes that it will be useful to people who need this information structured the way I do. Continue reading
Ayn Rand’s Atlas Shrugged is set in a world in which the death dance of capitalism has reached its final stages, the state itself becoming an instrument of direct appropriation of surplus value generated by the workers. As industrialists become aware of the extractive nature of the process in which they are participating, one by one, they convert to the radical anarchism of an agitator named John Galt,* and “go on strike” to an utopian community hidden in the mountains of Colorado: Galt’s Gulch.
In Galt’s Gulch, resources are allocated to whomever can use them most productively, in an informal process; since everyone can see how their interests converge, levels of trust are high, and hoarding and shirking are basically nonproblems. People pick up whatever tasks seem needed, regardless of their profession or the ability such tasks might give them to extract rents.
This raises the obvious question: Why does anyone use money in Galt’s Gulch? Continue reading
Growing up Jewish, I thought that the traditional rules around the Sabbath were silly. Then I forgot to bring a spare battery on a camping trip. Now I think that something like the traditional Jewish Sabbath is an important cultural adaptation to preserve leisure, that would otherwise be destroyed in an urbanized, technological civilization. Continue reading
In the past year, I have noticed that the Society of Friends (also known as the Quakers) has come to the right answer long before I or most people did, on a surprising number of things, in a surprising range of domains. And yet, I do not feel inclined to become one of them. Giving credit where credit is due is a basic part of good discourse, so I feel that I owe an explanation.
The virtues of the Society of Friends are the virtues of liberalism: they cultivate honest discourse and right action, by taking care not to engage in practices that destroy individual discernment. The failings of the Society of Friends are the failings of liberalism: they do not seem to have the organizational capacity to recognize predatory systems and construct alternatives.
Fundamentally, Quaker protocols seem like a good start, but more articulated structures are necessary, especially more closed systems of production. Continue reading
UPDATE: I'm currently staying in a cabin at a Quaker retreat center. This is basically the thing I needed. Thanks to everyone who reached out with suggestions or offers.
I’m currently coming to terms with just how much of human communication is marketing, like unto the Hobbesian war of all against all. I want to figure out a way for human beings to coordinate and create value when the dominant society is like that. That task is too big for me, so I need a team. I don’t know how to find the right people, but my best guess is that if I can clearly articulate what is needed, the right people have a good chance of recognizing my project as the one they want to be part of. But I can’t write about this in a reasonable way, because I can’t think about this in a reasonable way, because my intuitions are still all applying way too high a level of implicit trust, which means that I’m effectively deluged with spam and buying everything.
So, I want to get away, physically, to buy myself a little breathing room and get my head screwed on straight. A cabin, somewhere where I am not implicitly obligated to engage in more than incidental contact with other human beings, and can live in a reasonable amount of comfort for a while (e.g. includes basic temperature control and running water) is the essential thing. Other desiderata that are nice to have but not essential include:
- A high vantage point overlooking something, whether it be coastal cliffs, a mountainside, or something similar. One goes to a mountaintop to receive the law, so this feels aesthetically appropriate in a way that “a cabin in the woods” or something out in the more level desert does not.
- A land line telephone I can use.
- Drivable distance from the SF Bay Area or otherwise convenient for me to get to.
- Internet access, either on-site or within an hour of the site.
- ETA: Running water
Let me know if you have something like this to offer either free or for money, know of someone who does, or have advice beyond checking AirBnB for how to find something like this.
My best guess is that I should try this out for two weeks and then figure out what to do longer-term.
Simple consequentialist reasoning often appears to imply that you should trick others for the greater good. Paul Christiano recently proposed a simple consequentialist justification for acting with integrity:
I aspire to make decisions in a pretty simple way. I think about the consequences of each possible action and decide how much I like them; then I select the action whose consequences I like best.
To make decisions with integrity, I make one change: when I imagine picking an action, I pretend that picking it causes everyone to know that I am the kind of person who picks that option.
If I’m considering breaking a promise to you, and I am tallying up the costs and benefits, I consider the additional cost of you having known that I would break the promise under these conditions. If I made a promise to you, it’s usually because I wanted you to believe that I would keep it. So you knowing that I wouldn’t keep the promise is usually a cost, often a very large one.
Overall this seems like it’s on the right track – I endorse something similar. But it only solves part of the problem. In particular, it explains interpersonal integrity such as keeping one's word, but not integrity of character. Continue reading