[This is a transcript of a talk I gave at the Diffractions/Sbds event, "Wyrd Patchwork," in Prague on September 22, 2018. The video can be found here. My talk begins at around the 2-hour and 6-minute mark. I've added some links and an image.]
I want to talk about patchwork as an empirical model, but also a little bit as a normative model, because there's this idea that capitalism is increasingly collapsing the fact/value distinction. I tend to think that's true. And I think what that means is that, that which is empirically true increasingly looks to be normatively true also. Or if you're searching for a true model, you should be searching for models that are at once empirically well calibrated with reality and also one should be looking for normative or ethical consistency. And you can find the true model in any particular situation by kind of triangulating along the empirical and the normative. That's kind of how I think about patchwork.
I've been thinking about it in both of these dimensions and that has allowed me to converge on a certain vision of what I think patchwork involves or entails. And I've been writing a lot about that over the past couple months or so. So what I'm going to do in this talk specifically, is not just rehash some ideas that I've been thinking about and writing about and speaking about the past couple months, but I'm going to try to break a little bit of ground, at least in my own weird head, at the very least. And how these, some of these different ideas of mine connect, or can be integrated. In particular, I wrote a series of blog posts a few months ago on what I call reality forking (1, 2, 3). "Forking" is a term that comes from the world of software engineering. And so that's going to be one component of the talk.
You'll see it. It's very obvious how that connects to the idea of patchwork. And I'm also going to talk about this vision for a communist patch a lot of us have been interested in. And I've been talking with a lot of people about this idea of the communist patch and soliciting, you know, different people's impressions on it. And I also have written a few blog posts recently talking — kind of sketching, kind of hand-waving, if you will — at what a possibly communist patch might look like. A lot of people think, to this day, that patchwork has a very kind of right-wing connotation. People think primarily of Moldbug and Nick Land when they think of patchwork. But I think it's not at all obvious that patchwork necessarily has a right-wing flavor to it.
I think we can easily imagine left-wing patches that would be as competitive and as successful as more authoritarian patches. And so that's kind of what I've really been thinking a lot about recently. And even Nick Land himself told me that, you know, there's nothing wrong with trying to think about and even build a communist patch — it's all fair play. He's much less bullish on it than I am, but be that as it may. So those two ideas I'm going to discuss basically in turn and then try to connect them in a few novel ways. I have a few points or comments or extrapolations or connections between these two different ideas I've been working on, that I've never really written down or quite articulated yet. So that's what I'm going to try to do here.
So first of all, I was going to start this by talking a little bit about how patchwork I think is already happening in a lot of ways, but I deleted many of my bullet points because Dustin's presentation basically covered that better than I possibly could. So I'm not going to waste too much time talking about that. There's a lot of empirical data right now that looks a lot like fragmentation is the order of the day and there's a lot of exit dynamics and fragmentation dynamics that we're observing in many domains. And yeah, Dustin articulated a lot of them.
One thing I would say to kind of situate the talk, though, is that it's worth noting that not everyone agrees with this, you know... There's still a lot of integrative talk nowadays. There's a lot of discourse about the necessity of building larger and larger organizations. Especially when people are talking about global issues and major existential threats. Often in the educated discourse around preventing nuclear threats, for instance, or AI, things like runaway inhumane genetic testings, things like that. You could probably think of a few others. Climate change would be the obvious big one, right? A lot of these major global issues, the discourse around them, the expert opinions, tend to have a kind of integrative, centralized tendency to them. Actually just this morning I happened to be listening to a podcast that Sam Harris did with Yuval Harari. This guy who wrote the book, Sapiens, this mega global blockbuster of a book, and you know, he seemed like a nice guy, a smart guy of course, but everything he was saying was totally integrated. He was talking about how we need things like international organizations and more global international cooperation to solve all of these different problems and Sam Harris was just kind of nodding along happily. And that got me thinking actually, because even if you read people like Nick Bostrom and people who are kind of more hard-nosed and analytical about things like intelligence explosion, you find a lot of educated opinion is the opposite of a patchwork orientation, you find "We need to cooperate at a global level." Anyway, the reason I mentioned this is just to put in context that the ideas we're interested in and the empirical dynamics that were pinpointing are not at all obvious to everyone.
Even though, when you really look at all of the fragmentation dynamics now, I think it's increasingly hard to believe any idea, any proposal having to do with getting all of the nation states to cooperate on something. I just... I just don't see it. For instance, genetic engineering, you know China is off to the races and I just don't see any way in which somehow the US and China are going to negotiate some sort of pause to that. Anyway, so that's worth reflecting on. But one of the reasons I mention that is because I kind of have a meta-theory of precisely those discourses and that's what I'm going to talk about a little bit later in my talk when I talk about the ethical implications, because I think a lot of that is basically lying.
Okay. One of my theses is that when people are talking about how we have to organize some larger structure to prevent some moral problem — nine times out of ten, what they're actually doing is a kind of capitalist selling process. So that's actually just a kind of cultural capitalism in which they're pushing moral buttons to get a bunch of people to basically pay them. That is a very modern persona, that's a modern mold and that's precisely one of many things that I think is being melted down in the acceleration of capitalism. What's really happening is all that's really feasible in so many domains. All you can see for miles when you look in every possible direction is fragmentation, alienation, atomization, exits of all different kinds on all different kinds of levels.
And then you have people who are like, "Uh, we need to stop this, so give me your money and give me your votes." I think that's basically an unethical posture. I think it's a dishonest, disingenuous posture and it's ultimately about accruing power to the people who are promoting that — usually high-status, cultural elites in the "Cathedral" or whatever you want to call it. So that's why I think there are real ethical implications. I think if you want to not be a liar and not be a kind of cultural snake-oil salesman — which I think a lot of these people are — patchwork is not only what's happening but we're actually ethically obligated to hitch our wagon to patchwork dynamics. If only not to be a liar and a manipulator about the the nature of the real issues that we're going to have to try to navigate somehow.
I'll talk a little bit more about that, but I just wanted to kind of open up the talk with that reflection on the current debate around these issues. So, okay.
The one dimension of patchwork dynamics or exit dynamics that we're observing right now, that Dustin didn't talk about so much, is a patchwork dynamic that's taking place on the social-psychological level. To really drive this point home, I've had to borrow a term from the world of software engineering. I'll make this really quick and simple.
Basically, when you're developing software and you have a bunch of people contributing to this larger codebase, you need some sort of system or infrastructure for how a bunch of people can edit the code at the same time, right? You need to keep that orderly, right? So there's this simple term, it's called forking. So you have this codebase and if you want to make a change to the code base, you fork it. In a standard case, you might do what we call a soft fork. I'm butchering the technical language a little bit; if there are any hardcore programmers in the room, I'm aware I'm painting with broad strokes, but I'll get the point across effectively enough without being too nerdy about it.
A soft fork means that you pulled the codebase off for your own purposes, but it ultimately can merge back in — is the simple idea there. But a hard fork is when you pull the code base off to edit it, and there's no turning back. There's no reintegrating your edits to the shared master branch or whatever you want to call it. So I use this kind of technical distinction between a soft fork and a hard fork to think about what's actually going on with social, psychological reality and its distribution across Western societies today. The reason I do this is because I think you need this kind of language to really drive home how radical the social psychological problems are. I really think that we underestimate how much reality itself is being fragmented in different subpopulations.
I think we're talking about fundamental... We are now fundamentally entering into different worlds and it's not at all clear to me that there's any road back to having some sort of shared world. And so I sketched this out in greater detail. The traditional human society, you can think of it as a kind of system of constant soft forking, right? Individuals go off during the day or whatever, they go hunting and do whatever traditional societies do, and at the end of the night they integrate all of their experiences in a shared code base. Soft forks, which are then merged back to the master branch around the campfire or whatever you want to call it, however you want to think about that. But it's only now that, for the first time ever, we have the technological conditions in which individuals can edit the shared social codebase and then never really integrate back into the shared code base.
And so this is what I call the hard forking of reality. I think that is what we're living through right now. And I think that's why you see things like political polarization to a degree we've never seen before. That's why you see profound confusion and miscommunication, just deep inabilities to relate with each other across different groups, especially like the left vs. right divide, for instance. But you also see it with things like... Think about someone like Alex Jones, think these independent media platforms that are just on a vector towards outer space — such that it's hard to even relate it to anything empirical that you can recognize. You see more and more of these kinds of hard reality forks, or that's what I call them. I'm very serious.
I think educated opinion today underestimates how extreme that is and how much that's already taking place. It's not clear to me once this is underway, it's not clear to me how someone who is neck-deep in the world of Alex Jones — and that is their sense of what reality is — how that person is ever going to be able to sync back up with, you know, an educated person at Harvard University or something like that. It's not just that those people can't have dinner together — that happened several decades ago probably — but there's just no actual technical, infrastructural pathway through which these two different worlds could be negotiated or made to converge into something shared. The radicalism of that break is a defining feature of our current technological moment.
And that is an extraordinary patchwork dynamic. In other words, I think that patchwork is already here, especially strong in the socio-psychological dimension, and that's very invisible. So people underestimate it. People often think of patchwork as a territorial phenomenon and maybe one day it will be, but I think primarily for now it's social-psychological and that should not be underestimated because you can go into fundamentally different worlds even in the same territory. But that's what the digital plane opens up to us. So that's one half of what I'm bringing to the table in this talk.
There are a few antecedent conditions to explain, like why I think this is happening now. One is that there's been an extraordinary breakdown in trust towards all kinds of traditional, institutionalized, centralized systems. If you look at the public opinion data, for instance, on how people view Congress in the United States, or how people view Parliament or whatever, just trust in elected leaders... You look at the public opinion data since the fifties and it's really, really on the decline, a consistent and pretty rapid decline.
And this is true if you ask them about the mass media, politicians, a whole bunch of mainstream, traditional kinds of institutions that were the bedrock of modernized societies... People just don't take them seriously anymore at all. And I think that is because of technological acceleration, what's happened is that there is unprecedented complexity. There's just too much information. There's so much information that these modern institutions are really, really unwieldy. They're really unable to process the complexity that we now are trying to navigate and people are seeing very patently that all of these systems are just patently not able to manage. They're not able to do or give what they're supposed to be giving with this explosion of information that they were not designed to handle. So it's kind of like a bandwidth problem, really. But because of this, people are dropping their attention away from these institutions and they're looking outwards, they're looking elsewhere, they're looking for other forms of reality because that's ultimately what's at stake here.
These traditional institutions, they supplied the shared reality. Everyone referred back to these dominant institutions because — even if you didn't like those institutions in the 60s or 70s or whatever, even when people really didn't like those institutions, like the hippies or whatever — everyone recognized them as existing, as powerful. So even opposing them, you kind of referred back to them. We're now post- all of that, where people so mistrust these institutions that they're not even referring back to them anymore. And they're taking all their cues for what reality is from people like Alex Jones or people like Jordan Peterson or you name it, and you're going to see more and more fragmentation, more and more refinement of different types of realities for different types of subpopulations in an ever more refined way that aligns with their personalities and their preferences. These are basically like consumer preferences. People are going to get the realities that they most desire in a highly fragmented market. Anyway... So I think I've talked enough about that. That's my idea of reality forking and that's my model of a deep form of patchwork that I think is already underway in a way that people underestimate.
So now I want to talk a little bit more about the ethics of patchwork because I think the observations that I just prevent presented, they raise ethical questions. And so if I am right, that reality itself is already breaking up into multiple versions and multiple patches, well then that raises some interesting questions for us, not just in terms of what we want to do, but in terms of what should we do.
What does it mean to seek the good life if this is in fact what's happening? It seems to me that, right now, you're either going to be investing your efforts into somehow creatively co-constituting a new reality or you're going to be just consuming someone else's reality. And a lot of us, I think, do a combination of this. Like all the podcasts I listen to, and all the Youtube videos I watch, that's me outsourcing reality-creation to other people, to some degree. But then the reason I've gotten on Youtube and the reason I've gotten really into all of these platforms and invested myself in creating my own sense of the world is because I don't just want to be a consumer of other people's realities. I want to be... I want to create a world. That would, that sounds awesome. That would be the ideal, right? But the problem is that people are differently equipped to do so, to either create or consume realities and I think that this is difficult and very fraught. This is a very politically fraught problem. The left and the right will have debates about, you know, "the blank slate" versus the heritability of traits and all of that. And I don't want to get into that now, but however you want to interpret it, it is an obvious fact that some people are better equipped to do things like create systems, than other people. To me, this is the ethical-political question space.
The default mode right now is the one that I already described at the top of my talk: it's the moralist. It's the traditional left-wing (more or less) posture. "Here's a program for how we're going to protect a bunch of people. All it requires is for you to sign up and give your votes and come to meetings and give your money and somehow we're going to all get together and we're going to take state power and protect people" or something like that. As I already said — I won't beat a dead horse — but I think that's increasingly revealing itself to be a completely impractical and not serious posture that plays with our... it suits our moral tastebuds a little bit, but it's increasingly and patently not able to keep up with accelerating capitalism.
That's not gonna work. Why I think patchwork is an ethical obligation is because, if you're not going to manipulate people by trying to build some sort of large centralized institution, by manipulating their heartstrings, then what remains for us to do is to create our own realities, basically. And I think that the most ethical way to do that is to do it honestly and transparently, to basically reveal this, to reveal the source code of reality and theorize that and model that and make those blueprints and share those blueprints and then get together with people that you want to get together with and literally make your own reality. I feel like that doesn't just sound cool and fun, but you kind of have to do that or else you're going to be participating in this really harmful, delusional trade. That's my view anyway.
Now I'll just finish by telling you what I think the ideal path looks like ethically and practically. I've called it many different things, I haven't really settled on a convenient phrase to summarize this vision, but I think of it as a neo-feudal techno-communism. I think the ideal patch that will be both most competitive, most functional, most desirable and successful as a functioning political unit, but also that is ethically most reflective and consistent with the true nature of human being is... It's going to look something a little bit like European feudalism and it's going to be basically communist, but with contemporary digital technology.
Let me unpack that for you a little bit. You probably have a lot of questions [laughing]. One thing is that patchwork always sounds a little bit like "intentional communities." And on the Left, the "intentional communities" kind of have a bad rap because they've never really worked. You know, people who want to start a little group somewhere off in the woods or whatever, and make the ideal society, and then somehow that's going to magically grow and take over. It usually doesn't end well. It doesn't have a good historical track record. It usually ends up in some kind of cult or else it just fizzles out and it's unproductive or whatever. I think that the conditions now are very different, but I think if you want to talk about building a patch, you have to kind of explain why your model is different than all the other intentional communities that have failed.
One reason is that the digital revolution has been a game changer, I think. Most of the examples of failed intentional communities come from a pre-digital context, so that's one obvious point. I think the search-space, the solution-space, has not all been exhausted. That's kind of just a simple point.
But another thing I've thought a lot about, and I've written some about, is that, in a lot of the earlier intentional communities, one of the reasons they fail is because of self-selection. That's just a fancy social science term for... There's a certain type of person who historically has chosen to do intentional communities and they tend to have certain traits and I think for many reasons — I don't want to spend too much time getting into it — but it's not hard to imagine why that causes problems, right? If all the people are really good at certain things but really bad at other things, you have very lopsided communities in terms of personality traits and tendencies. I think that that's one of the reasons why things have led to failure. So what's new now, I think, is that because the pressure towards patchwork is increasingly going to be forced through things like climate change and technological shocks of all different kinds, because these are fairly random kinds of systemic, exogenous shocks, what that means is it's going to be forcing a greater diversity of people into looking for patches or maybe even needing patches. And I think that is actually valuable for those who want to make new worlds and make better worlds, because it's actually nature kind of imposing greater diversity on the types of people that will have to make different patches.
So what exactly does neo-feudal techno-communism look like? Basically it would have a producer elite, and this is where a lot of my left-wing friends start rolling their eyes, because it basically is kind of like an aristocracy. Like, look, there's going to be a small number of people who are exceptionally skilled at things like engineering and who can do things that most other people can't. You need at least a few people like that to engineer really sophisticated systems. Kind of like Casey said before, "the mayor as sys-admin." That's kind of a similar idea. You'd have a small number of elite engineer types and basically they can do all of the programming for the system that I'm about to describe, but what they also do is they make money in the larger techno-commercium. They would run a small business, basically, that would trade with other patches and it would make money, in probably very automated ways. So it would be a sleek, agile kind of little corporation of producer elites at the top of this feudal pyramid of a patch society. Then there would be a diversity of individuals including many poor unskilled, disabled, etc., people who don't have to do anything basically. Or they can do little jobs around the patch or whatever, to help out.
The first thing you might be thinking — this is the first objection I get from people — is why would the rich, these highly productive, potentially very rich, engineer types want to support this patch of poor people who don't do anything? Isn't the whole problem today, Justin, that the rich don't want to pay for these things and they will just exit and evade?
Well, my kind of novel idea here is that there is one thing that the rich today cannot get their hands on, no matter where they look. And I submit that it's a highly desirable, highly valuable human resource that most people really, really, really want. And that is genuine respect and admiration, and deep social belonging. Most of the rich today, they know that people have a lot of resentment towards them. Presumably they don't like the psychological experience of being on the run from national governments and putting their money in Swiss bank accounts. They probably don't like feeling like criminals who everyone more or less kind of resents and wants to get the money of, or whatever. So my hypothesis here is that if we could engineer a little social system in which they actually felt valued and desired and admired and actually received some respect for their skills and talents that they do have and the work that they do put in... I would argue that if you could guarantee that, that they would get that respect, and the poor would not try to take everything from them. If you could guarantee those things, then the communist patch would actually be preferable to the current status quo for the rich people. My argument is that this would be preferable; it would be a voluntary, preferable choice for the rich, because of this kind of unique, new agreement that the poor and normal people won't hate them and we'll actually admire them for what they deserve to be admired for. So then the question becomes, well, how do you guarantee that that's going to happen? This is where technology comes in.
The poor and normal people can make commitments to a certain type of, let's call them "good behaviors" or whatever. Then we can basically enforce that through trustless, decentralized systems, namely, of course, blockchain. So what I'm imagining is... Imagine something like the Internet of Things — you know, all of these home devices that we see more and more nowadays that have sensors built in and can passively and easily monitor all types of measures in the environment. Imagine connecting that up to a blockchain, and specifically Smart Contracts, so that basically the patch is being constantly measured, your behavior in the patch is being constantly measured. You might have, say, skin conductance measures on your wrist; there might be audio speakers recording everyone's voice at all times. I know that sounds a little authoritarian, but stick with me. Stick with me.
Basically, by deep monitoring of everything using the Internet of Things, what we can do is basically as a group agree on what is a fair measure of, say, a satisfactory level of honesty, for instance. Let's say the rich people say, "I'll guarantee you a dignified life by giving you X amount of money each month. You don't have to do anything for it as long as you respect me, you know, you don't tell lies about me, you don't plot to take all of my money" or whatever. So then you would have an Alexa or whatever, it would be constantly recording what everyone says, and that would be hooked up to a Smart Contract. And so if you tell some lie about the producer aristocrat, "He totally punched me the other day, he was a real ignoble asshole," and that's actually not true. Well, all of the speech that people are speaking would be constantly compared to some database of truth. It could be Wikipedia or whatever. And every single statement would have some sort of probability of being true or false, or something like that. That could all be automated through the Internet of Things feeding this information the internet, and basically checking it for truth or falsity. And then you have some sort of model that says, if a statement has a probability of being false that is higher than — maybe set it really high to be careful, right? — 95 percent, so only lies that can be really strongly confirmed... Those are going to get reported to the community as a whole.
If you have X amount of bad behaviors, then you lose your entitlement from the aristocrat producers. It's noblesse oblige, the old kind of feudal term for basically an aristocratic communism, the [obligatory] generosity of the noble. So that's all very skittish. A little sketch of how Internet of Things and Smart Contracts could be used to create this idea of a Rousseauean General Will.
The reason why this has never worked in history is because of lying, basically. People can always defect. People can always manipulate and say they're going to do one thing but then not deliver. That's on the side of the rich and also on the side of the poor. But what's at least in sight now, is the possibility that we could define very rigorously the ideal expectations of everyone in a community and program that in transparent Smart Contracts, hook those up to sensors that are doing all of the work in the background, and in this way basically automate a radically guaranteed, egalitarian, communist system in which people do have different abilities, but everyone has an absolutely dignified lifestyle guaranteed for them as long as they're not total [expletive] who break the rules of the group. You can actually engineer this in a way that rich people would find it preferable to how they're currently living. So to me that's a viable way of building communism that hasn't really been tried before. And I think it really suits a patchwork model. I think that this would be something like an absolutely ideal patch, and not just in a productive, successful way. This is the ideal way to make a large group of people maximally productive and happy and feel connected and integrated. Like everyone has a place and everyone belongs, even if there's a little bit of difference in aptitudes. The system, the culture, will reflect that. But in a dignified, and fair, and reasonable kind way, a mutually supportive way. I could say more, but I haven't been keeping time, and I feel like I've been talking enough.
I found out recently that — hat tip to my friend the Jaymo — the town of Tombsboro, Georgia is right now for sale, for only $1.7 million. I think that's a pretty good deal. It comes with a railroad station, a sugar factory, all kinds of stuff and you could easily build a little prototype patch that I just described. If you have a bunch of people and it's a major publicized project, it wouldn't be that hard to raise enough for a mortgage on a $1.7 million property. Especially if you have a compelling white paper along the lines that I just sketched. I'm not quite there yet, but that's what I'm thinking about, that's my model or my vision of the communist patch. So I'm going to cut myself off there. Thank you very much.