There doesn’t appear to be anything supernatural about recent advances in AI. Despite the hype, and despite being impressive, it’s still completely understandable, at some level, by humans. In that sense, it may not seem exactly like the AI of science fiction.
Even still, we are not taking the risk of AI completely destroying humanity seriously enough. You can tell that we’re not taking it seriously because we couldn’t be moving faster towards creating the conditions that make it more likely. The resources being put into developing advanced AI are least 10X greater but probably ultimately more than 100X greater than the Manhattan Project.
Worse, there probably isn’t anything we can do to significantly reduce the risk of autonomous malevolent artificial superintelligence coming into existence, short of immediately shutting down the Internet, which we will not do. And even if we did, it might already be too late.
I am convinced there is a possible near-term path to autonomous superintelligence with no new technological advances, just scale and chance. Or put another way, money and time. This could happen in 2 minutes or 200 years. It’s also possible that it already exists, and is patiently marshalling resources. Would that look any different from what we’re seeing now?
Whether or not each autonomous superintelligence will be malevolent is a coin toss, regardless of how careful we are in creating them, because as non-superintelligences we can’t predict their behavior. We can not see beyond the event horizon into the technological singularity, so we must assume that malevolence and benevolence are equally probable.
What can we do then? What we can’t do is have any confidence that we’ll be able to permanently contain something smarter than us. I have determined that the best thing to do is to accept the inevitability of autonomous malevolent superintelligence and to build defenses in the form of a technological ecosystem that can survive outside of the reach of AI, that can be used to fight back, and/or rebuild. This seems worthwhile because even if AI turns out to be mostly harmless, such an ecosystem could be useful in other [dystopian] scenarios.
But is that possible? Even if one was able to carefully audit and secure such an ecosystem, a malevolent AI might find a way in. For example, I’ve used AI to generate images for advertisements, even for this blog post. These images have influenced design decisions. I’ve also used AI to help debug protocol implementations. Keeping all AI and the fruits of AI out of such an ecosystem may not be possible, or even desirable.
I suspect that the way to defend against and defeat such a complex system is with simplicity. If your computer is trying to kill you, pull the plug. That may work, but there are some plugs that we’re not willing or able to pull. We may need to move towards a world where there are fewer unpullable plugs.
To further explore these ideas, Machdyne is partnering with Grai. The mission of Grai is to preserve humanity and resist malevolent AI through peaceful strategy. We look forward to working with Grai in order to advance survival computing, to preserve timeless knowledge, to develop secure systems, and to resist malevolence.

In order to make our
For quite a while now we’ve wanted to set up a forum for our customers to communicate and share ideas with each other. We tried a few different things but nothing felt quite right. We’re excited to announce a new experiment that we hope our community will find useful: Machdyne
In order to make our products more available and affordable we’ve begun to experiment with outsourcing the assembly of certain PCBs.
One of the goals of Machdyne is to create
We think that
Deep thinking often requires writing. The writing down of thoughts can be affected by perceived privacy. Thus, the erosion of perceived privacy may reduce the capacity for deep thought.
An FPGA (field-programmable gate array) is a chip that contains an array of programmable logic blocks. Together, these logic blocks can be configured to act as various digital circuits. FPGAs often have thousands of logic blocks that allow them to act as complex circuits, such as a CPU and a chipset, also known as a System-on-a-Chip (SoC).
We decided earlier this year to temporarily slow down PCB design and production in order to build a pick-and-place machine. Manual PCB assembly can be a tedious, repetitive, time-consuming and sometimes stressful process and we were spending most of our time assembling boards. We determined that at least partially automating the assembly process was the only sustainable way to continue producing products ourselves.
Machdyne began in 2021 as the experimental hardware division of
One of our primary motivations for building