A skateboard for the mind


This is a writeup I did of a solo podcast I recorded two years ago. My attempt was to gather together some of my ideas on how computers could be better. Because I often find myself referencing "that skateboard blog" I did, but actually never posted, I figured I would post it here in my grand tradition of posting well-aged and somewhat discarded content here.

The computer is "a bicycle for our minds," as Steve Jobs once said.

Do you ever watch those velodrome events in the Olympics? They either have a rolling start, or they have to have someone holding them up for a standing start. The bikers look silly and helpless until they get moving. And then they go in a loop, round and round, very, very fast. If you turn a little too hard your wheels slide out from under you and you eat it. When bikers crash in the velodrome their paper-thin costumes tear to shreds. The road rash is horrific.

What if we thought of the computer as a "skateboard for the mind?"

Skateboarding is improvisation. Repeated failure. Creative recombination. There's no "right way" to skate. It's also much slower than riding a bike. But much cooler.


The tradeoffs we've made in the development of computers since the 60s have favored the "efficiency" and "raw speed" of the velodrome over the ability to turn your own way.

The greatest bicycle I've ever owned is my iPad. It's always fast, always responsive, and incredibly secure. And yet, it can only do what Apple and its blessed legion of third-party developers allow it to do. Trying to code on the iPad is like riding the wrong way in a velodrome. "That's not what it's for!" Coding applications FOR the iPad ON the iPad is impossible. Unless you write some Swift in a text editor that you copy into Xcode later like a jokester.

What would a "skateboard" computer be like? How do we get there?

The Cuckoo's Egg

“The future is already here – it's just not evenly distributed." - William Gibson

I brought this up on the last episode, but I'm reading a book called "The Cuckoo's Egg." It's making me very jealous of how much agency this guy has over his machines. In 1989!

He's tracking a hacker that has been poking around in the network of computers he's responsible for. So he writes tools to help himself track and fight the hacker, while also keeping up with his "day job" of writing software for his fellow astronomers.

Here are some "skateboard" things the Cuckoo's Egg guy does with his computer:

  1. (Physically) intercept the hacker's connection to his system and print out the hacker's every action onto paper.

  2. When the hacker connects to his system, his computer dials his pager and sends a message in morse code.

  3. A computer simulation of telescope bending (not related to hacking, but pretty cool!)

  4. Statistical interpretation of 5,000 phone calls (based on end / start times)

This sort of tool-builder will always be an outlier. It feels like the math stuff is especially difficult to democratize. But it's hard to know HOW outlier and HOW math-wiz this has to be because we've gone so far in the other direction.

The Unix Philosophy

Here's a widely cited, rarely followed, way to use computers, known as The Unix Philosophy:

"Do one thing and do it well," is a beautiful thought, and I believe it's largely right. It's also responsible for some of the worst computer architecture crimes of this century. Microservices and JavaScript's leftpad debacle are two recent examples.

You have to know when you're winning by making simple composable pieces. Otherwise the coordination cost of composing those pieces overwhelms all the original wins.

As mad as I am about iOS's allergy to user-built software, the iOS music ecosystem is my favorite example of programs that work well together in this way. You can pipe midi from one program into another program's synth, then send the audio through AUv3 effect, then back to the first program for output.

lol nice job on this one unix

This is "yak shaving." It's easy to have the detour to make tools, and then the cascading detours that detour generates, take days, then weeks, then years, and then you're dead. It's very difficult for a tool-builder to know which and to build another tool.

I guess the solution is to build more tools!


This is a bit of a tangent. But it's another example of how computers could be different and improvisational.

Smalltalk has a concept called "reflection." Reflection has largely, though not entirely, disappeared from most of our modern systems.

A reflective program is one that reasons about itself. A fully reflective procedural architecture [Smith 1983] is one in which a process can access and manipulate a full, explicit, causally connected representation of its own state. "Causally connected" means that any changes made to a process's self-representation are immediately reflected in its actual state and behavior.

Reflective Facilities in Smalltalk-80

What can you change about your computer while it's running? The color of the menu bars? The desktop picture? Can you touch the logic?

This of course is a very difficult thing to build, not to mention make performant. But it sure is interesting, isn't it?


I don't want to brag but I've been disappointed by computers for a while now. Here's a piece I wrote about the history of wearable computers in 2012, and how I was pretty sure Google Glass was the wrong direction.

One of the first applications of a "wearable computer" was Beating roulette. Nerds have been bringing computers into casinos as early as the 60s, culminating in the 80s with serious wins at roulette. Thomas Bass, who wrote the definitive piece on wearables for Wired in 1998, smuggled a shoe computer to a table, gaining a 44% advantage over the house.

Predicting roulette is using a computer for something only a computer could do. It's taking in the same visuals as a human, but then running complicated calculations to predict the deceleration of the wheel, the movement of the ball, and the eventual landing spot. "The problem is similar to landing a spaceship on the Moon," writes Bass, "except all the calculations have to be done within the few seconds between the launch of the game and the croupier's call to place your bets."

The piece also digs into a concept called affective computing (referencing that same Thomas Bass article):

In Bass's Wired piece, he spends time with Jennifer Healey, another Media Lab graduate student working in the Affective Computing Group. She lived life with electrodes on her face to track when she was smiling and frowning, rings to measure the conductivity of her skin (sweat), and a heart rate monitor. A Palm Pilot tracked the data, and an earpiece could let her know how she was doing.

"Your wearable computer could whisper in your ear," writes Negroponte, "perhaps after playing for a few too many hours with a few too many kids, 'Patience, the birthday party is almost over.'"

Sounds creepy if Google or Facebook tried to do this for you. These are the sort of capabilities that only make sense if computing is skateboard-like. If you try this in the velodrome you'll crash and burn.

Dyson sphere

Along the lines of affective computing, another piece I wrote even longer ago attempted to explain how much our computers are dropping on the floor when they interact with us.

My comparison was to a Dyson sphere. Here's the Wikipedia definition:

A Dyson sphere is a hypothetical megastructure that completely encompasses a star and captures a large percentage of its power output.

Here's my wrap-up paragraph:

Ultimately, it's not that I want to use less of my computers, I just want my computers to use more of me. Luckily, most of the things I've outlined here are obvious and inevitable evolutions of computer interaction. What scares me is that in the meantime we'll get so hung up on "natural user interface" that we'll just use fancy new gestures and voice commands to do the exact same tasks we've been doing with perfect speed and accuracy for years.

Again, this is creepy to say if you're asking Google and Facebook to scan your body 60 times per second to decide what's best for you. But a skateboard computer, composed of hand-built tools, could be much more interesting.

One example of Dyson sphere-ification is Kinect Star Wars vs. Beatsaber.

In Kinect Star Wars you pose and hope the system recognizes that pose. Your whole human body was being translated into a small handful of button presses.

Meanwhile, Beatsaber is "analog." You swing the lightsabers with a full six degrees of freedom, and you hit those stupid blocks in 3D space or you don't.

Fun anecdote: I recently super-doxxed myself to Google. I needed a scan of my driver's license and Social Security card. I accidentally switched to Google Lens. Lol.

But this copy and paste thing seems cool. It feels a bit more like a small tool, and not simply an appendage of the Google beast.

Maybe we can have nice things one day. Probably not. But maybe?


There are some really cool people in the computer industry who are challenging the assumptions of inherent complexity. Jonathan Blow, Casey Muratori, Raph Levien. Sometimes they might come across as curmudgeons on Twitter. But I think they're fighting the good fight.

There's also a really good talk about what's gone wrong in computing called "The Mess We're In" by Joe Armstrong, RIP.

Finally, I recommend the On The Metal podcast for a technical perspective. There's a whole battle raging to make it simpler and more secure to boot a computer. I know not everyone can dip into this particular battle, but it's at least cool to know it's happening.

I love riding my bike. And I'm bad at skateboarding. But I want to get better.