FutureBlind Podcast
FutureBlind Podcast

FutureBlind Podcast

Max Olson

Overview
Episodes

Details

This is the audio edition of the FutureBlind blog. Episodes will be rare: on occasion I'll record an audio version of an essay that tries to take advantage of the audio medium with clips from others, good sound design, and more. futureblind.com

Recent Episodes

Roundup #7: Augmented Intelligence & The Education Frontier
JUL 12, 2023
Roundup #7: Augmented Intelligence & The Education Frontier
<p>This is going to be something a little bit different: <strong>a podcast roundup</strong>.</p><p>For the past few years I’ve been doing a roundup post and essay about once a quarter. This will be an audio version of that. I wanted to try a podcast format where it’s primarily clips from others with me narrating along the way.</p><p>Why? It sounded like fun and I love experimenting with how to make content on different mediums. I think it turned out pretty well, although much longer than I initially thought.</p><p>Going through all these audio files was a pain, so I wrote a script that transcribed them all and allows me to search by topic and group related content. This was kinda fun to make as well.</p><p>Most of this episode is on AI, similar to my last few roundups, but I also did a section on the future of education, and I’ve got some carveouts for other interesting content at the end as well.</p><p>The podcast is almost an hour long, you should be able to skip to relevant sections in the description. And <strong>if you’re not an audio person, there’s the usual written version below.</strong></p><p>* 01:37 — 🤖 <strong>The A.I. Frontier</strong></p><p>* 08:45 — AI: How useful are language models?</p><p>* 14:05 — AI: What does AI allow us to do?</p><p>* 14:33 — AI: Easier to communicate</p><p>* 18:16 — AI: Creativity</p><p>* 25:50 — AI: Augmented Intelligence</p><p>* 33:29 — 🧑‍🏫 <strong>The Education Frontier</strong></p><p>* 43:00 — 🔗 <strong>Interesting Content</strong></p><p>🤖 A.I. Frontier</p><p>In this section about AI, I’m going to first cover some basic fundamentals of how these AI models work, and then really get into their use cases and what they allow us to do.</p><p>It’s been 3 years since the original GPT-3 release, 6 months since ChatGPT, and we’ve had GPT-4 and other open source models for more than 3 months now.</p><p><strong>So with that said, I want to talk about the current state of GenAI, in particular language models because these have more generalized promise.</strong></p><p>Here’s <a target="_blank" href="https://www.youtube.com/watch?v=bZQun8Y4L2A">Andrej Karpathy in a talk called “State of GPT”</a> with more on what models are out there and what they can do:</p><p>Now since then we've seen an entire evolutionary tree of base models that everyone has trained. Not all of these models are available. For example, the GPT-4 base model was never released. The GPT-4 model that you might be interacting with over API is not a base model, it's an assistant model. And we're going to cover how to get those in a bit. GPT-3 base model is available via the API under the name DaVinci. And GPT-2 base model is available even as weights on our GitHub repo. But currently the best available base model probably is the LLaMA series from Meta, although it is not commercially licensed.</p><p>Now one thing to point out is base models are not assistants. They don't want to answer to you, they don't want to make answers to your questions. They just want to complete documents. So if you tell them, write a poem about the bread and cheese, it will answer questions with more questions. It's just completing what it thinks is a document. However, you can prompt them in a specific way for base models that is more likely to work.</p><p>A review of what we have now:</p><p>* Monolithic foundation models from OpenAI, Google, Anthropic, etc.</p><p>* Assistant models built on top of these like ChatGPT, Bard, and Claude.</p><p>* Open source models — particularly the semi-open-sourced LLaMA model from Meta, where you can run it locally and even fine-tune it with your own data.</p><p>State-of-the-art GPT-4 is actually a “mixture of experts” assistant model that routes your prompt to whichever large model that can complete it best. This is a big difference from the original GPT-3.</p><p>In the early days of GPT-3, if you “asked” it a question, it had a high chance of spitting out something rude, offensive, or just didn’t really answer the question. How do they get from the base models to the assistant models like ChatGPT that are pleasant to talk to?</p><p>So instead, we have a different path to make actual GPT assistants, not just base model document completers. And so that takes us into supervised fine tuning. So in the supervised fine tuning stage, we are going to collect small but high-quality data sets. And in this case, we're going to ask human contractors to gather data of the form prompt and ideal response. And we're going to collect lots of these, typically tens of thousands or something like that. And then we're going to still do language modeling on this data. So nothing changed algorithmically. We're just swapping out a training set. So it used to be internet documents, which is a high-quantity, low-quality for basically QA prompt response kind of data. And that is low-quantity, high-quality. So we would still do language modeling. And then after training, we get an SFT model. And you can actually deploy these models. And they are actual assistants. And they work to some extent.</p><p>The performance of these models seems to have surprised even the engineers that created them. How and why do these models work?</p><p>Stephen Wolfram released a great essay about this earlier this year aptly called “<a target="_blank" href="https://writings.stephenwolfram.com/2023/02/what-is-chatgpt-doing-and-why-does-it-work/">What is ChatGPT Doing… and Why Does It Work?</a>” The essay was basically a small book, and Wolfram actually did turn it into a small book. So it’s long, but if you’re interested in these, I’d really recommend it, even if not everything makes sense. He explains models, neural nets, and a bunch of other building blocks. Here’s the gist of what ChatGPT does:</p><p>The basic concept of ChatGPT is at some level rather simple. Start from a huge sample of human-created text from the web, books, etc. Then train a neural net to generate text that’s “like this”. And in particular, make it able to start from a “prompt” and then continue with text that’s “like what it’s been trained with”.</p><p>As we’ve seen, the actual neural net in ChatGPT is made up of very simple elements—though billions of them. And the basic operation of the neural net is also very simple, consisting essentially of passing input derived from the text it’s generated so far “once through its elements” (without any loops, etc.) for every new word (or part of a word) that it generates.</p><p>. . . The specific engineering of ChatGPT has made it quite compelling. But ultimately (at least until it can use outside tools) ChatGPT is “merely” pulling out some “coherent thread of text” from the “statistics of conventional wisdom” that it’s accumulated. But it’s amazing how human-like the results are. And as I’ve discussed, this suggests something that’s at least scientifically very important: that human language (and the patterns of thinking behind it) are somehow simpler and more “law like” in their structure than we thought. ChatGPT has implicitly discovered it. But we can potentially explicitly expose it, with semantic grammar, computational language, etc.</p><p>In a YouTube lecture he put out, Wolfram explains that the ChatGPT's ability to generate plausible text is due to its ability to use the known structures of language, specifically the regularity in grammatical syntax.</p><p>We know that sentences aren't random jumbles of words. Sentences are made up with nouns in particular places, verbs in particular places, and we can represent that by a parse tree in which we say, here's the whole sentence, there's a noun phrase, a verb phrase, another noun phrase, these are broken down in certain ways. This is the parse tree, and in order for this to be a grammatically correct sentence, there are only certain possible forms of parse tree that correspond to a grammatically correct sentence. So this is a regularity of language that we've known for a couple of thousand years.</p><p>How useful are language models?</p><p>How useful really are language models, and what are they good and bad at?</p><p>I have occasionally used it to summarize super long emails, but I've never used it to write one. I actually summarize documents is something I use it for a lot. It's super good at that. I use it for translation. I use it to learn things.</p><p>That was Sam Altman, founder of OpenAI, saying he primarily uses it for summarizing. <a target="_blank" href="https://matt-rickard.com/chatgpt-plugins-dont-have-pmf">At another event</a> he admitted that ChatGPT plug-ins haven’t really caught on yet, despite their potential.</p><p>What are they good for then? What are the LLM “primitives”— or the base capabilities that they’re actually good at?</p><p>The major primitives are: Summarization, text expansion, basic reasoning, semantic search in meaning space. Most of all translation — between human languages, computer languages, and any combination of them. I’ll talk about this in a little bit.</p><p>They’re clearly not good at retrieving information, or being a store of data. The same way human brains aren’t very good at this naturally without heavy training. This might never really be a core primitive of generalized LLMs. They are not a database and likely will never be. And that’s ok — they can always be supplemented with vector or structured data sources.</p><p>As I mentioned, the context window of a transformer is its working memory. If you can load the working memory with any information that is relevant to the task, the model will work extremely well, because it can immediately access all that memory.</p><p>Andrej Karpathy talks here about what LLMs aren’t good at, but also how they can easily be supplemented:</p><p>And the emerging recipe there is you take relevant documents, you split them up into chunks, you embed all of them, and you basically get embedding vectors that represent that data. You store that in the vector store, and then at test time, you make some kind of a query to your vector store, and you fetch chunks that might be relevant to your task, and you stuff them into the prompt, and then you generate. So this can work quite well in practice.</p><p>So this is I think similar to when you and I solve problems. You can do everything from your memory and transformers have very large and extensive memory, but also it really helps to reference some primary documents. So whenever you find yourself going back to a textbook to find something or whenever you find yourself going back to documentation of a library to look something up, the transformers definitely want to do that too. You have some memory over how some documentation of a library works, but it's much better to look it up. So the same applies here.”</p><p>I can attest to this working very well, and nearly eliminating any hallucinations.</p><p>Advanced reasoning also isn’t there yet, but it may be in the near future. Other methods for both prompting and training techniques are currently being explored.</p><p>I think more generally, a lot of these techniques fall into the bucket of what I would say, recreating our system 2. So you might be familiar with the system 1, system 2 thinking for humans. System 1 is a fast, automatic process. And I think kind of corresponds to like an LLM, just sampling tokens. And system 2 is the slower, deliberate planning sort of part of your brain. And so this is a paper actually from just last week, because this space is pretty quickly evolving. It's called “Tree of Thought”. And in “Tree of Thought”, the authors of this paper propose maintaining multiple completions for any given prompt. And then they are also scoring them along the way and keeping the ones that are going well, if that makes sense. And so a lot of people are really playing around with kind of prompt engineering to basically bring back some of these abilities that we sort of have in our brain for LLMs.</p><p>Karpathy mentions “Tree of Thought” as one solution to LLM problem solving. This is one of a bunch of different exploratory efforts going on to hugely improve results without even retraining the base models.</p><p><a target="_blank" href="https://openai.com/research/improving-mathematical-reasoning-with-process-supervision">OpenAI also recently said they aligned a model</a> via process supervision, rather than outcome supervision. What this means is that, when training, instead of rewarding it for getting the “right” answer they reward it for thinking through the steps of the problem itself, called “chain-of-thought” reasoning. It seems to me that models like this will really allow future GPT versions to do more advanced reasoning.</p><p>And if you pay for ChatGPT Pro, as of last week you now have access to Code Interpreter. This is <a target="_blank" href="https://www.latent.space/p/code-interpreter#details">being described by some as GPT-4.5</a> because of its ability to write and troubleshoot code to solve problems. (It’s seriously good — try if you can!)</p><p>What does AI allow us to do?</p><p>Abilities fall into 3 major categories for me:</p><p>* They make us smarter.</p><p>* They make us more creative.</p><p>* They make it much easier to communicate with computers and other people.</p><p>3. Easier to communicate</p><p><strong>Language models act as a translation or mapping layer between humans and machines</strong>. So if I describe something in English to an LLM it could translate to Mandarin so someone in China could understand, Python so it can be executed as code, or be reworded in English so a 5 year old gets it.</p><p>Part of this “mapping” is routing tasks to the right tools, where we then can use them and have finer UI controls. In <a target="_blank" href="https://www.newyorker.com/science/annals-of-artificial-intelligence/there-is-no-ai">a recent New Yorker article</a>, the writer talked about AI making computers less rigid:</p><p>Many of the uses of A.I. that I like rest on advantages we gain when computers get less rigid. Digital stuff as we have known it has a brittle quality that forces people to conform to it, rather than assess it. We’ve all endured the agony of watching some poor soul at a doctor’s office struggle to do the expected thing on a front-desk screen. The face contorts; humanity is undermined. The need to conform to digital designs has created an ambient expectation of human subservience. A positive spin on A.I. is that it might spell the end of this torture, if we use it well. We can now imagine a Web site that reformulates itself on the fly for someone who is color-blind, say, or a site that tailors itself to someone’s particular cognitive abilities and styles. A humanist like me wants people to have more control, rather than be overly influenced or guided by technology. Flexibility may give us back some agency.</p><p>Design researcher Jakob Nielsen, the business partner of Don Norman (of “Design of Everyday Things” fame) <a target="_blank" href="https://www.nngroup.com/articles/ai-paradigm">wrote a short post recently on how generative AI is introducing the third user-interface paradigm in computing history</a>. Unlike the current paradigm of command-based interaction design, where users issue commands to the computer one at a time to produce a desired outcome, users now tell the computer directly what they want.</p><p>With the new AI systems, the user no longer tells the computer what to do. Rather, the user tells the computer what outcome they want. Thus, the third UI paradigm, represented by current generative AI, is intent-based outcome specification. . . . the user tells the computer the desired result but does not specify how this outcome should be accomplished. Compared to traditional command-based interaction, this paradigm completely reverses the locus of control.</p><p>But are we destined to only be interacting with these AI models through speaking or writing? Hopefully not, <a target="_blank" href="https://futureblind.com/p/roundup-6-design-is-a-meta-skill">as I discussed in my last essay</a>. Nielsen continues:</p><p>The current chat-based interaction style also suffers from requiring users to write out their problems as prose text. Based on recent literacy research, I deem it likely that half the population in rich countries is not articulate enough to get good results from one of the current AI bots. That said, the AI user interface represents a different paradigm of the interaction between humans and computers — a paradigm that holds much promise. . . . Future AI systems will likely have a hybrid user interface that combines elements of both intent-based and command-based interfaces while still retaining many GUI elements.</p><p>2. The Creativity Frontier</p><p><strong>The second ability is that AI makes us more creative</strong>. It turn us into explorers of possibility space.</p><p>What do I mean by possibility space? You’ll also see it called meaning space or latent space, which for these purposes are interchangeable.</p><p>It can be a complicated topic to explain, especially without any visual aids. I’ll put a few links in the footnotes, and eventually I hope to write a post specifically about this. Here’s a passage from Stephen Wolfram’s “What is ChatGPT Doing?” essay:</p><p>One can think of an embedding as a way to try to represent the “essence” of something by an array of numbers—with the property that “nearby things” are represented by nearby numbers.</p><p>And so, for example, we can think of a word embedding as trying to lay out words in a kind of “meaning space” in which words that are somehow “nearby in meaning” appear nearby in the embedding.</p><p>. . . Roughly the idea is to look at large amounts of text (here 5 billion words from the web) and then see “how similar” the “environments” are in which different words appear. So, for example, “alligator” and “crocodile” will often appear almost interchangeably in otherwise similar sentences, and that means they’ll be placed nearby in the embedding. But “turnip” and “eagle” won’t tend to appear in otherwise similar sentences, so they’ll be placed far apart in the embedding.</p><p>Everyone is familiar with seeing data represented on a 2 dimensional chart with Y and X axes. That’s all these “embeddings” are — coordinates in a space, but rather than 2 dimensions, they have thousands. So we can’t really visualize them easily. Humans can’t see more than 3 dimensions so we have to just take a slice of them in 2D or 3D (or otherwise reduce the dimensionality).</p><p>In this meaning space, as Wolfram said, similar things are closer together. It works the same way in the text and art models. Alligator and crocodile are close in word space. In image space, an apple and an orange are close together, just as the color red would be close to roses.</p><p>What ChatGPT and Midjourney are doing is allowing us to “navigate” this space by using writing and text to explore it. Language models complete the text by finding higher probability combinations nearby your prompt. Art models associate your text prompt with nearby imagery in the space.</p><p>Daniel Gross put it similarly <a target="_blank" href="https://stratechery.com/2022/an-interview-with-daniel-gross-and-nat-friedman-about-chatgpt-and-the-near-term-future-of-ai/">in an interview last year with Ben Thompson</a>:</p><p>The models are just roaming through this embedded space they’ve created looking for things that are near each other in this multi-dimensional world. And sometimes the statistically most-next-likely thing isn’t the thing we as humans sort of agree on is truth.</p><p>He was referring to the negative side here of course, of hallucinations. But what’s a negative for discovering facts actually helps us be more creative.</p><p>One of my favorite authors, Kevin Kelly, has a new book out and has been making the podcast rounds recently. <a target="_blank" href="https://tim.blog/2023/04/28/kevin-kelly-excellent-advice-for-living-transcript/">Here he is on The Tim Ferriss Show</a> talking about how he uses MidJourney:</p><p>It’s like photography. I feel I have some of the same kind of a stance that I have and I’m photographing, I’m kind of hunting, searching through it, I’m trying to find a good position, a good area where there’s kind of promise. And I’m moving around and I’m trying, and I’m whispering to the AI, how about this? I’m changing the word order. I’m actually interacting, having a conversation with it over time. And it might take a half hour or more to get an image that I am happy with. And I’m at that point, very comfortable in putting my name as a co-creator of it because I have, me and the intern have worked together to make this thing.</p><p>This is exactly what Kevin is doing. The MidJourney model is generating the image for him, but mostly it’s helping him explore possibility space until he finds something he likes.</p><p>Here he is <a target="_blank" href="https://www.libertyrpf.com/p/kevin-kelly-excellent-advice-for#details">in another interview on Liberty’s Highlights</a> talking about how these models make us more creative and add leverage to our capabilities:</p><p>Yeah, the way you think of the kind of the prompters, which are the new artists, is that they’re like directors of a film. They’re commanding a whole bunch of different entities and agencies that do the craft. And their craft is orchestrating creative ideas and they have crafts people who do the cinematography and do the music, the score and everything. And they're kind of curating or directing the thing, or a producer in music. So we now have that kind of on images. And the important thing is that enables many more people to participate in creativity. And I say we have synthetic creativity. These engines are absolutely 100% lowercase creative. They are creative. There’s no getting around it. We can’t deny it. But it's a lowercase, small “c”, creativity. It's not the uppercase creativity that we have, like with a Guernica painting or Picasso or something. So the idea is that those we can't synthesize yet, maybe someday, but not right now. And so that lowercase creativity is synthesized and it's available to people. And that’s a huge innovation.</p><p>So this exploration of possibility space is a key feature of generative AI models.</p><p>At the same time, it’s pretty clear to me that <strong>we need a better way to navigate this space</strong>. Right now we’re doing it very haphazardly using trial and error with text prompting. But this doesn’t have to be the case.</p><p>Just like a geneticist discovering which genes code for certain traits (maybe it’s a few genes for eye color, or a lot for personality traits), we can discover which neurons and dimensions represent certain areas of meaning space.</p><p>This is why further research on how LLMs work is so useful.</p><p>If we can tweak parameters that represent attributes like humor, verbosity, complexity of writing… then we can manually adjust them like on a dashboard. You could increase the humor of your writing by 10% and decrease the length by 5%. Same goes for art models. How cool would it be to have slider adjustments for realism, color, detail, mood, or whatever?</p><p>I could keep going, but I’ll save it for another time.</p><p>1. AI = “Augmented Intelligence”</p><p>Rather than Artificial Intelligence, it’s better to think of AI as “Augmented Intelligence”. More than anything, it acts as a multiplier of our own intelligence — just like computers did decades ago, but much more so.</p><p>This is Marc Andreessen on a recent episode of Lex Fridman’s podcast talking about the benefits that AI will bring:</p><p><strong>Marc</strong>: And certainly at the collective level, we could talk about the collective effect of just having more intelligence in the world, which will have very big payoff. But there's also just at the individual level, like what if every person has a machine, and it's the concept of augment —Doug Engelbart's concept of augmentation. What if everybody has a an assistant, and the assistant is 140 IQ, and you happen to be 110 IQ. And you've got something that basically is infinitely patient and knows everything about you, and is pulling for you in every possible way — wants you to be successful. And anytime you find anything confusing, or want to learn anything, or have trouble understanding something, or want to figure out what to do in a situation — right, when I figure out how to prepare for a job interview, like any of these things, like it will help you do it. And it will therefore — the combination will effectively be — because it will effectively raise your IQ will therefore raise the odds of successful life outcomes in all these areas.</p><p><strong>Lex</strong>: So people below the hypothetical 140 IQ, it’ll pull them off towards 140 IQ.</p><p><strong>Marc</strong>: And then of course, people at 140 IQ will be able to have a peer, right, to be able to communicate, which is great. And then people above 140 IQ will have an assistant that they can farm things out to.</p><p>Packy McCormick wrote a few good essays on this topic recently. In “<a target="_blank" href="https://www.notboring.co/p/intelligence-superabundance">Intelligence Superabundance</a>” he addresses the concern that AI will steal our jobs and obsolete humans:</p><p>. . . <strong>the increased supply of intelligence will create more demand for tasks that require intelligence</strong>, that we’ll turn gains in intelligence efficiency into ways of doing new things that weren’t previously feasible.</p><p>. . . We can examine it a little more rigorously with a little mental flip. Instead of thinking about AI as something separate from human intelligence, we should think of it as an <strong>increase in the overall supply of intelligence</strong>. . . . Then the important question becomes not “Will I lose my job?” but <strong>“What would I do with superabundant intelligence?”</strong></p><p>Essentially, demand for intelligence is very elastic. Meaning when either the price goes down or the supply goes up, the demand goes up a lot to match it. If AIs give us more intelligence, our demand and uses for it will increase.</p><p>Packy writes more <a target="_blank" href="https://www.notboring.co/p/evolving-minds">in another essay called “Evolving Minds”</a>:</p><p>In that case, I don’t think the framing of AI as simply a tool that we’ll use to do things for us is quite right. I think we’ll use what we learn building AI to better understand how our minds work, define the things that humans are uniquely capable of, and use AI as part of a basket of techniques we use to enhance our uniquely human abilities.</p><p>Even as AI gets smarter, it will fuel us to get smarter and more creative, too. Chess and Go provide useful case studies.</p><p>He goes on to write about how AI became much better than us at chess and Go, but instead of giving up, humans got better — and more creative — as well:</p><p>Some people will be happy letting AI do the grunt work. Not everyone will see AI’s advance as a challenge to meet, just as I have not taken advantage of chess and Go engines to become better at either of those games. But I think, as in chess and Go, <em>more</em> people will get smarter. The process will produce more geniuses, broadly defined, ready to deliver the fresh, human insights required to ignite the next intellectual revolution.</p><p>If we can walk the line, the result, I think, will be a deeper understanding of the universe and ourselves, and a meaningful next leg in the neverending quest to understand and shape our realities.</p><p>When we think about this augmented intelligence, it’s not necessarily just “ChatGPT” or some other generic AI. Future versions will be very much customized. Not only will they have your data, but they’ll share your system of values. This is Sam Altman on this topic from an interview early this year:</p><p>. . . the future I would like to see is where access to AI is super democratized, where there are several AGIs in the world that can kind of help allow for multiple viewpoints and not have any one get too powerful, and that the cost of intelligence and energy, because it gets commoditized, trends down and down and down, and the massive surplus there, access to the systems, eventually governance of the systems, benefits all of us.</p><p>. . . We now have these language models that can understand language, and so we can say, hey model here's what we'd like you to do, here are the values we'd like you to align to. And we don't have it working perfectly yet, but it works a little and it'll get better and better.</p><p>And the world can say all right, here are the rules, here's the very broad-bounds absolute rules of a system. But within that people should be allowed very different things that they want their AI to do. And so if you want the super never offend safe for work model you should get that. And if you want an edgier one that is sort of creative and exploratory but says some stuff you might not be comfortable with or some people might not be comfortable with, you should get that.</p><p>So this all sounds great, but if base-level intelligence is commoditized, how can humans stand out and differentiate? I think there’s a lot of answers to this — some of them maybe even philosophical. I’ll leave those for another time, and just leave you with a clip from Liberty’s interview with Kevin Kelly:</p><p><strong>Liberty</strong>: And in the book that would be compressed as don't be the best, be the only, this reminds me of — Naval has a way of saying it, which is “escape competition through authenticity”, right? If someone else is doing the exact same thing, well, maybe a robot can do it. Right?</p><p><strong>Kevin</strong>: Exactly. That's another piece of advice in the book is that if your views on one matter can be predicted from your views on another matter, you should examine yourself because you may be in a grip of an ideologue. And what you actually want is you actually want to have a life and views that are not predictable by your previous ones. And the advantage there is that you are going to be less replaceable by an AI.</p><p><strong>Liberty</strong>: Yeah. If everybody's thinking the same, a lot of people are not thinking in there.</p><p><strong>Kevin</strong>: Well, but the AIs are auto-complete. They're just kind of auto-completing. And if you're easy to auto-complete, you're easy to be replaced.</p><p><strong>Liberty</strong>: That's a good way to put it. Optimize for being hard to auto-complete.</p><p>🧑‍🏫 Education Frontier</p><p>Education — specifically K-12 education — has a lot of room for improvement. I’ve written a few things about the future of education many years ago, but haven’t really talked about it much. With a 2-year-old now, I follow it a little closer so I want to share some interesting content I’ve seen recently.</p><p>Ryan Delk founded a startup called Primer that’s working to build a new education system in the US by helping teachers launch their own micro-schools with 10-25 students that provide more personalized education without all the bureaucracy that comes with big school systems.</p><p><a target="_blank" href="https://overcast.fm/+u18yASOBA">He was interviewed recently on a podcast called The Deep End</a>, and here is is giving some more detail on what they do:</p><p>So the micro schools today that we have are, they're all led by one teacher as sort of the main micro school leader. Depending on the number of kids, this is part of what's really cool about it is not only can teachers make way more money than they do in the traditional system. So on average, our teachers probably make 40% to 50% more than what they'd make in the existing districts that they teach in. But they also can get an aid to help them out in their class.</p><p>And so at a certain threshold, they unlock an aid that will help them and can also sort of assist with different subjects and stuff. So you have one micro-school leader. For elementary, as you mentioned, it's very standard in the US that an elementary teacher teaches all subjects. And at some middle schools or a lot of middle schools, sometimes that's also true, where one teacher's teaching subjects. And so we use that model.</p><p>Right now, most of our kids are in elementary, third through fifth, third through sixth grade. And so that's the current model. And then how we augment that is with online, what we call them as subject matter experts. And so we have, for example, a really exceptional ELA and a really exceptional reading teacher, really exceptional math teachers. And so for specific subjects where kids, whether it's that they need remedial help, or they're advanced and they need more support, and the micro school leader can't support them directly or just isn't set up to support them directly because of the structure of the class or whatever, kids have these resources.</p><p>This brings us to the question of — what is school actually providing to students?</p><p>I think a lot of people that are in the education sort of startup space have this sort of like, oh, we're going to take over the world and disrupt everything. And, you know, we need to blow up the whole existing system. My view on it is maybe a bit more nuanced than that. And I think that the public school system in the U.S. right now, I think it serves like three very distinct functions. And we've bundled all that into one name, which is like public schools.</p><p>Everyone thinks about the academic component. Hey, we're sort of this is the engine that is responsible for the academic development of the children of the U.S. And that's certainly one component of it. And we could debate how effective it is at that. But another component that's sort of the two other components I think are maybe equally important societally, but are very under discussed.</p><p>One is that for millions of kids, it's actually basically the way that they're able to get food and have food security each day. And we sort of totally dismiss this, but it's actually, I think, a very important and very sad and unfortunate sort of component of public schools. You can do the numbers and the research on this, and it will blow your mind.</p><p>And the second component, the third component is that, you know, most parents are working and they need a place for their kids to go from you know eight to, with aftercare, eight to five every day so they can work. And so there's this babysitting component of school that's very important.</p><p>This reminds me of the famous Good Will Hunting scene (and meme):</p><p>As we all know, especially with the internet, you can learn everything they teach you at Harvard yourself. But obviously that’s not the real reason why people want to go to Harvard. Just like universities, all school is a bundle of services:</p><p>* Learning</p><p>* Daycare / babysitting — even if a computer could handle 100% of education, you wouldn’t leave a 7 year old alone all day while you work.</p><p>* Social connections</p><p>* Tutoring</p><p>* Working with others</p><p>* And as Ryan mentioned, providing food.</p><p>So it’s pretty clear we’re going to have somewhat of an “unbundling” of these services over time. Into what, I don’t know. What I do know is that AI will play a bigger and bigger role going forward.</p><p><a target="_blank" href="https://www.ted.com/talks/sal_khan_how_ai_could_save_not_destroy_education/c">Sal Khan did a TED talk earlier this year</a> where he talked about The 2 Sigma Problem, a study from the 1980s where if you give one-on-one tutoring to students, you can improve their outcomes by 2 standard deviations. This takes an average student and turns them into an exceptional student. The problem here of course is that this is very hard to scale.</p><p>Enter AI. Khan Academy has a new feature called Khanmigo that integrates GPT-4:</p><p>This is really going to be a super tutor. And it's not just exercises. It understands what you're watching. It understands the context of your video. It can answer the age-old question, why do I need to learn this? And it asks, socratically, well, what do you care about? And let's say the student says, I want to be a professional athlete. And it says, well, learning about the size of cells, which is what this video is about, is really useful for understanding nutrition and how your body works, et cetera. It can answer questions, it can quiz you, it can connect it to other ideas. You can now ask as many questions of a video as you could ever dream of. </p><p>So another big shortage out there, I remember the high school I went to, the student-to-guidance counselor ratio was about 200 or 300 to one. In a lot of the country, it's worse than that. We can use Khanmigo to give every student a guidance counselor, academic coach, career coach, life coach, which is exactly what you see right over here.</p><p>With this, students can get truly custom learning experiences. That’s pretty awesome to me because in school there can be more hands-on activities and applications. Things that are hard to beat on a computer.</p><p>In <a target="_blank" href="https://www.infiniteloopspodcast.com/ethan-mollick-how-ai-changes-everything-ep165/">an interview on Jim O’Shaughnessy’s podcast “Infinite Loops”, Ethan Mollick</a> has a great bit about this:</p><p>So, I think in much of the world, we're going to see a real transformation of teaching, where teaching is not high quality right now and we're going to see a leap into another model.</p><p>But in the US, in Europe and probably in China, Japan, other places with that already have fairly highly developed educational systems, I think the future looks like one where you are going to still go to school as you did before, during the same class period times, but you're going to be basically... The easiest version to think about, is you're going to be doing a much more engaging version of homework in school. Project-based learning, experiences, building stuff, trying to create things, trying to apply things. Outside of school you're going, your homework assignments are going to be monitored by AI and you'll be working with an AI tutor that will generate lessons for you and teach you concepts, and make sure you're on track and will let the teacher know "The person's struggling at this, despite this," and will help the teacher set up teams so that you don't have those horrible team assignments and track grades. So, I think that there's a model that we already know works. We'll still give lectures. It's like 5,000 years old, giving lectures, and we know that it doesn’t work, but we keep doing them. So, it's going to force us to shake ourselves out of a bunch of stuff that doesn't work well. I'm sure there'll still be some classes that look like classes. I think English composition will still look like English composition. I think math is still going to look a lot like math, but I think a lot of your other projects will be more...</p><p>I’ll finish this section with another clip from Ryan Delk on blending the virtual and real-world experiences:</p><p>So I think of it as the best possible experience is absolutely a best teacher in person in the smallest possible environment, meaning the fewest amount of students to the teacher ratio. That's the best. That's by far the best situation that you could get any kid into. But that's sort of like a pipe dream for most kids in the US right now, that you're just not going to have this amazing world-class teacher that can sit down with 12 kids and go deep about some specific thing you're excited about. So I think I agree that that's the optimal scenario.</p><p>I'm not one of these people that thinks virtual is inherently better. But I do think that virtual with a world-class teacher, especially if you can build interesting product experiences, which we're building on top of the virtual experience, when needed, and accompanying an in-person teacher, is better than a sort of average in-person experience.</p><p>🔗 Interesting Content</p><p><em>Recent content I’ve enjoyed and think you might also.</em></p><p>David Senra on Steve Jobs</p><p>One of my favorite recent episodes of The Founders podcast was <a target="_blank" href="https://www.joincolossus.com/episodes/12538515/senra-steve-jobs-make-something-wonderful">where David talked about the recently released book</a> “<a target="_blank" href="https://book.stevejobsarchive.com/">Make Something Wonderful</a>” which is a selection of Steve Jobs’ writings.</p><p>That is what your life is. Think of your life as a rainbow. This is again, maybe my favorite quote in the entire book. Think of your life as a rainbow arcing across the horizon of this world. You appear, have a chance to blaze in the sky, and then you disappear. What is he saying? Go for it like your life depends on it. Why? Because it does. It literally does. The two endpoints of everyone's rainbow. This is, this is again, I think one of the, he's got a ton of great ideas, right? But something you'll, if you study Steve, what he says over and over again, it's like if you need to use your inevitable death as a tool to make decisions on how you're going to spend your time, the two endpoints of everyone's rainbow are birth and death. We all experience both completely alone. And yet most people of your age have not thought about these events very much, much less even seen them in others.</p><p>How many of you have seen the birth of another human? It's a miracle. And how many of you have witnessed the death of a human? It is a mystery beyond our comprehension. No human alive knows what happens to us after our death. Some believe this, others that, but no one really knows. Again, most people your age have not thought about these events very much. It's as if we shelter you from them, afraid that the thought of mortality will somehow wound you. For me, it's the opposite. To know my arc will fall makes me want to blaze while I'm in the sky. Not for others, but for myself, for the trail that I know that I'm leaving. And I love the way he ends his speech. Now, as you live your arc across the sky, you want to have as few regrets as possible. Remember, regrets are different from mistakes.</p><p>Lots of good wisdom in this one, so make sure to check it out. I’ll call out <a target="_blank" href="https://www.joincolossus.com/episodes/72775590/senra-arnold-schwarzenegger-barbara-outland-baker">his just-released one on Arnold Schwarzenegger</a> as well, which is a good compliment to the Netflix documentary I highly recommend.</p><p>Thomas Edison’s Lab</p><p>Eric Gilliam <a target="_blank" href="https://worksinprogress.co/issue/thomas-edison-tinkerer">wrote a great essay for Works in Progress on Thomas Edison</a>. Highly recommend this and I got a lot out of it. He also did <a target="_blank" href="https://www.freaktakes.com/p/tales-of-edisons-lab#details">a companion solo podcast called “Tales of Edison’s Lab”</a>. Edison was one of those once-in-a-century type people that to me really just fits the American “man in the arena” prototype.</p><p>In a field full of uncertain potentials, he often did seem to proceed to his triumphs by a certain kind of laborious sorcery, ill prepared in mathematics, disrespectful of the boundaries set by available theory. He took on all the forces of nature with a combination, as he said, rather unfairly about himself, of 2% inspiration and 98% hard work. The inspiration was informed by years of practice and shrewd observation. The hard work was simply amazing. And if the means by which he brought off his extraordinary efforts are not wholly clear, neither is the cause for his obsessive labors, no diver into nature's deepest mysteries, caring next to nothing for the advancement of knowledge and even less for the world's goods. He would become absorbed in making something work well enough to make money. The test in the marketplace was for him, apparently, the moment of truth for his experiments. And once this was passed, he became forthwith, absorbed in making something else. By the time of Edison's foray into lighting, he was a commercially successful inventor many times over.</p><p>Acquired on Lockheed Martin</p><p>Ben and David from the <a target="_blank" href="https://www.acquired.fm/episodes/lockheed-martin">Acquired Podcast recently did an episode on Lockheed</a> .</p><p>It’s a 3+ hour episode so there’s a lot there. They cover Kelly Johnson’s creation of Lockheed’s famous Skunkworks division:</p><p>So in this circus tent in a parking lot, Kelly and this super elite team from Lockheed build the first prototype US fighter jet named the Lulu Bell in 143 days start to finish. This is just wild. For years the US had been working on this technology and they hadn't gotten it operationalized. The Germans beat them to it and then in 143 days, Kelly and Lockheed go from zero to flying prototype. What a testament to him and to this organization in the circus tent that he has built, the Skunkworks.</p><p>I used Skunkworks as an example a few times in my essay and podcast from last year, <a target="_blank" href="https://futureblind.com/p/take-the-iterative-path">Take the Iterative Path</a> on how Skunkworks and SpaceX both use iterative design to build so quickly. So check that out if that interests you.</p><p>On Acquired they also talk about Lockheed’s space and missile division, which I had actually never heard of. And apparently it did pretty well for them:</p><p>Well, I think at times in the 60s and 70s and 80s, LMSC was the largest business by revenue. But almost through the whole time, it was by far the most profitable division within Lockheed. And at times, when we'll get into, Lockheed fell on some really hard times in the 70s. There were years where LMSC generated more than 100% of the profits of Lockheed. So all of the rest of Lockheed, Skunkworks included, was in the red. Unprofitable, bleeding money, and LMSC was keeping the company afloat.</p><p>. . .This is much more technology problems and computing problems that LMSC is tackling here. Yes, they're building missiles. Yes, they're building rockets and all that. But the core value components of those rockets is computing and silicon and ultimately software. And as we talk about all the time on this show, like, well, that's really good margins. Definitely better margins than building airplanes.</p><p>Apple Vision Pro</p><p>There’s been lots of good content out — from testers and analysts — about Apple’s Vision Pro mixed reality device.</p><p>As usual, <a target="_blank" href="https://stratechery.com/2023/apple-vision/">Ben Thompson had a good writeup</a> and was really impressed with the half-hour demo he got:</p><p>As I noted above, I have been relatively optimistic about VR, in part because <a target="_blank" href="https://stratechery.com/2022/meta-meets-microsoft/"><strong>I believe the most compelling use case is for work</strong></a>. First, if a device actually makes someone more productive, it is far easier to justify the cost. Second, while it is a barrier to actually put on a headset — to go back to my VR/AR framing above, a headset is a destination device — work is a destination.</p><p>. . . Again, I’m already biased on this point, given both my prediction and personal workflow, but if the Vision Pro is a success, I think that an important part of its market will to at first be used alongside a Mac, and as the native app ecosystem develops, to be used in place of one. To put it even more strongly, the Vision Pro is, I suspect, the future of the Mac.</p><p>I agree with Ben’s belief that Vision is ultimately set up to replace some segment of the Mac’s jobs-to-be-done, and ultimately become an amazing productivity tool.</p><p>Just as websites on smartphones were originally the same desktop version, Apple Vision shows users interacting with several 2D screens floating in space. This alone can be a huge productivity boost, as evident by the boost you get from multiple monitors today.</p><p>But the real boost will come when the mixed reality interface breaks out of the past paradigm and you can interact with both the physical world and an infinite digital canvas. <a target="_blank" href="https://notes.andymatuschak.org/Vision%20Pro">Andy Matuschak wrote about this following the announcement</a>, talking about ideas like having apps or files live in physical space. Like a timer above your stove, or the Instacart app in your fridge.</p><p>It will be interesting to see what kinds of stories and experiences come out of the mixed reality world. Steven Johnson, who’s one of my favorite writers, <a target="_blank" href="https://sites.libsyn.com/468759/views-of-xr-from-a-historic-futurist">did an interview on VR and AR right before the Vision Pro announcement</a>. One of my favorite parts was when he talks about the history of immersive experiences:</p><p>There's this wonderful period, I don't know, like 250 years ago, where there was this amazing explosion of immersive, I call them kind of palaces of illusion. And they were everything, you know, there's a wonderful example of like the panorama, which was the 360 degree painting where you would come in and see a famous military battle reenacted or a skyline view of London and things like that. And there was just a huge number of experiments that happened back then, where they would try all these different configurations of ways to like, create the illusion of being in some other world.</p><p>Interestingly, they were not narrative environments, they really didn't tell a story. It was just the kind of like the body, the sensory feeling of, oh, I'm in this like other world, and I can occupy that world for a bit of time and then go back into reality. And that there was just, again, a lot of mucking around here for, you know, figuring out ways to kind of captivate people.</p><p>And it was kind of like a Cambrian explosion of all these different possibilities that then eventually consolidated basically into the cinema, right? There are all these different ways of creating the illusion that they're off in some other world. And then we kind of settled on this one approach and all those other experiments died off, basically. And the other thing that happened in that period is directly related to VR and AR is the invention of the stereoscope, which was one of these great cases where it was a scientific innovation about stereo vision.</p><p>That’s all for now — I hope you learned something interesting from this experiment!</p> <br/><br/>Get full access to FutureBlind at <a href="https://futureblind.com/subscribe?utm_medium=podcast&#38;utm_campaign=CTA_4">futureblind.com/subscribe</a>
play-circle icon
54 MIN
Take the Iterative Path
SEP 27, 2022
Take the Iterative Path
<p>One of the greatest business successes over the last 20 years has been SpaceX’s rise to dominance. SpaceX now launches more rockets to orbit than any other company (or nation) in the world. They seem to move fast on every level, out executing and out innovating everyone in the industry.</p><p>Their story has been rightfully told as one of engineering brilliance and determination.</p><p>But at its core, the key their success is much simpler.</p><p>There’s a clue in this NASA report on the Commercial Crew Program:</p><p>SpaceX and Boeing have very different philosophies in terms of how they develop hardware. SpaceX focuses on rapidly iterating through a build-test-learn approach that drives modifications toward design maturity. Boeing utilizes a well-established systems engineering methodology targeted at an initial investment in engineering studies and analysis to mature the system design prior to building and testing the hardware. Each approach has advantages and disadvantages.</p><p><em>This</em> is the heart of why SpaceX won. They take an iterative path.</p><p>Taking the determinate path</p><p>Let’s talk about the Boeing philosophy first, which is the most common approach taken by other traditional aerospace companies. “There are basically two approaches to building complex systems like rockets: linear and iterative design,” Eric Berger writes in the book “Liftoff” about the early history of SpaceX:</p><p>The linear method begins with an initial goal, and moves through developing requirements to meet that goal, followed by numerous qualification tests of subsystems before assembling them into the major pieces of the rocket, such as its structures, propulsion, and avionics. With linear design, years are spent engineering a project before development begins. This is because it is difficult, time-consuming, and expensive to modify a design and requirements after beginning to build hardware.</p><p>I call this the “determinate path” — in trying to accomplish a goal, the path to get there is planned and fixed in advance.</p><p>In project management this method is called <em>waterfall</em>, an “approach that emphasizes a linear progression from beginning to end of a project. This methodology, often used by engineers, is front-loaded to rely on careful planning, detailed documentation, and consecutive execution.”</p><p>Spend a lot of time scoping and planning carefully upfront, then move progressively forward step-by-step. This is the “measure twice, cut once” approach.</p><p>You may be familiar with it as it’s very common in organizations everywhere.</p><p>There can be many reasons why this path would be taken:</p><p>* If from the start you have very clear, unambiguous requirements (from customer, management, etc.)</p><p>* If you think you can figure out how exactly to build something before building it, you’d probably want to plan it all in advance.</p><p>* If your fixed costs are high, it can <em>force</em> you to make decisions up front. Take traditional auto manufacturing. A door mold machine might cost $50 or $100M, so you have to figure out what the design of the door will be first. (But this means if later they have a new idea for a better car door, they don’t want to change it because of the sunk costs of the mold machine.)</p><p>* You have a lot of resources, which makes you think you can just brute force it and overwhelm the problem with money and people. (Many overfunded startups are guilty of this.)</p><p>But there <em>is</em> another way . . .</p><p>Taking the iterative path</p><p>When I think of the most impactful technologies over the last 100 years, nearly all were created by small teams of tinkerers.</p><p>Why? It’s easier for these teams to take an iterative path.</p><p>Taking this path means rapid prototyping, testing concepts against reality, failing, and adapting. Continuing from the book “Liftoff”:</p><p>The iterative approach begins with a goal and almost immediately leaps into concept designs, bench tests, and prototypes. The mantra with this approach is build and test early, find failures, and adapt.</p><p>Focus more on building and finding failure modes than making things perfect. Project managers call it “agile”, or at Facebook, “move fast and break things.”</p><p>The canonical example of this to me is the Wright brothers, previously bicycle mechanics, building iterations of their airplane design over and over, and failing until they succeeded.</p><p>This approach ended up being common in the origin stories of all airplane manufacturers and defense companies — Martin Marietta, Lockheed, Northrop Grumman, etc., where again you had relatively small teams of self-taught tinkerers building complex machines through a process of iteration, failure, and learning until they succeed.</p><p>How can you reconcile this “fail fast” approach with the care that’s needed to reliably build things where human lives are on the line?</p><p>The answer is that these can be two different parts of the organization. Working together, but with different focuses. “[SpaceX is] launching 5 or 6 times a month and on their pads they need operational excellence with zero risk — you know, they’re doing innovation but it’s minimal innovation. Blowing things up on the pad is not a good idea — you want that down to zero because human lives and certainly lots of capital is at risk.” This is <a target="_blank" href="https://overcast.fm/+a0Bqzb_2Q/13:26">Steve Blank on a recent Village Global podcast</a>. He continues:</p><p>But on the other hand, they have another part of the company that in fact believes in not only blowing things up on the test pad — because if you’re not doing that you’re not pushing the envelope fast enough — it’s the cycle time of doing that. So they have an agile innovation process.</p><p>Now think about that. This is the same company doing two very different things with two different groups of people, two different risk profiles, but more importantly they’re talking to each other. It’s not “here are the smart people, and here are the people turning the crank,” they’re learning from each other. The guys building the raptor engines and Starship need to know where the GFC plugs in and what the right materials and things they need to get right on the next rocket. And the people doing the existing rockets can learn about new materials and incremental upgrades so they are innovating but innovating with minimal risk.</p><p>The iterative path is easier to take when you’re nimble and the cost of failure is low. This is why it’s so common in software. But as the previously mentioned companies have shown, it’s also the best approach in hardware and complex, frontier tech.</p><p>And just as the traditional aerospace companies have demonstrated, organizations that are very bureaucratic <em>now</em> were almost always more iterative in the past.</p><p>The early history of Lockheed’s Skunk Works division is informative, which I believe later served as one of the models for SpaceX’s approach. Skunk Works was an R&D group created by Kelly Johnson within Lockheed during the war in 1943 when they got the contract to build the P-80 Shooting Star. <a target="_blank" href="https://youtu.be/bukQbeaP_Cw?t=788">From a documentary on the birth of Skunk Works:</a></p><p>Lockheed was already swamped in terms of manpower, tooling, and facilities with wartime contracts but this was a blessing in disguise, an opportunity to implement an idea he’d been pestering Robert Gross about for years. Let him round up a small group of talented people: designers, engineers and shop men. Put them under one roof where they could all work closely together and give him complete authority over everything from procurement to flight tests.</p><p>Johnson gathered 28 engineers including himself, and 105 “shop men” (I assume this just means workers who can build what the engineers design) and built a small facility out of discarded shipping crates using a circus tent for a roof. He then laid out the original rules that would become the foundation for Skunk Works over the next 30 years:</p><p>. . . he’d be responsible for all decisions. Paperwork and red tape would be cut to the minimum. Each engineer would be designer, shop contact, parts chaser, and mechanic, and each would remain within a stone’s throw of the shop at all times. . . . Forcefully reminded that simplicity is the keynote of good design, the designers jumped into their work. But this was a new kind of operation, and instead of moving from stage to stage, the schedule demanded an extraordinary degree of concurrency.</p><p>The time from initial concept to delivery of the first P-80 to test pilots would be only 5 months. In fact, nearly all of the early planes coming out of Lockheed took less than 6 months — less than 6 months from concept to delivery. Crazy!</p><p>Even the famous A-12 (later the SR-71 Blackbird) look less than 4 years from initial idea to roll out. This may seem like a lot when you’re used to super-fast software timelines, but this is 4 years for one of the fastest, most successful aircraft <em>ever built</em>.</p><p>The scrappy culture lived on in later Skunk Works projects. <a target="_blank" href="https://www.youtube.com/watch?v=I9wAEjox2WI">This is Ben Rich, who led the division in later years, on their building of the F-117</a> (this is the <a target="_blank" href="https://en.wikipedia.org/wiki/Lockheed_F-117_Nighthawk">Darth-Vader-looking stealth fighter</a> you’ve probably seen before):</p><p>On the F-117, we had to get the guy to climb into the cockpit. So I went to the local builders mart, and bought one of these ladders for 50 bucks, and we just used it. . . . We didn’t have to spend thousands of dollars designing it for Mil spec — military specification — and we did simple things like that.</p><p>The more you learn about the history of building things, the more you hear stories like this, even with highly complex innovations. The development of the Sidewinder missile is another interesting example: again, small team, rapid iteration, creative solutions to problems.</p><p>Why is iteration better?</p><p>Taking the iterative path tests your model against reality, getting to the truth as fast as possible.</p><p>There are a few major downsides to the <em>linear</em> approach:</p><p>* Clear specs and requirements from the outset may seem like a good thing. Much of the time though they don’t match reality though. This is especially true in areas that are pushing the boundaries of innovation.</p><p>Over time, “the spec” becomes the most important thing. Here’s Ben Rich again, <a target="_blank" href="https://youtu.be/I9wAEjox2WI?t=305">on one of the requirements for building the SR-71 Blackbird</a>:</p><p>Some general insisted that there was a military spec that the SR-71 had to say ‘U.S. Airforce’ and the stars and bars. I said ‘General . . . you’re crazy.’ I said, you know, this has the temperature of an oven. Have you ever taken a piece of metal, painted it and stick it under your broiler? You can’t keep the paint on the metal. He said ‘No, the spec says you gotta say U.S. Air Force on our airplanes.’ I said we’ll develop it. So we spent a million dollars developing a paint that could show red, white and blue, and we put it on the airplane. . . . I mean, who’s going to see you at 90,000 feet?</p><p>* In this example and many others, politicians start dictating <em>how</em> work should be done, rather than just setting the goal like they should be. Conditions for funding become completely removed from the outcome itself, like mandates to use certain suppliers or base employees in certain states.</p><p>* The technical scope is too large, so that when there’s a problem, it’s hard to find the root cause. When there’s a problem you have to go back to the drawing board, but you may not even be able to do that given the cost to start over.</p><p>* You become too risk averse, fearing failure. This is pretty simple: if the costs to start something or fail are high, people don’t want to do new things. From the book “Liftoff”:</p><p>At most other aerospace companies, no employee wanted to make a mistake, lest it reflect badly on an annual performance review. Musk, by contrast, urged his team to move fast, build things, and break things.</p><p>* And from an executive of Blue Origin on what they do wrong:</p><p>I believe we study a little too much and do too little . . . More test [rather than] more analysis will allow us to progress more quickly, iterate, and eventually succeed.</p><p>A <em>good</em> iterative approach creates tight feedback loops, like John Boyd’s classic OODA loop: observe, orient, decide, act. Observe what’s going on, orient yourself to the environment, decide what needs to be done to make progress, act on that decision, and return to observing the results from your action. From the book “Certain to Win” on Boyd’s philosophy:</p><p>What does it take to win? <em>Agility</em> — the ability to rapidly change one’s orientation (worldview) in response to what’s happening in the external world.</p><p>This was referring to combat. But it’s just as true in business and engineering.</p><p>Tight feedback loops lead to a high rate of innovation and adaptation, quickly finding better solutions and what not to do. Speed is a tactical advantage.</p><p>“Innovation per year is what matters. Not innovation absent time. . . . What is your rate of innovation? that matters. And is the rate of innovation, is that accelerating or decelerating?” — <a target="_blank" href="https://www.youtube.com/watch?v=ReXzsnyDq_M"><strong>Elon Musk</strong></a></p><p>Brian Armstrong, founder of Coinbase, has a good saying that “action produces information”. You can’t predict the future, so just start building. And that’s what SpaceX did. Eric Berger writes that “the engineers designing the Falcon 1 rocket spent much of their time on the factory floor, testing ideas, rather than debating them. Talk less, do more.”</p><p>The iterative path in practice</p><p>So I’ve convinced you that iteration is best. What does it mean in practice?</p><p>Agile is straightforward for software, but not so much hardware.</p><p>Historically, linear/waterfall has been easier to do in hardware, especially for large engineering projects. A lot of upfront costs are needed — so spend a lot of time gathering requirements and planning before actually building.</p><p>For companies building big, complex, expensive things, it seems like a reasonable assumption that you have to know exactly what you’re doing, plan a lot, and be risk averse. As you know, this isn’t true! You can be fast, nimble, and agile even in megaprojects.</p><p>In complex hardware what this means is:</p><p>* Being hardware rich — having lots of spare parts and backups.</p><p>This lets you move quickly, and continue to try things over and over, because you have all these “at bats”. I’d include 3D printing as it allows you to create parts on an ad hoc basis.</p><p>* Using simulations — move atoms to bits when possible, giving you the freedom to quickly test and have all the benefits of software. If you can simulate what’s happening in the real world with enough accuracy, you can fail as much as you want. This is another area that has changed a lot in past decade or so. <a target="_blank" href="https://cliffberg.medium.com/spacexs-use-of-agile-methods-c63042178a33">Here is Cliff Berg on SpaceX’s use of simulation</a>:</p><p>SpaceX has invested a great deal of effort in automating the process of designing and validating through simulation, and delivering the machinery that they build through automation. They use traditional computer-aided design (CAD) tools such as CATIA, but they also invested in an end-to-end 3D modeling system from which they can view and simulate entire assemblies and automatically print parts. Importantly, the software is fast, even when handling complex assemblies, so that engineers do not have to wait, which encourages a rapid iterative design approach.</p><p>* And further <a target="_blank" href="https://www.youtube.com/watch?v=vYA0f6R5KAI&#38;t=376s">from a slide in a presentation</a> given by SpaceX Director of Research Adam Lichtl and Lead Software Engineer Stephen Jones:</p><p>Why Simulation?1. Investigate what cannot be measured2. Reduce need for testing3. Design optimisation: narrow design space4. Proactive instead of reationary design</p><p>* Constantly testing the whole system. Simulation is great but in the end you have to actually test <em>everything</em> out.</p><p>In these complex systems, in the end what really matters is how the whole thing behaves in the real world. This reminded me of how George Mueller, who led NASA’s spaceflight program during the Apollo missions in the 60s, approached building the Saturn rocket:</p><p>At a system level you’re much better off testing the system [rather than the individual parts] because in the end that system has to work. And then the only way you find out is if you test it as a system.</p><p>* Subsystems of the rocket would only be tested if needed.</p><p>* Utilizing “pathfinders”. In manufacturing, a pathfinder is an early build of something that won’t end up seeing the light of day.</p><p>You build a pathfinder to see where problems are, and how it can be done better. You know it will fail or be suboptimal, you’re just looking for how to do it better.</p><p>This is similar to a “tracer bullet”. You fire it first with no expectation of hitting the target, watch it, and then adjust your aim.</p><p>Doing all of these things, like SpaceX, leads to much faster iteration. This is something many other companies can learn from.</p><p>In summary: <strong>When a goal is big and complicated, an iterative, fail-fast approach is much better than a linear approach. Determinate paths lead to slower, poorly-adapted solutions, whereas iteration finds problems faster with a result that’s better adapted to the real world.</strong></p><p><em>Thanks to Rohit for providing feedback on the draft. All illustrations in this essay are a combination of DALL-E generations and my personal edits.</em></p><p><em>Intro/outro music: </em><a target="_blank" href="https://dg.lnk.to/Luminous"><em>“Cinnabar” by Roger & Brian Eno, from the album Luminous</em></a><em>.</em> </p> <br/><br/>Get full access to FutureBlind at <a href="https://futureblind.com/subscribe?utm_medium=podcast&#38;utm_campaign=CTA_4">futureblind.com/subscribe</a>
play-circle icon
19 MIN