rw-book-cover

Metadata

Highlights

  • I wanted to give some quick thoughts on the Apple AI (sorry, “Apple Intelligence”) release. I haven’t used it myself, and we don’t know everything about their approach, but I think the release highlights something important happening in AI right now: experimentation with four kinds of models - AI models, models of use, business models, and mental models of the future. What is worth paying attention to is how all the AI giants are trying many different approaches to see what works. (View Highlight)
  • Remember BloombergGPT, which was a specially trained finance LLM, drawing on all of Bloomberg’s data? It made a bunch of firms decide to train their own models to reap the benefits of their special information and data. You may not have seen that GPT-4 (the old, pre-turbo version with a small context window), without specialized finance training or special tools, beat BloombergGPT on almost all finance tasks. This demonstrates a pattern: the most advanced generalist AI models often outperform specialized models, even in the specific domains those specialized models were designed for. That means that if you want a model that can do a lot - reason over massive amounts of text, help you generate ideas, write in a non-robotic way — you want to use one of the three frontier models: GPT-4o, Gemini 1.5, or Claude 3 Opus. (View Highlight)
  • But these models are expensive to train and slow and expensive to run, which leaves room for much smaller models that aren’t as good as the frontier models but can run cheaply and easily - even on a PC or phone. This isn’t new. Back in December, I was able to run Mistral 7b, a model slightly less advanced than the original ChatGPT, directly on my phone without an internet connection. I also ran Mixtral, a model from the same company that slightly outperforms the original ChatGPT, on my gaming computer. (View Highlight)
  • All of the tech companies have been releasing these sorts of small, fairly powerful models, with the idea that they can handle simple questions on the hardware of your devices, and then call a larger model “in the cloud” when they need help. You aren’t getting anywhere near the smarts of a frontier model, but if you want to do straightforward things (make a Siri that works or “make my photos more vivid”), these models are often more than enough. Many of the companies betting on frontier models, like Google, have also released faster and cheaper models to fill this niche, and are deploying them to phones as well. (View Highlight)
  • Apple does not have a frontier model, Google and Microsoft/OpenAI have a large lead in that space. But they have created a bunch of small models that run on the AI-focused chips in Apple products. And they have built a medium-sized model that the iPhone can call in the cloud when it needs help. The model that runs on your phone is pretty close in abilities to the version of Mistral that I was using above (but much faster and optimized to reduce errors) and the version that runs in the cloud is better than the original ChatGPT, but not that much better. These smaller, weaker models give Apple a lot of control over AI use on their systems and offloads a lot of work to the phone or computer. But they still don’t have a frontier model, so they working with OpenAI to send GPT-4 the questions that are too hard for Apple’s models to answer. Companies are clearly still experimenting with which models, or sets of models, to offer. (View Highlight)
  • The makers of frontier models do not have strong views about how their systems can be used, and so they are not optimized for any one task. Working with advanced models is more like working with a human being, a smart one that makes mistakes and has weird moods sometimes. Frontier models are more likely to do extraordinary things but are also more frustrating and often unnerving to use. Contrast this with Apple’s narrow focus on making AI get stuff done for you. (View Highlight)
  • For many people the second use case is actually the more natural, intuitive, and useful one. A machine that can do anything much of the time, but also sometimes does something entirely different, is harder to understand than a narrow AI that just does what you want (with the caveat we don’t know how well the Apple system works). Google is also going to be releasing smaller AI models that are local to phones. And Microsoft is taking a similar approach to Apple, with a business twist. They have implemented Copilots in their key office apps. They do a really good job of providing easily understood “it just works” (mostly) integration of AI into work in easy ways. But both the Apple and app-specific Copilot models are constrained, which limits their upside, as well as their downside. (View Highlight)
  • The potential gains to AI, the productivity boosts and innovation, along with the weird risks, come from the larger, less constrained models. And the benefits come from figuring out how to apply AI to your own use cases, even though that takes work. Frontier models thus have a very different approach to use cases than more constrained models. Take a look at this demo, from OpenAI, where GPT-4o (rather flirtatiously?) helps someone work through an interview, and compare it to this demo of Apple’s AI-powered Siri, helping with appointments. Radically different philosophies at work. (View Highlight)
  • The best access to an advanced model costs you 2B](https://www.ft.com/content/81ac0e78-5b9b-43c2-b135-d11c47480119)). To a large extent, the future of AI will be shaped by the degree to which AI companies figure out sustainable business models, so expect to see more experimentation. (View Highlight)
  • What every one of these companies needs to succeed, however, is trust. There are a lot of reasons why people don’t trust AI companies; from their unclear use of training data to their plans for an AI-dominated future to their often-opaque management. But what most people mean by trust is the question of privacy (“will AI use what I give it as training data?”) and that has long been answered. All of the AI companies offer options where they agree to not use your data for training, and the legal implications for breaching these agreements would be dire. But Apple goes many steps further, putting extra work into making sure it could never learn about your data, even if it wanted to. Only the local AI on your phone accesses personal data, and anything handed to the cloud AI is encrypted, processed anonymously and instantly erased in ways that would be very hard for anyone to intercept. To the extent that data is given to OpenAI, it is also anonymous and requires explicit permission. Between the limited use cases and the privacy focus, this is a very “ethical” use of AI (though we still know little about Apple’s training data). We will see if that is enough to get the public to trust AI more. (View Highlight)
  • There is a specter haunting all AI development, the specter of AGI - Artificial General Intelligence, the hypothetical machine better than humans at every intellectual tasks. This is the explicit goal of OpenAI and Anthropic, and it is something they hope to achieve in the near term. For people who genuinely believe they are building AGI soon, almost nothing else is important. The AI models along the way to AGI are mere stepping stones, not anything you want to build a business around, because they will be replaced by better models soon. OpenAI’s systems may feel unpolished because the company believes that future models will significantly advance AI capabilities. As a result, they may not be investing heavily in refining systems that will likely be outdated as new models are released. I do not know if AGI is achievable, but I know that the mere idea of AGI being possible soon bends everything around it, resulting in wide differences in approach and philosophy in AI implementations. (View Highlight)
  • While Apple is building narrow AI systems that can accurately answer questions about your personal data (“tell me when my mother is landing”), OpenAI wants to build autonomous agents that would complete complex tasks for you (“You know those emails about the new business I want to start, could you figure out what I should do to register it so that it is best for my taxes and do that.”). The first is, as Apple demonstrated, science fact, while the second is science fiction, at least for now. Every major AI company argues the technology will evolve further and has teased mysterious future additions to their systems. In contrast, what we are seeing from Apple is a clear and practical vision of how AI can help most users, without a lot of effort, today. In doing so, they are hiding much of the power, and quirks, of LLMs from their users. Having companies take many approaches to AI is likely to lead to faster adoption in the long term. And, as companies experiment, we will learn more about which sets of models are correct. (View Highlight)