Where Copilots Work

A simple checklist for builders

"happy, clean, jeeves the butler smiling sparkling, holding a magic wand, digital art"

Sponsored By: Lever

Hire smarter with Lever—the only complete hiring solution that provides modern talent acquisition leaders with complete ATS and robust CRM capabilities in one product: LeverTRM

Luke Skywalker had R2-D2’s whistles and beeps. Maverick had Goose. Bertie had his butler Jeeves, who shimmered in and out of the room to perform tasks well before he’d even been asked to. 

These stories are popular because everyone wants a copilot—a partner who makes you better, and who (sometimes) becomes a friend you can lean on when things get hard.

This sort of thing is exactly what a lot of people in AI are building right now.

GitHub CoPilot is the first large-scale AI use case that has significant traction—reportedly writing 40% of the code for developers who use it. Reid Hoffman thinks there will be a copilot for every profession. Microsoft is building an AI copilot into Office. Diagram is building a copilot for designers. The list goes on.

These systems work like a superpowered autocomplete. They predict what you’re about to do, and then offer that to you before you have a chance to do it yourself. It saves time and effort. 

If you’re a builder hacking away on side projects, you’re probably thinking about building a copilot too. GPT-3 makes this kind of thing pretty easy to pull together over a weekend. I know because I’ve been doing it too: 

I built a little copilot for my mind. I want it to help make me smarter: to make connections between ideas, bring up pieces of supporting evidence for points I’m making, and suggest quotes to use as I’m writing. 

It takes in any chunk of text, and then attempts to complete the chunk using quotes it finds in my Readwise database.

It’s a cool demo, but it isn’t anywhere close to being a usable product yet. It doesn’t always pull great quotes for me, and it doesn’t always complete them in a way that actually supports the point I’m trying to make. It also doesn’t demonstrate sufficient understanding of my writing, or the writing of the authors it’s pulling from to be useful.

As I wrote in the End of Organizing, I’m quite optimistic about the future of technologies like this. I find myself reading fewer and fewer physical books and taking fewer physical notes. I’m increasingly confident that every digital highlight I make will be made 10x more useful by these tools in the next year or so. 

The question for me and other builders like me, is this: Where can these kinds of copilot experiences actually deliver value with the technology that’s available today? And what are the bottlenecks that need to be resolved to make these useful for more use cases? 

Let’s take these one at a time.

Lever is the leading Talent Acquisition Suite that makes it easy for talent teams to reach their hiring goals and to connect companies with top talent. 

With LeverTRM, talent leaders can scale and grow their people pipeline, build authentic long-lasting relationships, and source the right people to hire. And, thanks to LeverTRM's Analytics, you get customized reports with data visualization, offers completed, interview feedback, and much more—so you can make better, more informed, and strategic hiring decisions.

Where copilots can be built today

AI-based copilots are quite useful with out-of-the-box technology in situations where small pieces of lightly transformed boilerplate text provide a lot of value. 

This is particularly true in areas where:

  1. Text can be checked for accuracy quickly with little user effort
  2. The cost of inaccuracies is low
  3. Relevant text can be found reliably with embeddings search

GitHub CoPilot is a great example of this. But other examples are things like grant writing, contract writing, tax prep, many types of email replies, RFP responses, medical recommendations to doctors, and more.

If you’re building a copilot (or thinking of building one), I’ve put together a little checklist for you to go through in order to figure out whether it will be possible to get good results with today’s technology.

Can you build a copilot for it? A checklist.

If you want to build a copilot for a specific domain using today’s technology here’s the list of things you need to check off:

  1. Is there a corpus of relevant text completions to be used by this copilot?
  2. Can relevant text for completions be found reliably with embeddings search over this text corpus?
  3. Can those pieces of text, without more context needed, be lightly transformed and inserted as an accurate completion?
  4. Can completions be checked for accuracy with little to no user effort?

Is there a corpus of relevant text completions to be used by this copilot?

You want your copilot to be smart and not make things up. It should have access to some source of knowledge that it can bring to the user when they need it. Ideally, this source of knowledge is accurate, up to date, and maybe even personal to the user—for example, it might include all of their emails or their company’s internal wiki. 

If you have this, you’re ready to go to the next step.

Can relevant text for completions be found reliably with embeddings search over this text corpus?

Once you have a knowledge base for your copilot to use, you need the copilot to be able to accurately identify chunks of that knowledge base to return to the user when they need it. For example, in my copilot for thought demo, I needed my copilot to find quotes from my Readwise that were relevant to whatever I was currently writing.

The standard way to do this is to use embeddings search. Embeddings are a condensed mathematical representation of a piece of text. Just like latitude and longitude can help you tell how close two cities are on a map, embeddings do the same kind of thing for text chunks. If you want to know if two pieces of text are similar, calculate the embeddings for them and compare them. Text chunks with embeddings that are “closer” together are similar.

Embeddings are useful because when a user is typing something that the copilot wants to autocomplete, it can just look through its knowledge base to find pieces of text that are “close” to whatever the user is typing. 

But embeddings aren’t perfect, and it’s where a lot of copilot use cases fail for now. Your copilot quality is going to be bounded by your ability to find relevant chunks of information in your knowledge base to help the user. If you’re not getting relevant results, completion accuracy will suffer.

If you can get relevant results, then you can go to the next step.

Can those pieces of text, without more context needed, be lightly transformed and inserted as an accurate completion?

Once you can find the most relevant pieces of information in your knowledge base from embeddings search, your copilot is going to need to intelligently package them up as a completion for the user.

This works best if they only need to be slightly transformed before they can be suggested. For example, you’re often going to want to rearrange the text so that it carries the same information but is rephrased so that it completes the user's sentence. This kind of transformation is easy to do with GPT-3, but more advanced transformations are harder to do.

Can completions be checked for accuracy with little to no user effort?

Once your copilot suggests a completion, it works best if the user knows whether or not the completion is accurate without a lot of work. If the user has to spend a lot of time figuring out if the completion is accurate or not, they’ll just ignore it.

This is one of the big levers for copilots. If you can make it easy to check a completion without a lot of work, your copilot can return a lot of wrong answers because it doesn’t cost the user much to consider them. I think this is part of why GitHub Copilot is successful: You can just run the code to see if it’s right, so the computer generates the code and then checks it for you. 

Other use cases that require more user input will require correspondingly higher rates of accuracy for the user to feel motivated enough to check.

What might change this list?

The limits of a copilot are the limits of the AI’s context window. The context window is the amount of tokens you can feed into the AI in the prompt, and the amount of tokens it can give back in a completion. 

Because context windows are limited, you have to use embeddings search to find little pieces of information you can feed to your AI for it to generate a copilot completion. This means that while context windows are still small, the quality of your copilot is bounded by the quality of your embeddings search.

GPT-3’s current context window is 4,096 tokens, which is about 3,000 words. OpenAI is rumored to soon be releasing a version of its models that have a 32K token context window—roughly 8x the current size. This, I think, would be a giant step change in the quality of the responses that are returned for copilot use cases. 

You’d be able to return far more information for the AI to reason over and turn into a usable response, which would have a direct impact on accuracy. 

The other big limiters here are inference cost, inference speed, embedding cost, and access to usable data. I expect cost to go down and speed to go up significantly enough that I’m not worried about them as true bottlenecks. But access to usable data is a big deal.

Right now, I’m using Readwise as my data source. But my completions for my copilot would be a lot better if it had access to the books that I am pulling from. The average number of tokens in a book is on the order of 80,000 tokens. So in order to increase the quality of my responses, I need to figure out how to make that data available to the AI, and also clean it so that it’s easy for it to find relevant passages. 

Advice for builders

If you’re building or investing in this space, my recommendations for creating better copilot experiences are as follows:

Tighten your feedback loops

You can think of a copilot completion as a sequential chain:

  1. Get user input
  2. Query for relevant documents
  3. Prompt model with documents
  4. Return a result

As you’re developing a copilot experience, you’ll want to be able to iterate as quickly as possible on each of these parts of the chain, with as little code as possible. I recommend building tools to help you do this quickly.

As I was building my notes copilot, I built a little UI to visualize and quickly swamp out each part of the chain:

This worked well for me, but you should explore your own solutions.

Get creative with embeddings search

For now, the quality of your completions is limited by the quality of your embeddings search. Because of this, I’d recommend spending time focusing on increasing the quality of your embeddings search.

There are many ways to enhance embeddings search to help you get more relevant documents. For example, check out HyDE for a creative solution to this problem from the query side. Or, try using GPT-3 to summarize the data in your knowledge base to make it easier for embeddings to find a usable text chunk.

Decrease cost of checking for accuracy

The other big lever here to create a good experience with existing technology is to lower the cost to the user if the accuracy of completions is low. An easy one: Before you display anything to the user, use GPT-3 to check if it thinks the completion is any good. If not, then don’t display it.

But there are lots of other ways to do this. For example, make sure completions are quite short. Another example: Make sure that all of the context information the user would need to check for accuracy is included in the completion—so they don’t have to do research or think too hard.

Wrapping up

This is the ideal copilot in my mind:

Every time you touch your keyboard it brings to bear your entire archive of notes, and everything you’ve ever read, to help you complete your next sentence. 

It would help you make connections between ideas, bring up pieces of supporting evidence, and suggest quotes to use. It might also bring up writers you love who disagree with the point you’re making—so you could change your mind, or sharpen your argument in response to theirs.

Ideally, it would do this in a fashion that’s seamless, highly accurate, and easily checked. In other words, usually if it completes something it’s making a good point, and it’s easy for you to tell if the point is good or not, without lots of extra effort.

This is far from the reality today. If we want to advance these kinds of tools beyond just being interesting demos, we’re going to have to build them ourselves. 

I hope this post pushes a few of you in that direction. I’ll keep you posted as I keep discovering more.

Like this?
Become a subscriber.

Subscribe →

Or, learn more.

Thanks to our Sponsor: Lever

Hire smarter with Lever—the only complete hiring solution designed for modern talent acquisition teams.

Read this next:

Chain of Thought

The Knowledge Economy Is Over. Welcome to the Allocation Economy

In the age of AI, every maker becomes a manager

7 Jan 19, 2024 by Dan Shipper

Chain of Thought

How Hard Should I Push Myself?

What the science of stress tells us about peak performance

2 Oct 17, 2023 by Dan Shipper

Chain of Thought

GPT-4 Is a Reasoning Engine

Reason is only as good as the information we give it

4 Mar 31, 2023 by Dan Shipper

Chain of Thought

Transcript: ChatGPT for Radical Self-betterment

'How Do You Use ChatGPT?’ with Dr. Gena Gorlin

🔒 Jan 31, 2024 by Dan Shipper

Napkin Math

Crypto’s Prophet Speaks

A16z’s Chris Dixon hasn’t abandoned the faith with his new book, 'Read Write Own'

13 Feb 1, 2024 by Evan Armstrong

Thanks for rating this post—join the conversation by commenting below.

Comments

You need to login before you can comment.
Don't have an account? Sign up!

Every smart person you know is reading this newsletter

Get one actionable essay a day on AI, tech, and personal development

Subscribe

Already a subscriber? Login