On Thursday, AI company Anthropic announced it has given its ChatGPT-like Claude AI language model the ability to analyze an entire book’s worth of material in under a minute. This new ability comes from expanding Claude’s context window to 100,000 tokens, or about 75,000 words.
Like OpenAI’s GPT-4, Claude is a large language model (LLM) that works by predicting the next token in a sequence when given a certain input. Tokens are fragments of words used to simplify AI data processing, and a “context window” is similar to short-term memory—how much human-provided input data an LLM can process at once.
A larger context window means an LLM can consider larger works like books or participate in very long interactive conversations that span “hours or even days,” according to Anthropic:
The average person can read 100,000 tokens of text in ~5+ hours, and then they might need substantially longer to digest, remember, and analyze that information. Claude can now do this in less than a minute. For example, we loaded the entire text of The Great Gatsby into Claude-Instant (72K tokens) and modified one line to say Mr. Carraway was “a software engineer that works on machine learning tooling at Anthropic.” When we asked the model to spot what was different, it responded with the correct answer in 22 seconds.
While it may not sound impressive to pick out changes in a text (Microsoft Word can do that, but only if it has two documents to compare), consider that after feeding Claude the text of The Great Gatsby, the AI model can then interactively answer questions about it or analyze its meaning. 100,000 tokens is a big upgrade for LLMs. By comparison, OpenAI’s GPT-4 LLM boasts context window lengths of 4,096 tokens (about 3,000 words) when used as part of ChatGPT and 8,192 or 32,768 tokens via the GPT-4 API (which is currently only available via waitlist).
To understand how a larger context window leads to a longer conversation with a chatbot like ChatGPT or Claude, we made a diagram for an earlier article that shows how the size of the prompt (which is held in the context window) enlarges to contain the entire text of the conversation. That means a conversation can an last longer before the chatbot loses its “memory” of the conversation.
According to Anthropic, Claude’s enhanced capabilities extend past processing books. The enlarged context window could potentially help businesses extract important information from multiple documents through a conversational interaction. The company suggests that this approach may outperform vector search-based methods when dealing with complicated queries.
While not as big of a name in AI as Microsoft and Google, Anthropic has emerged as a notable rival to OpenAI in terms of competitive offerings in LLMs and API access. Former OpenAI VP of Research Dario Amodei and his sister Daniela founded Anthropic in 2021 after a disagreement over OpenAI’s commercial direction. Notably, Anthropic received a $300 million investment from Google in late 2022, with Google acquiring a 10 percent stake in the firm.
Anthropic says that 100K context windows are available now for users of the Claude API, which is currently restricted by a waitlist.
https://arstechnica.com/?p=1938873