← Back to OpenAI updates ← Terug naar OpenAI-updates
OpenAI ARTICLE ARTIKEL 25 January 2022 25 januari 2022

Introducing text and code embeddings Introducing text and code embeddings

Title: Introducing text and code embeddings Title: Introducing text and code embeddings

Article details Artikelgegevens
AI maker AI-maker OpenAI Type Type Article Artikel Published Gepubliceerd 25 January 2022 25 januari 2022 Updates Updates Videos Video's View original article Bekijk origineel artikel
Why it matters Waarom dit telt

Quick editorial signal Snelle redactionele duiding

5 min
Impact Impact

Worth checking before choosing or changing a subscription. Handig om te checken voordat je een abonnement kiest of wijzigt.

Audience Voor wie Developers Developers
Level Niveau Expert Expert
  • Track this as a OpenAI update, not just a standalone headline. Bekijk dit als OpenAI-update, niet alleen als losse headline.
  • Check plan details before changing subscriptions or advising a team. Controleer plandetails voordat je abonnementen wijzigt of een team adviseert.
  • Likely worth revisiting after people have used the release in practice. Waarschijnlijk de moeite waard om opnieuw te bekijken zodra mensen het in praktijk gebruiken.
model apps pricing developers

Introducing text and code embeddings | OpenAI

We are introducing embeddings, a new endpoint in the OpenAI API that makes it easy to perform natural language and code tasks like semantic search, clustering, topic modeling, and classification.

Listen to article

Embeddings are numerical representations of concepts converted to number sequences, which make it easy for computers to understand the relationships between those concepts. Our embeddings outperform top models in 3 standard benchmarks, including a 20% relative improvement in code search.

Embeddings are useful for working with natural language and code, because they can be readily consumed and compared by other machine learning models and algorithms like clustering or search.

Embeddings that are numerically similar are also semantically similar. For example, the embedding vector of “canine companions say” will be more similar to the embedding vector of “woof” than that of“meow.”

The new endpoint uses neural network models, which are descendants of GPT‑3, to map text and code to a vector representation—“embedding” them in a high-dimensional space. Each dimension captures some aspect of the input.

The new/embeddings⁠(opens in a new window)endpoint in theOpenAI API⁠(opens in a new window)provides text and code embeddings with a few lines of code:

import openai

response = openai.Embedding.create(

input="canine companions say",

engine="text-similarity-davinci-001")

We’re releasing three families of embedding models, each tuned to perform well on different functionalities: text similarity, text search, and code search. The models take either text or code as input and return an embedding vector.

Models Use Cases

Text similarity: Captures semantic similarity between pieces of text.text-similarity-{ada, babbage, curie, davinci}-001 Clustering, regression, anomaly detection, visualization

Text search: Semantic information retrieval over documents.text-search-{ada, babbage, curie, davinci}-{query, doc}-001 Search, context relevance, information retrieval

Code search: Find relevant code with a query in natural language.code-search-{ada, babbage}-{code, text}-001 Code search and relevance

Drag to pan, scroll or pinch to zoom

animal

athlete

film

transportation

village

Embeddings from the text-similarity-babbage-001 model, applied to the DBpedia dataset. We randomly selected 100 samples from the dataset covering 5 categories, and computed the embeddings via the /embeddings endpoint. The different categories show up as 5 clear clusters in the embedding space. To visualize the embedding space, we reduced the embedding dimensionality from 2048 to 3 using PCA. The code for how to visualize embedding space in 3D dimension is available here.

To compare the similarity of two pieces of text, you simply use thedot product⁠(opens in a new window)on the text embeddings. The result is a “similarity score”, sometimes called “cosine similarity⁠(opens in a new window),” between –1 and 1, where a higher number means more similarity. In most applications, the embeddings can be pre-computed, and then the dot product comparison is extremely fast to carry out.

One popular use of embeddings is to use them as features in machine learning tasks, such as classification. In machine learning literature, when using a linear classifier, this classification task is called a “linear probe.” Our text similarity models achieve new state-of-the-art results on linear probe classification inSentEval⁠(opens in a new window)(Conneau et al., 2018⁠(opens in a new window)), a commonly used benchmark for evaluating embedding quality.

Linear probe classification over 7 datasets

text-similarity-davinci-001

92.2%

Text search models provide embeddings that enable large-scale search tasks, like finding a relevant document among a collection of documents given a text query. Embedding for the documents and query are produced separately, and then cosine similarity is used to compare the similarity between the query and each document.

Embedding-based search can generalize better than word overlap techniques used in classical keyword search, because it captures the semantic meaning of text and is less sensitive to exact phrases or words. We evaluate the text search model’s performance on theBEIR⁠(opens in a new window)(Thakur, et al. 2021⁠(opens in a new window)) search evaluation suite and obtain better search performance than previous methods. Ourtext search guide⁠(opens in a new window)provides more details on using embeddings for search tasks.

Average accuracy over 11 search tasks in BEIR

text-search-davinci-{doc, query}-001

52.8%

Average accuracy over 6 programming languages

code-search-babbage-{doc, query}-001

93.5%

FineTune Learning⁠(opens in a new window)is a company building hybrid human-AI solutions for learning, likeadaptive learning loops⁠(opens in a new window)that help students reach academic standards.

OpenAI’s embeddings significantly improved the task of finding textbook content based on learning objectives. Achieving a top-5 accuracy of 89.1%, OpenAI’s text-search-curie embeddings model outperformed previous approaches like Sentence-BERT (64.5%). While human experts are still better, the FineTune team is now able to label entire textbooks in a matter of seconds, in contrast to the hours that it took the experts.

Fabius⁠(opens in a new window)helps companies turn customer conversations into structured insights that inform planning and prioritization. OpenAI’s embeddings allow companies to more easily find and tag customer call transcripts with feature requests.

For instance, customers might use words like “automated” or “easy to use” to ask for a better self-service platform. Previously, Fabius was using fuzzy keyword search to attempt to tag those transcripts with the self-service platform label. With OpenAI’s embeddings, they’re now able to find 2x more examples in general, and 6x–10x more examples for features with abstract use cases that don’t have a clear keyword customers might use.

All API customers can get started with theembeddings documentation⁠(opens in a new window)for using embeddings in their applications.

Help shape what we cover next Help bepalen wat we hierna volgen

Anonymous feedback, no frontend account needed. Anonieme feedback, zonder front-end account.

Watch related videos Bekijk gerelateerde video's

Open videos → Open video's →

More from OpenAI Meer van OpenAI

All updates Alle updates

Gemini komt eraan