Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

It’t that precisely embeddings API from OpenAI? It has all context so very useful for search


Not quite. My understanding is that OpenAI's various embeddings APIs return only a single vector per document, instead of the sequence of hidden states corresponding to each predicted next token in the response generated by a GPT-type LLM.

Imagine getting generated text from a GPT LLM that comes with a deep embedding of each generated token's "contextual meaning":

  [(text_token, deep_emb), (text_token, deep_emb), ...]
allowing higher-level models and apps to use all the information in those rich representations as inputs.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: