Discover more from hlb's Substack
Random Note on June 29, 2023
The current prevailing approach to equip Large Language Models (LLMs) with Retrieval Augmented Generation (RAG) capability involves the use of embeddings connected to a vector store. However, the approximation of these embeddings may not be directly relevant to the actual needs of the user.
For example, when a user asks, "How's the weather today?", a series of complex decisions are needed: the current date and time, the user's location (based on IP, Geo), the external tool to be used for the inquiry, etc. Or, if a user asks about a word, the context at that moment is necessary for making assumptions.
By using embeddings and a vector store, it can feel like we're disregarding the advancements made in search engines, which seems counterproductive.
However, it prompts the question, what would the suitable search engine or contextual judgment look like for use in LLMs? Could it be a Knowledge Graph?
目前要讓 LLM 具備 Retrieval Augmented Generation (RAG) 能力的主流做法，是透過 embedding 串接 vector store。但是 embedding 的近似，跟用戶的實際需求，可能是完全無關的兩件事情。
用 embedding & vector store，基本上是拋棄掉搜尋引擎累積的成果，感覺上是個彎路。
不過反過來問，適合給 LLM 使用的搜尋引擎、情境判斷，又會是什麼樣子呢？是 Knowledge Graph 嗎？我存疑。
Thanks for reading hlb's Substack! Subscribe for free to receive new posts and support my work.