I did some really simple experiments in a notebook with langchain and a few PDFs. It’s a neat technique.
The first thing that jumped out at me was that the retrieval step is an upstream bottleneck on the LLM, so any methods that you can use to get better retrieval performance are fair game. Embeddings and vector databases are hot right now, but there’s no reason that you can’t augment them with traditional search methods.
I did some really simple experiments in a notebook with langchain and a few PDFs. It’s a neat technique.
The first thing that jumped out at me was that the retrieval step is an upstream bottleneck on the LLM, so any methods that you can use to get better retrieval performance are fair game. Embeddings and vector databases are hot right now, but there’s no reason that you can’t augment them with traditional search methods.