Discussion about this post

User's avatar
Dario's avatar

I enjoyed reading your article and found it to be quite insightful. I've been experiencing similar issues with RAG. Currently, I'm trying to optimize the formatting of PDFs to create a vectorized document, which I believe will make it more efficient for the LLMS model to process the responses. However, I haven't yet found a reliable way to leverage the LLM model itself to generate the formatted file. Typically, the output truncates the document, and I'm dealing with PDFs that range from 25 to 35 pages.

Expand full comment
Work-Work Balance's avatar

For structured output from LLMs you should have used https://github.com/boundaryml/baml to make your life easier.

Also, why do you send the entire conversation back and forth to the LLMs, instead of the last things added?

Expand full comment
2 more comments...

No posts