Cocoindex update #656
-
I am new to Cocoindex and I am trying to build a knowledge graph by ingesting a single pdf file containing 30 pages. I have created a flow function in python file and I am trying to run cocoindex update main.py:PdfToKnowledgeGraph. I am running neo4j, postgres and Ollama with llama3.1:8b in a docker container which has CUDA enabled with specific memory allotments.(I am hoping that the below error is not a memory issue.) The command is stuck at: for hours. Is there anything wrong that I am doing? |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment
-
Thanks for describing the problem in detail! Sorry for the inconvenience. This is mostly likely caused by issue of ollama: ollama/ollama#8200 While 30 pages is long. It won't perform well for most models, and makes the ollama bug more vulnerable. You likely need to split it into chunks first. Our If it's still stuck after splitting it, you may try to switch to a different LLM API. OpenAI and Google Gemini are usually quite stable, and LiteLLM provides a proxy that integrate with a variety of LLM APIs. I also created #658 to make such issues easier to debug in the future, and created #659 to provide vLLM integration - it's potentially a more stable alternative to Ollama. |
Beta Was this translation helpful? Give feedback.
Thanks for describing the problem in detail! Sorry for the inconvenience.
This is mostly likely caused by issue of ollama: ollama/ollama#8200
While 30 pages is long. It won't perform well for most models, and makes the ollama bug more vulnerable. You likely need to split it into chunks first. Our
docs_to_knowledge_graph
example had an earlier version that split it into chunks first (we simplified the example so there's no step of splitting, but for largedocs it should actually be split first).If it's still stuck after splitting it, you may try to switch to a different LLM API. OpenAI and Google Gemini are usually quite stable, and LiteLLM provides a proxy that integrate with a variety of…