Debugging Unexpected LLM Invocations in Dify's Knowledge Retrieval Node
Context A Dify instance was deployed locally on a GPU server, integrated with Xinference hosting the THUDM/glm-4-9b-chat model. A RAG workflow was built using Dify’s default task flow template and a local knowledge base. During chat execution, the knowledge retrieval node failed with an error: Model...