Indexing stuck at 'Workflow extract_graph started.' stage. Fails with 18.18 percent complete
The indexing stage of the Quickstart notebook is stuck at the following stage:
{ 'status_code': 200, 'index_name': 'graphrag-index16', 'storage_name': 'filesforgraphrag16', 'status': 'running', 'percent_complete': 18.18, 'progress': 'Workflow extract_graph started.', }
after few mins
{ 'status_code': 200, 'index_name': 'graphrag-index16', 'storage_name': 'filesforgraphrag16', 'status': 'failed', 'percent_complete': 18.18, 'progress': 'Workflow extract_graph started.', }
I get the same error as you, same files etc worked with the old solution accelerator.
Did you get any errors with the CosmosDB during creation etc? My guess it's something with assigning access to CosmosDB that isn't working correctly during deployment.
Check the logs for your graphrag-index-manager kubernetes job and your indexing-job-
You may be able to find out what the error was.
sorry, I don't know how to. The deployment was successful so I don't think there was an issue. Is there a way to try old accelerator. I just want to see how to deploy /run it for learning perspective
For me it was the autogenerated prompts for extracting entities that had an error in it. The generated prompt had } where it shouldn't be, after removing the extra } it worked. I would start double checking your prompts to see if you can find anything, or try with default prompts.
are you referring to quickstart or advanced notebook?
are you referring to quickstart or advanced notebook?
The GraphRAG Accelerator API gives you the possibility to generate prompts, there is an example in the advanced notebook. Those generated prompts caused my problem, but best is of course to try and dig into the logs and see what you can find.
I am also experiencing this. we see the following error in the Storage acct logs/report folder. Upon checking we see that extract_entities/graph_extractor.py does not exist in the graphrag.
"stack": "Traceback (most recent call last):\n File "/usr/local/lib/python3.10/site-packages/graphrag/index/operations/extract_entities/graph_extractor.py", line 127, in call\n result = await self._process_document(text, prompt_variables)\n File "/usr/local/lib/python3.10/site-packages/graphrag/index/operations/extract_entities/graph_extractor.py", line 156, in _process_document\n self._extraction_prompt.format(**{\nValueError: unexpected '{' in field name\n",
That's the same issue I had, but with }, if you change the manual prompt (remove that unexpected '{' from the prompt or close it if needed) it should work.
I hit the exactly same issue. Graphrag event in AKS indicates nodes are not available. But I suppose AKS should manage auto scaling, right? Just use sample wiki data for indexing
0/2 nodes are available: 2 node(s) didn't match Pod's node affinity/selector. preemption: 0/2 nodes are available: 2 Preemption is not helpful for scheduling.
Hi all, I encountered a similar issue while indexing a .txt file from my local Windows system after successfully deploying the GraphRAG Accelerator in Azure. Although I passed the correct container name (book) in the 1-Quickstart.ipynb script, the file was not uploaded there—instead, a new container with a random name (e.g., 47d9c246a2294c7f0798323f6a7dfa05) was created.
The indexing workflow fails at step (3/11): extract_graph, after completing the first two steps (create_base_text_units, create_final_documents). OpenAI deployment seems correctly configured in AI Foundry with the expected name and endpoint.
Could someone please help me resolve this?
{ 'type': 'on_workflow_start', 'data': 'Index: cenik-graph-index-05 -- Workflow (3/11): extract_graph started.', 'details': { 'workflow_name': 'extract_graph', 'index_name': 'cenik-graph-index-05', }, } { 'type': 'error', 'data': 'Error Invoking LLM', 'cause': "Error code: 404 - {'error': {'code': '404', 'message': 'Resource not found'}}", 'stack': ( .......
' entry = await func()\n' ' File "/usr/local/lib/python3.10/site-packages/openai/resources/chat/completions/completions.py", line 200' '0, in create\n' ' return await self._post(\n' ' File "/usr/local/lib/python3.10/site-packages/openai/_base_client.py", line 1767, in post\n' ' return await self.request(cast_to, opts, stream=stream, stream_cls=stream_cls)\n' ' File "/usr/local/lib/python3.10/site-packages/openai/_base_client.py", line 1461, in request\n' ' return await self._request(\n' ' File "/usr/local/lib/python3.10/site-packages/openai/_base_client.py", line 1562, in _request\n' ' raise self._make_status_error_from_response(err.response) from None\n' "openai.NotFoundError: Error code: 404 - {'error': {'code': '404', 'message': 'Resource not found'}}\n" ),
I got the same error. The progress stops at 18.18%. I see
"openai.NotFoundError: Error code: 404 - {'error': {'code': '404', 'message': 'Resource not found'}}\n"
in the logs.