2026-03-21 06:03:32.0180 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71525/106718
2026-03-21 06:03:32.0180 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71526/106718
2026-03-21 06:03:32.0180 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71527/106718
2026-03-21 06:03:32.0180 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71528/106718
2026-03-21 06:03:32.0180 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71529/106718
2026-03-21 06:03:32.0180 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71530/106718
2026-03-21 06:03:32.0181 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71531/106718
2026-03-21 06:03:32.0181 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71532/106718
2026-03-21 06:03:32.0181 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71533/106718
2026-03-21 06:03:32.0181 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71534/106718
2026-03-21 06:03:32.0181 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71535/106718
2026-03-21 06:03:32.0181 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71536/106718
2026-03-21 06:03:32.0181 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71537/106718
2026-03-21 06:03:32.0181 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71538/106718
2026-03-21 06:03:32.0181 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71539/106718
2026-03-21 06:03:32.0181 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71540/106718
2026-03-21 06:03:32.0181 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71541/106718
2026-03-21 06:03:32.0182 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71542/106718
2026-03-21 06:03:32.0303 - ERROR - graphrag_llm.middleware.with_logging - Async request failed with exception=litellm.BadRequestError: OpenAIException - We could not parse the JSON body of your request. (HINT: This likely means you aren't using your HTTP library correctly. The OpenAI API expects a JSON payload, but what was sent was not valid JSON. If you have trouble figuring out how to fix this, please contact us through our help center at help.openai.com.)
Traceback (most recent call last):
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/llms/openai/openai.py", line 929, in acompletion
headers, response = await self.make_openai_chat_completion_request(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/litellm_core_utils/logging_utils.py", line 297, in async_wrapper
result = await func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/llms/openai/openai.py", line 460, in make_openai_chat_completion_request
raise e
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/llms/openai/openai.py", line 437, in make_openai_chat_completion_request
await openai_aclient.chat.completions.with_raw_response.create(
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/openai/_legacy_response.py", line 384, in wrapped
return cast(LegacyAPIResponse[R], await func(*args, **kwargs))
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/openai/resources/chat/completions/completions.py", line 2714, in create
return await self._post(
^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/openai/_base_client.py", line 1884, in post
return await self.request(cast_to, opts, stream=stream, stream_cls=stream_cls)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/openai/_base_client.py", line 1669, in request
raise self._make_status_error_from_response(err.response) from None
openai.BadRequestError: Error code: 400 - {'error': {'message': "We could not parse the JSON body of your request. (HINT: This likely means you aren't using your HTTP library correctly. The OpenAI API expects a JSON payload, but what was sent was not valid JSON. If you have trouble figuring out how to fix this, please contact us through our help center at help.openai.com.)", 'type': 'invalid_request_error', 'param': None, 'code': None}}
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/main.py", line 620, in acompletion
response = await init_response
^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/llms/openai/openai.py", line 989, in acompletion
raise OpenAIError(
litellm.llms.openai.common_utils.OpenAIError: Error code: 400 - {'error': {'message': "We could not parse the JSON body of your request. (HINT: This likely means you aren't using your HTTP library correctly. The OpenAI API expects a JSON payload, but what was sent was not valid JSON. If you have trouble figuring out how to fix this, please contact us through our help center at help.openai.com.)", 'type': 'invalid_request_error', 'param': None, 'code': None}}
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/middleware/with_logging.py", line 64, in _request_count_middleware_async
return await async_middleware(**kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/middleware/with_request_count.py", line 72, in _request_count_middleware_async
result = await async_middleware(**kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/middleware/with_cache.py", line 145, in _cache_middleware_async
response = await async_middleware(**kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/middleware/with_retries.py", line 55, in _retry_middleware_async
return await retrier.retry_async(
^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/retry/exponential_retry.py", line 102, in retry_async
return await func(**input_args)
^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/middleware/with_metrics.py", line 80, in _metrics_middleware_async
response = await async_middleware(**kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/completion/lite_llm_completion.py", line 290, in _base_completion_async
response = await litellm.acompletion(
^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/utils.py", line 2072, in wrapper_async
raise e
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/utils.py", line 1892, in wrapper_async
result = await original_function(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/main.py", line 639, in acompletion
raise exception_type(
^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/litellm_core_utils/exception_mapping_utils.py", line 2456, in exception_type
raise e
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/litellm_core_utils/exception_mapping_utils.py", line 478, in exception_type
raise BadRequestError(
litellm.exceptions.BadRequestError: litellm.BadRequestError: OpenAIException - We could not parse the JSON body of your request. (HINT: This likely means you aren't using your HTTP library correctly. The OpenAI API expects a JSON payload, but what was sent was not valid JSON. If you have trouble figuring out how to fix this, please contact us through our help center at help.openai.com.)
2026-03-21 06:03:32.0324 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71543/106718
2026-03-21 06:03:32.0325 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71544/106718
2026-03-21 06:03:32.0715 - ERROR - graphrag.index.run.run_pipeline - error running workflow extract_graph
Traceback (most recent call last):
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/llms/openai/openai.py", line 929, in acompletion
headers, response = await self.make_openai_chat_completion_request(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/litellm_core_utils/logging_utils.py", line 297, in async_wrapper
result = await func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/llms/openai/openai.py", line 460, in make_openai_chat_completion_request
raise e
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/llms/openai/openai.py", line 437, in make_openai_chat_completion_request
await openai_aclient.chat.completions.with_raw_response.create(
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/openai/_legacy_response.py", line 384, in wrapped
return cast(LegacyAPIResponse[R], await func(*args, **kwargs))
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/openai/resources/chat/completions/completions.py", line 2714, in create
return await self._post(
^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/openai/_base_client.py", line 1884, in post
return await self.request(cast_to, opts, stream=stream, stream_cls=stream_cls)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/openai/_base_client.py", line 1669, in request
raise self._make_status_error_from_response(err.response) from None
openai.BadRequestError: Error code: 400 - {'error': {'message': "We could not parse the JSON body of your request. (HINT: This likely means you aren't using your HTTP library correctly. The OpenAI API expects a JSON payload, but what was sent was not valid JSON. If you have trouble figuring out how to fix this, please contact us through our help center at help.openai.com.)", 'type': 'invalid_request_error', 'param': None, 'code': None}}
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/main.py", line 620, in acompletion
response = await init_response
^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/llms/openai/openai.py", line 989, in acompletion
raise OpenAIError(
litellm.llms.openai.common_utils.OpenAIError: Error code: 400 - {'error': {'message': "We could not parse the JSON body of your request. (HINT: This likely means you aren't using your HTTP library correctly. The OpenAI API expects a JSON payload, but what was sent was not valid JSON. If you have trouble figuring out how to fix this, please contact us through our help center at help.openai.com.)", 'type': 'invalid_request_error', 'param': None, 'code': None}}
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag/index/run/run_pipeline.py", line 135, in _run_pipeline
result = await workflow_function(config, context)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag/index/workflows/extract_graph.py", line 61, in run_workflow
entities, relationships, raw_entities, raw_relationships = await extract_graph(
^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag/index/workflows/extract_graph.py", line 143, in extract_graph
entities, relationships = await get_summarized_entities_relationships(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag/index/workflows/extract_graph.py", line 168, in get_summarized_entities_relationships
entity_summaries, relationship_summaries = await summarize_descriptions(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag/index/operations/summarize_descriptions/summarize_descriptions.py", line 115, in summarize_descriptions
return await get_summarized(entities_df, relationships_df, semaphore)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag/index/operations/summarize_descriptions/summarize_descriptions.py", line 80, in get_summarized
edge_results = await asyncio.gather(*edge_futures)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/.local/share/uv/python/cpython-3.12.13-linux-x86_64-gnu/lib/python3.12/asyncio/tasks.py", line 385, in __wakeup
future.result()
File "/home/dobin/.local/share/uv/python/cpython-3.12.13-linux-x86_64-gnu/lib/python3.12/asyncio/tasks.py", line 314, in __step_run_and_handle_result
result = coro.send(None)
^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag/index/operations/summarize_descriptions/summarize_descriptions.py", line 102, in do_summarize_descriptions
results = await run_summarize_descriptions(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag/index/operations/summarize_descriptions/summarize_descriptions.py", line 139, in run_summarize_descriptions
result = await extractor(id=id, descriptions=descriptions)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag/index/operations/summarize_descriptions/description_summary_extractor.py", line 68, in __call__
result = await self._summarize_descriptions(id, descriptions)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag/index/operations/summarize_descriptions/description_summary_extractor.py", line 105, in _summarize_descriptions
result = await self._summarize_descriptions_with_llm(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag/index/operations/summarize_descriptions/description_summary_extractor.py", line 124, in _summarize_descriptions_with_llm
response: LLMCompletionResponse = await self._model.completion_async(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/completion/lite_llm_completion.py", line 193, in completion_async
response = await self._completion_async(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/middleware/with_logging.py", line 64, in _request_count_middleware_async
return await async_middleware(**kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/middleware/with_request_count.py", line 72, in _request_count_middleware_async
result = await async_middleware(**kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/middleware/with_cache.py", line 145, in _cache_middleware_async
response = await async_middleware(**kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/middleware/with_retries.py", line 55, in _retry_middleware_async
return await retrier.retry_async(
^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/retry/exponential_retry.py", line 102, in retry_async
return await func(**input_args)
^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/middleware/with_metrics.py", line 80, in _metrics_middleware_async
response = await async_middleware(**kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/completion/lite_llm_completion.py", line 290, in _base_completion_async
response = await litellm.acompletion(
^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/utils.py", line 2072, in wrapper_async
raise e
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/utils.py", line 1892, in wrapper_async
result = await original_function(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/main.py", line 639, in acompletion
raise exception_type(
^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/litellm_core_utils/exception_mapping_utils.py", line 2456, in exception_type
raise e
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/litellm_core_utils/exception_mapping_utils.py", line 478, in exception_type
raise BadRequestError(
litellm.exceptions.BadRequestError: litellm.BadRequestError: OpenAIException - We could not parse the JSON body of your request. (HINT: This likely means you aren't using your HTTP library correctly. The OpenAI API expects a JSON payload, but what was sent was not valid JSON. If you have trouble figuring out how to fix this, please contact us through our help center at help.openai.com.)
2026-03-21 06:03:32.0737 - ERROR - graphrag.api.index - Workflow extract_graph completed with errors
2026-03-21 06:03:32.0740 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71545/106718
2026-03-21 06:03:32.0744 - INFO - graphrag_llm.metrics.log_metrics_writer - Metrics for openai/text-embedding-3-small: {
"attempted_request_count": 1,
"successful_response_count": 1,
"failed_response_count": 0,
"failure_rate": 0.0,
"requests_with_retries": 0,
"retries": 0,
"retry_rate": 0.0,
"compute_duration_seconds": 2.2788126468658447,
"compute_duration_per_response_seconds": 2.2788126468658447,
"cache_hit_rate": 0.0,
"streaming_responses": 0,
"responses_with_tokens": 1,
"prompt_tokens": 9,
"total_tokens": 9,
"tokens_per_response": 9.0,
"cost_per_response": 0.0
}
2026-03-21 06:03:32.0745 - INFO - graphrag_llm.metrics.log_metrics_writer - Metrics for openai/gpt-5.4-mini: {
"attempted_request_count": 21831,
"successful_response_count": 21829,
"failed_response_count": 2,
"failure_rate": 9.161284412074573e-05,
"requests_with_retries": 0,
"retries": 0,
"retry_rate": 0.0,
"compute_duration_seconds": 48129.54413795471,
"compute_duration_per_response_seconds": 2.204844204404907,
"cached_responses": 6,
"cache_hit_rate": 0.0002748637134087682,
"streaming_responses": 0,
"responses_with_tokens": 21829,
"prompt_tokens": 22409866,
"completion_tokens": 8838469,
"total_tokens": 31248335,
"tokens_per_response": 1431.5055659901966,
"cost_per_response": 0.0
}
Do you need to file an issue?
Describe the bug
When indexing, i get the following error message:
Steps to reproduce
Expected Behavior
No error
GraphRAG Config Used
I changed model, and file pattern:
Logs and screenshots
No response
Additional Information