Загрузка данных
{"user_message_id": 31, "reserved_assistant_message_id": 32}
{"error": "\u041f\u0440\u043e\u0438\u0437\u043e\u0448\u043b\u0430 \u043d\u0435\u043f\u0440\u0435\u0434\u0432\u0438\u0434\u0435\u043d\u043d\u0430\u044f \u043e\u0448\u0438\u0431\u043a\u0430. \u041f\u043e\u0436\u0430\u043b\u0443\u0439\u0441\u0442\u0430, \u043f\u043e\u043f\u0440\u043e\u0431\u0443\u0439\u0442\u0435 \u0441\u043d\u043e\u0432\u0430.", "stack_trace": "Traceback (most recent call last):\n File \"/usr/local/lib/python3.11/site-packages/litellm/llms/openai/openai.py\", line 745, in completion\n raise e\n File \"/usr/local/lib/python3.11/site-packages/litellm/llms/openai/openai.py\", line 628, in completion\n return self.streaming(\n ^^^^^^^^^^^^^^^\n File \"/usr/local/lib/python3.11/site-packages/litellm/llms/openai/openai.py\", line 918, in streaming\n headers, response = self.make_sync_openai_chat_completion_request(\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/usr/local/lib/python3.11/site-packages/litellm/litellm_core_utils/logging_utils.py\", line 237, in sync_wrapper\n result = func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/usr/local/lib/python3.11/site-packages/litellm/llms/openai/openai.py\", line 489, in make_sync_openai_chat_completion_request\n raise e\n File \"/usr/local/lib/python3.11/site-packages/litellm/llms/openai/openai.py\", line 471, in make_sync_openai_chat_completion_request\n raw_response = openai_client.chat.completions.with_raw_response.create(\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/usr/local/lib/python3.11/site-packages/openai/_legacy_response.py\", line 364, in wrapped\n return cast(LegacyAPIResponse[R], func(*args, **kwargs))\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/usr/local/lib/python3.11/site-packages/sentry_sdk/utils.py\", line 1761, in runner\n return sentry_patched_function(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/usr/local/lib/python3.11/site-packages/sentry_sdk/integrations/openai.py\", line 145, in new_chat_completion\n raise e from None\n File \"/usr/local/lib/python3.11/site-packages/sentry_sdk/integrations/openai.py\", line 141, in new_chat_completion\n res = f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/usr/local/lib/python3.11/site-packages/openai/_utils/_utils.py\", line 286, in wrapper\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/usr/local/lib/python3.11/site-packages/openai/resources/chat/completions/completions.py\", line 1156, in create\n return self._post(\n ^^^^^^^^^^^\n File \"/usr/local/lib/python3.11/site-packages/openai/_base_client.py\", line 1259, in post\n return cast(ResponseT, self.request(cast_to, opts, stream=stream, stream_cls=stream_cls))\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/usr/local/lib/python3.11/site-packages/openai/_base_client.py\", line 1047, in request\n raise self._make_status_error_from_response(err.response) from None\nopenai.InternalServerError: Error code: 500 - {'error': {'message': 'litellm.MidStreamFallbackError: litellm.APIConnectionError: [Errno 13] Permission denied\\nTraceback (most recent call last):\\n File \"/app/litellm/litellm_core_utils/streaming_handler.py\", line 2017, in __anext__\\n async for chunk in self.completion_stream: # type: ignore[union-attr]\\n ...<81 lines>...\\n return processed_chunk\\n File \"/app/gigachat_onprem_handler.py\", line 146, in __anext__\\n chunk = next(self.response_iterator)\\n File \"/app/gigachat_onprem_handler.py\", line 1120, in _stream_lines\\n http_client = client or httpx.Client(**httpx_kwargs)\\n ~~~~~~~~~~~~^^^^^^^^^^^^^^^^\\n File \"/usr/lib/python3.13/site-packages/httpx/_client.py\", line 688, in __init__\\n self._transport = self._init_transport(\\n ~~~~~~~~~~~~~~~~~~~~^\\n verify=verify,\\n ^^^^^^^^^^^^^^\\n ...<5 lines>...\\n transport=transport,\\n ^^^^^^^^^^^^^^^^^^^^\\n )\\n ^\\n File \"/usr/lib/python3.13/site-packages/httpx/_client.py\", line 731, in _init_transport\\n return HTTPTransport(\\n verify=verify,\\n ...<4 lines>...\\n limits=limits,\\n )\\n File \"/usr/lib/python3.13/site-packages/httpx/_transports/default.py\", line 153, in __init__\\n ssl_context = create_ssl_context(verify=verify, cert=cert, trust_env=trust_env)\\n File \"/usr/lib/python3.13/site-packages/httpx/_config.py\", line 67, in create_ssl_context\\n ctx.load_cert_chain(*cert)\\n ~~~~~~~~~~~~~~~~~~~^^^^^^^\\nPermissionError: [Errno 13] Permission denied\\n. Received Model Group=GigaChat-Max\\nAvailable Model Group Fallbacks=None', 'type': None, 'param': None, 'code': '500'}}\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File \"/usr/local/lib/python3.11/site-packages/litellm/main.py\", line 2153, in completion\n raise e\n File \"/usr/local/lib/python3.11/site-packages/litellm/main.py\", line 2125, in completion\n response = openai_chat_completions.completion(\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/usr/local/lib/python3.11/site-packages/litellm/llms/openai/openai.py\", line 756, in completion\n raise OpenAIError(\nlitellm.llms.openai.common_utils.OpenAIError: Error code: 500 - {'error': {'message': 'litellm.MidStreamFallbackError: litellm.APIConnectionError: [Errno 13] Permission denied\\nTraceback (most recent call last):\\n File \"/app/litellm/litellm_core_utils/streaming_handler.py\", line 2017, in __anext__\\n async for chunk in self.completion_stream: # type: ignore[union-attr]\\n ...<81 lines>...\\n return processed_chunk\\n File \"/app/gigachat_onprem_handler.py\", line 146, in __anext__\\n chunk = next(self.response_iterator)\\n File \"/app/gigachat_onprem_handler.py\", line 1120, in _stream_lines\\n http_client = client or httpx.Client(**httpx_kwargs)\\n ~~~~~~~~~~~~^^^^^^^^^^^^^^^^\\n File \"/usr/lib/python3.13/site-packages/httpx/_client.py\", line 688, in __init__\\n self._transport = self._init_transport(\\n ~~~~~~~~~~~~~~~~~~~~^\\n verify=verify,\\n ^^^^^^^^^^^^^^\\n ...<5 lines>...\\n transport=transport,\\n ^^^^^^^^^^^^^^^^^^^^\\n )\\n ^\\n File \"/usr/lib/python3.13/site-packages/httpx/_client.py\", line 731, in _init_transport\\n return HTTPTransport(\\n verify=verify,\\n ...<4 lines>...\\n limits=limits,\\n )\\n File \"/usr/lib/python3.13/site-packages/httpx/_transports/default.py\", line 153, in __init__\\n ssl_context = create_ssl_context(verify=verify, cert=cert, trust_env=trust_env)\\n File \"/usr/lib/python3.13/site-packages/httpx/_config.py\", line 67, in create_ssl_context\\n ctx.load_cert_chain(*cert)\\n ~~~~~~~~~~~~~~~~~~~^^^^^^^\\nPermissionError: [Errno 13] Permission denied\\n. Received Model Group=GigaChat-Max\\nAvailable Model Group Fallbacks=None', 'type': None, 'param': None, 'code': '500'}}\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File \"/app/gigab2b/chat/process_message.py\", line 1295, in stream_chat_message_objects\n yield from run_chat_llm_with_state_containers(\n File \"/app/gigab2b/chat/chat_state.py\", line 118, in run_chat_llm_with_state_containers\n raise pkt.obj.exception\n File \"/app/gigab2b/chat/chat_state.py\", line 87, in run_with_exception_capture\n func(emitter, *args, **kwargs_with_state)\n File \"/app/gigab2b/chat/llm_loop.py\", line 1014, in run_llm_loop\n packet = next(step_generator)\n ^^^^^^^^^^^^^^^^^^^^\n File \"/app/gigab2b/chat/llm_step.py\", line 387, in run_llm_step\n for packet in llm.stream(\n File \"/app/gigab2b/llm/chat_llm.py\", line 1074, in _stream_implementation\n self._completion(\n File \"/app/gigab2b/llm/chat_llm.py\", line 825, in _completion\n raise e\n File \"/app/gigab2b/llm/chat_llm.py\", line 693, in _completion\n response = litellm.completion(\n ^^^^^^^^^^^^^^^^^^^\n File \"/usr/local/lib/python3.11/site-packages/litellm/utils.py\", line 1371, in wrapper\n raise e\n File \"/usr/local/lib/python3.11/site-packages/litellm/utils.py\", line 1244, in wrapper\n result = original_function(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/usr/local/lib/python3.11/site-packages/litellm/main.py\", line 3767, in completion\n raise exception_type(\n ^^^^^^^^^^^^^^^\n File \"/usr/local/lib/python3.11/site-packages/litellm/litellm_core_utils/exception_mapping_utils.py\", line 2273, in exception_type\n raise e\n File \"/usr/local/lib/python3.11/site-packages/litellm/litellm_core_utils/exception_mapping_utils.py\", line 502, in exception_type\n raise InternalServerError(\nlitellm.exceptions.InternalServerError: litellm.InternalServerError: InternalServerError: OpenAIException - litellm.MidStreamFallbackError: litellm.APIConnectionError: [Errno 13] Permission denied\nTraceback (most recent call last):\n File \"/app/litellm/litellm_core_utils/streaming_handler.py\", line 2017, in __anext__\n async for chunk in self.completion_stream: # type: ignore[union-attr]\n ...<81 lines>...\n return processed_chunk\n File \"/app/gigachat_onprem_handler.py\", line 146, in __anext__\n chunk = next(self.response_iterator)\n File \"/app/gigachat_onprem_handler.py\", line 1120, in _stream_lines\n http_client = client or httpx.Client(**httpx_kwargs)\n ~~~~~~~~~~~~^^^^^^^^^^^^^^^^\n File \"/usr/lib/python3.13/site-packages/httpx/_client.py\", line 688, in __init__\n self._transport = self._init_transport(\n ~~~~~~~~~~~~~~~~~~~~^\n verify=verify,\n ^^^^^^^^^^^^^^\n ...<5 lines>...\n transport=transport,\n ^^^^^^^^^^^^^^^^^^^^\n )\n ^\n File \"/usr/lib/python3.13/site-packages/httpx/_client.py\", line 731, in _init_transport\n return HTTPTransport(\n verify=verify,\n ...<4 lines>...\n limits=limits,\n )\n File \"/usr/lib/python3.13/site-packages/httpx/_transports/default.py\", line 153, in __init__\n ssl_context = create_ssl_context(verify=verify, cert=cert, trust_env=trust_env)\n File \"/usr/lib/python3.13/site-packages/httpx/_config.py\", line 67, in create_ssl_context\n ctx.load_cert_chain(*cert)\n ~~~~~~~~~~~~~~~~~~~^^^^^^^\nPermissionError: [Errno 13] Permission denied\n. Received Model Group=GigaChat-Max\nAvailable Model Group Fallbacks=None\n"}