Загрузка данных


{"user_message_id": 31, "reserved_assistant_message_id": 32}
{"error": "\u041f\u0440\u043e\u0438\u0437\u043e\u0448\u043b\u0430 \u043d\u0435\u043f\u0440\u0435\u0434\u0432\u0438\u0434\u0435\u043d\u043d\u0430\u044f \u043e\u0448\u0438\u0431\u043a\u0430. \u041f\u043e\u0436\u0430\u043b\u0443\u0439\u0441\u0442\u0430, \u043f\u043e\u043f\u0440\u043e\u0431\u0443\u0439\u0442\u0435 \u0441\u043d\u043e\u0432\u0430.", "stack_trace": "Traceback (most recent call last):\n  File \"/usr/local/lib/python3.11/site-packages/litellm/llms/openai/openai.py\", line 745, in completion\n    raise e\n  File \"/usr/local/lib/python3.11/site-packages/litellm/llms/openai/openai.py\", line 628, in completion\n    return self.streaming(\n           ^^^^^^^^^^^^^^^\n  File \"/usr/local/lib/python3.11/site-packages/litellm/llms/openai/openai.py\", line 918, in streaming\n    headers, response = self.make_sync_openai_chat_completion_request(\n                        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n  File \"/usr/local/lib/python3.11/site-packages/litellm/litellm_core_utils/logging_utils.py\", line 237, in sync_wrapper\n    result = func(*args, **kwargs)\n             ^^^^^^^^^^^^^^^^^^^^^\n  File \"/usr/local/lib/python3.11/site-packages/litellm/llms/openai/openai.py\", line 489, in make_sync_openai_chat_completion_request\n    raise e\n  File \"/usr/local/lib/python3.11/site-packages/litellm/llms/openai/openai.py\", line 471, in make_sync_openai_chat_completion_request\n    raw_response = openai_client.chat.completions.with_raw_response.create(\n                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n  File \"/usr/local/lib/python3.11/site-packages/openai/_legacy_response.py\", line 364, in wrapped\n    return cast(LegacyAPIResponse[R], func(*args, **kwargs))\n                                      ^^^^^^^^^^^^^^^^^^^^^\n  File \"/usr/local/lib/python3.11/site-packages/sentry_sdk/utils.py\", line 1761, in runner\n    return sentry_patched_function(*args, **kwargs)\n           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n  File \"/usr/local/lib/python3.11/site-packages/sentry_sdk/integrations/openai.py\", line 145, in new_chat_completion\n    raise e from None\n  File \"/usr/local/lib/python3.11/site-packages/sentry_sdk/integrations/openai.py\", line 141, in new_chat_completion\n    res = f(*args, **kwargs)\n          ^^^^^^^^^^^^^^^^^^\n  File \"/usr/local/lib/python3.11/site-packages/openai/_utils/_utils.py\", line 286, in wrapper\n    return func(*args, **kwargs)\n           ^^^^^^^^^^^^^^^^^^^^^\n  File \"/usr/local/lib/python3.11/site-packages/openai/resources/chat/completions/completions.py\", line 1156, in create\n    return self._post(\n           ^^^^^^^^^^^\n  File \"/usr/local/lib/python3.11/site-packages/openai/_base_client.py\", line 1259, in post\n    return cast(ResponseT, self.request(cast_to, opts, stream=stream, stream_cls=stream_cls))\n                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n  File \"/usr/local/lib/python3.11/site-packages/openai/_base_client.py\", line 1047, in request\n    raise self._make_status_error_from_response(err.response) from None\nopenai.InternalServerError: Error code: 500 - {'error': {'message': 'litellm.MidStreamFallbackError: litellm.APIConnectionError: [Errno 13] Permission denied\\nTraceback (most recent call last):\\n  File \"/app/litellm/litellm_core_utils/streaming_handler.py\", line 2017, in __anext__\\n    async for chunk in self.completion_stream:  # type: ignore[union-attr]\\n    ...<81 lines>...\\n        return processed_chunk\\n  File \"/app/gigachat_onprem_handler.py\", line 146, in __anext__\\n    chunk = next(self.response_iterator)\\n  File \"/app/gigachat_onprem_handler.py\", line 1120, in _stream_lines\\n    http_client = client or httpx.Client(**httpx_kwargs)\\n                            ~~~~~~~~~~~~^^^^^^^^^^^^^^^^\\n  File \"/usr/lib/python3.13/site-packages/httpx/_client.py\", line 688, in __init__\\n    self._transport = self._init_transport(\\n                      ~~~~~~~~~~~~~~~~~~~~^\\n        verify=verify,\\n        ^^^^^^^^^^^^^^\\n    ...<5 lines>...\\n        transport=transport,\\n        ^^^^^^^^^^^^^^^^^^^^\\n    )\\n    ^\\n  File \"/usr/lib/python3.13/site-packages/httpx/_client.py\", line 731, in _init_transport\\n    return HTTPTransport(\\n        verify=verify,\\n    ...<4 lines>...\\n        limits=limits,\\n    )\\n  File \"/usr/lib/python3.13/site-packages/httpx/_transports/default.py\", line 153, in __init__\\n    ssl_context = create_ssl_context(verify=verify, cert=cert, trust_env=trust_env)\\n  File \"/usr/lib/python3.13/site-packages/httpx/_config.py\", line 67, in create_ssl_context\\n    ctx.load_cert_chain(*cert)\\n    ~~~~~~~~~~~~~~~~~~~^^^^^^^\\nPermissionError: [Errno 13] Permission denied\\n. Received Model Group=GigaChat-Max\\nAvailable Model Group Fallbacks=None', 'type': None, 'param': None, 'code': '500'}}\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n  File \"/usr/local/lib/python3.11/site-packages/litellm/main.py\", line 2153, in completion\n    raise e\n  File \"/usr/local/lib/python3.11/site-packages/litellm/main.py\", line 2125, in completion\n    response = openai_chat_completions.completion(\n               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n  File \"/usr/local/lib/python3.11/site-packages/litellm/llms/openai/openai.py\", line 756, in completion\n    raise OpenAIError(\nlitellm.llms.openai.common_utils.OpenAIError: Error code: 500 - {'error': {'message': 'litellm.MidStreamFallbackError: litellm.APIConnectionError: [Errno 13] Permission denied\\nTraceback (most recent call last):\\n  File \"/app/litellm/litellm_core_utils/streaming_handler.py\", line 2017, in __anext__\\n    async for chunk in self.completion_stream:  # type: ignore[union-attr]\\n    ...<81 lines>...\\n        return processed_chunk\\n  File \"/app/gigachat_onprem_handler.py\", line 146, in __anext__\\n    chunk = next(self.response_iterator)\\n  File \"/app/gigachat_onprem_handler.py\", line 1120, in _stream_lines\\n    http_client = client or httpx.Client(**httpx_kwargs)\\n                            ~~~~~~~~~~~~^^^^^^^^^^^^^^^^\\n  File \"/usr/lib/python3.13/site-packages/httpx/_client.py\", line 688, in __init__\\n    self._transport = self._init_transport(\\n                      ~~~~~~~~~~~~~~~~~~~~^\\n        verify=verify,\\n        ^^^^^^^^^^^^^^\\n    ...<5 lines>...\\n        transport=transport,\\n        ^^^^^^^^^^^^^^^^^^^^\\n    )\\n    ^\\n  File \"/usr/lib/python3.13/site-packages/httpx/_client.py\", line 731, in _init_transport\\n    return HTTPTransport(\\n        verify=verify,\\n    ...<4 lines>...\\n        limits=limits,\\n    )\\n  File \"/usr/lib/python3.13/site-packages/httpx/_transports/default.py\", line 153, in __init__\\n    ssl_context = create_ssl_context(verify=verify, cert=cert, trust_env=trust_env)\\n  File \"/usr/lib/python3.13/site-packages/httpx/_config.py\", line 67, in create_ssl_context\\n    ctx.load_cert_chain(*cert)\\n    ~~~~~~~~~~~~~~~~~~~^^^^^^^\\nPermissionError: [Errno 13] Permission denied\\n. Received Model Group=GigaChat-Max\\nAvailable Model Group Fallbacks=None', 'type': None, 'param': None, 'code': '500'}}\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n  File \"/app/gigab2b/chat/process_message.py\", line 1295, in stream_chat_message_objects\n    yield from run_chat_llm_with_state_containers(\n  File \"/app/gigab2b/chat/chat_state.py\", line 118, in run_chat_llm_with_state_containers\n    raise pkt.obj.exception\n  File \"/app/gigab2b/chat/chat_state.py\", line 87, in run_with_exception_capture\n    func(emitter, *args, **kwargs_with_state)\n  File \"/app/gigab2b/chat/llm_loop.py\", line 1014, in run_llm_loop\n    packet = next(step_generator)\n             ^^^^^^^^^^^^^^^^^^^^\n  File \"/app/gigab2b/chat/llm_step.py\", line 387, in run_llm_step\n    for packet in llm.stream(\n  File \"/app/gigab2b/llm/chat_llm.py\", line 1074, in _stream_implementation\n    self._completion(\n  File \"/app/gigab2b/llm/chat_llm.py\", line 825, in _completion\n    raise e\n  File \"/app/gigab2b/llm/chat_llm.py\", line 693, in _completion\n    response = litellm.completion(\n               ^^^^^^^^^^^^^^^^^^^\n  File \"/usr/local/lib/python3.11/site-packages/litellm/utils.py\", line 1371, in wrapper\n    raise e\n  File \"/usr/local/lib/python3.11/site-packages/litellm/utils.py\", line 1244, in wrapper\n    result = original_function(*args, **kwargs)\n             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n  File \"/usr/local/lib/python3.11/site-packages/litellm/main.py\", line 3767, in completion\n    raise exception_type(\n          ^^^^^^^^^^^^^^^\n  File \"/usr/local/lib/python3.11/site-packages/litellm/litellm_core_utils/exception_mapping_utils.py\", line 2273, in exception_type\n    raise e\n  File \"/usr/local/lib/python3.11/site-packages/litellm/litellm_core_utils/exception_mapping_utils.py\", line 502, in exception_type\n    raise InternalServerError(\nlitellm.exceptions.InternalServerError: litellm.InternalServerError: InternalServerError: OpenAIException - litellm.MidStreamFallbackError: litellm.APIConnectionError: [Errno 13] Permission denied\nTraceback (most recent call last):\n  File \"/app/litellm/litellm_core_utils/streaming_handler.py\", line 2017, in __anext__\n    async for chunk in self.completion_stream:  # type: ignore[union-attr]\n    ...<81 lines>...\n        return processed_chunk\n  File \"/app/gigachat_onprem_handler.py\", line 146, in __anext__\n    chunk = next(self.response_iterator)\n  File \"/app/gigachat_onprem_handler.py\", line 1120, in _stream_lines\n    http_client = client or httpx.Client(**httpx_kwargs)\n                            ~~~~~~~~~~~~^^^^^^^^^^^^^^^^\n  File \"/usr/lib/python3.13/site-packages/httpx/_client.py\", line 688, in __init__\n    self._transport = self._init_transport(\n                      ~~~~~~~~~~~~~~~~~~~~^\n        verify=verify,\n        ^^^^^^^^^^^^^^\n    ...<5 lines>...\n        transport=transport,\n        ^^^^^^^^^^^^^^^^^^^^\n    )\n    ^\n  File \"/usr/lib/python3.13/site-packages/httpx/_client.py\", line 731, in _init_transport\n    return HTTPTransport(\n        verify=verify,\n    ...<4 lines>...\n        limits=limits,\n    )\n  File \"/usr/lib/python3.13/site-packages/httpx/_transports/default.py\", line 153, in __init__\n    ssl_context = create_ssl_context(verify=verify, cert=cert, trust_env=trust_env)\n  File \"/usr/lib/python3.13/site-packages/httpx/_config.py\", line 67, in create_ssl_context\n    ctx.load_cert_chain(*cert)\n    ~~~~~~~~~~~~~~~~~~~^^^^^^^\nPermissionError: [Errno 13] Permission denied\n. Received Model Group=GigaChat-Max\nAvailable Model Group Fallbacks=None\n"}