The response was filtered due to the prompt triggering Azure OpenAI's content management policy. Please modify your prompt and retry. To learn more about our content filtering policies please read our documentation: https://***.com/***?linkid=***
17天前
错误信息: The selected model requires paid balance. Your paid balance is insufficient. Please top up and try again. 堆栈信息: AI_APICallError at file:///C:/Program%20Files/Cherry%20Studio/resources/app.asar/out/renderer/assets/dist-gixTQuVo.js:5146:11 at async postToApi (file:///C:/Program%20Files/Cherry%20Studio/resources/app.asar/out/renderer/assets/dist-gixTQuVo.js:5028:24) at async OpenAICompatibleChatLanguageModel$1.doStream (file:///C:/Program%20Files/Cherry%20Studio/resources/app.asar/out/renderer/assets/store-Dy89cVhB.js:243247:48) at async wrapStream (file:///C:/Program%20Files/Cherry%20Studio/resources/app.asar/out/renderer/assets/store-Dy89cVhB.js:38758:31) at async fn (file:///C:/Program%20Files/Cherry%20Studio/resources/app.asar/out/renderer/assets/store-Dy89cVhB.js:36943:17) at async file:///C:/Program%20Files/Cherry%20Studio/resources/app.asar/out/renderer/assets/store-Dy89cVhB.js:34321:19 at async _retryWithExponentialBackoff (file:///C:/Program%20Files/Cherry%20Studio/resources/app.asar/out/renderer/assets/store-Dy89cVhB.js:34423:10) at async streamStep (file:///C:/Program%20Files/Cherry%20Studio/resources/app.asar/out/renderer/assets/store-Dy89cVhB.js:36911:108) at async fn (file:///C:/Program%20Files/Cherry%20Studio/resources/app.asar/out/renderer/assets/store-Dy89cVhB.js:37190:5) at async file:///C:/Program%20Files/Cherry%20Studio/resources/app.asar/out/renderer/assets/store-Dy89cVhB.js:34321:19 错误原因: "[undefined]" 状态码: 403 请求路径: https://api.siliconflow.cn/v1/chat/completions 请求体: { "model": "Pro/moonshotai/Kimi-K2-Thinking", "user": "[undefined]", "max_tokens": "[undefined]", "temperature": 1, "top_p": "[undefined]", "frequency_penalty": "[undefined]", "presence_penalty": "[undefined]", "response_format": "[undefined]", "stop": "[undefined]", "seed": "[undefined]", "serviceTier": "[undefined]", "reasoning_effort": "[undefined]", "verbosity": "[undefined]", "messages": [ { "role": "user", "content": "nh" } ], "tools": "[undefined]", "tool_choice": "[undefined]", "stream": true, "stream_options": { "include_usage": true } } 响应首部: { "content-length": "144", "content-type": "application/json; charset=utf-8", "date": "Wed, 10 Dec 2025 06:15:10 GMT", "document-policy": "include-js-call-stacks-in-crash-reports", "x-siliconcloud-trace-id": "ti_yues7n5xbtjdgj4e6y" } 响应内容: { "code": 30011, "message": "The selected model requires paid balance. Your paid balance is insufficient. Please top up and try again.", "data": null } 数据: "[undefined]"
17天前
Traceback (most recent call last): File "D:\新建文件夹\langchain-base\ollama_python.py", line 3, in <module> response = chat( model="deepseek-r1:7b", ...<2 lines>... ] ) File "C:\Users\27717\AppData\Roaming\Python\Python313\site-packages\ollama\_client.py", line 365, in chat return self._request( ~~~~~~~~~~~~~^ ChatResponse, ^^^^^^^^^^^^^ ...<14 lines>... stream=stream, ^^^^^^^^^^^^^^ ) ^ File "C:\Users\27717\AppData\Roaming\Python\Python313\site-packages\ollama\_client.py", line 189, in _request return cls(**self._request_raw(*args, **kwargs).json()) ~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^ File "C:\Users\27717\AppData\Roaming\Python\Python313\site-packages\ollama\_client.py", line 129, in _request_raw r = self._client.request(*args, **kwargs) File "D:\software\Python\Lib\site-packages\httpx\_client.py", line 825, in request return self.send(request, auth=auth, follow_redirects=follow_redirects) ~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\software\Python\Lib\site-packages\httpx\_client.py", line 914, in send response = self._send_handling_auth( request, ...<2 lines>... history=[], ) File "D:\software\Python\Lib\site-packages\httpx\_client.py", line 942, in _send_handling_auth response = self._send_handling_redirects( request, follow_redirects=follow_redirects, history=history, ) File "D:\software\Python\Lib\site-packages\httpx\_client.py", line 979, in _send_handling_redirects response = self._send_single_request(request) File "D:\software\Python\Lib\site-packages\httpx\_client.py", line 1014, in _send_single_request response = transport.handle_request(request) File "D:\software\Python\Lib\site-packages\httpx\_transports\default.py", line 250, in handle_request resp = self._pool.handle_request(req) File "D:\software\Python\Lib\site-packages\httpcore\_sync\connection_pool.py", line 256, in handle_request raise exc from None File "D:\software\Python\Lib\site-packages\httpcore\_sync\connection_pool.py", line 236, in handle_request response = connection.handle_request( pool_request.request ) File "D:\software\Python\Lib\site-packages\httpcore\_sync\http_proxy.py", line 206, in handle_request return self._connection.handle_request(proxy_request) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^ File "D:\software\Python\Lib\site-packages\httpcore\_sync\connection.py", line 103, in handle_r return self._connection.handle_request(request) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^ st raise exc File "D:\software\Python\Lib\site-packages\httpcore\_sync\http11.py", line 106, in handle_request ) = self._receive_response_headers(**kwargs) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^ File "D:\software\Python\Lib\site-packages\httpcore\_sync\http11.py", line 177, in _receive_response_headers event = self._receive_event(timeout=timeout) File "D:\software\Python\Lib\site-packages\httpcore\_sync\http11.py", line 217, in _receive_event data = self._network_stream.read( self.READ_NUM_BYTES, timeout=timeout ) File "D:\software\Python\Lib\site-packages\httpcore\_backends\sync.py", line 128, in read return self._sock.recv(max_bytes) ~~~~~~~~~~~~~~~^^^^^^^^^^^
17天前
Traceback (most recent call last): File "D:\新建文件夹\langchain-base\ollama_python.py", line 3, in <module> response = chat( model="deepseek-r1:7b", ...<2 lines>... ] ) File "C:\Users\27717\AppData\Roaming\Python\Python313\site-packages\ollama\_client.py", line 365, in chat return self._request( ~~~~~~~~~~~~~^ ChatResponse, ^^^^^^^^^^^^^ ...<14 lines>... stream=stream, ^^^^^^^^^^^^^^ ) ^ File "C:\Users\27717\AppData\Roaming\Python\Python313\site-packages\ollama\_client.py", line 189, in _request return cls(**self._request_raw(*args, **kwargs).json()) ~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^ File "C:\Users\27717\AppData\Roaming\Python\Python313\site-packages\ollama\_client.py", line 129, in _request_raw r = self._client.request(*args, **kwargs) File "D:\software\Python\Lib\site-packages\httpx\_client.py", line 825, in request return self.send(request, auth=auth, follow_redirects=follow_redirects) ~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\software\Python\Lib\site-packages\httpx\_client.py", line 914, in send response = self._send_handling_auth( request, ...<2 lines>... history=[], ) File "D:\software\Python\Lib\site-packages\httpx\_client.py", line 942, in _send_handling_auth response = self._send_handling_redirects( request, follow_redirects=follow_redirects, history=history, ) File "D:\software\Python\Lib\site-packages\httpx\_client.py", line 979, in _send_handling_redirects response = self._send_single_request(request) File "D:\software\Python\Lib\site-packages\httpx\_client.py", line 1014, in _send_single_request response = transport.handle_request(request) File "D:\software\Python\Lib\site-packages\httpx\_transports\default.py", line 250, in handle_request resp = self._pool.handle_request(req) File "D:\software\Python\Lib\site-packages\httpcore\_sync\connection_pool.py", line 256, in handle_request raise exc from None File "D:\software\Python\Lib\site-packages\httpcore\_sync\connection_pool.py", line 236, in handle_request response = connection.handle_request( pool_request.request ) File "D:\software\Python\Lib\site-packages\httpcore\_sync\http_proxy.py", line 206, in handle_request return self._connection.handle_request(proxy_request) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^ File "D:\software\Python\Lib\site-packages\httpcore\_sync\connection.py", line 103, in handle_r return self._connection.handle_request(request) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^ st raise exc File "D:\software\Python\Lib\site-packages\httpcore\_sync\http11.py", line 106, in handle_request ) = self._receive_response_headers(**kwargs) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^ File "D:\software\Python\Lib\site-packages\httpcore\_sync\http11.py", line 177, in _receive_response_headers event = self._receive_event(timeout=timeout) File "D:\software\Python\Lib\site-packages\httpcore\_sync\http11.py", line 217, in _receive_event data = self._network_stream.read( self.READ_NUM_BYTES, timeout=timeout ) File "D:\software\Python\Lib\site-packages\httpcore\_backends\sync.py", line 128, in read return self._sock.recv(max_bytes) ~~~~~~~~~~~~~~~^^^^^^^^^^^ KeyboardInterrupt
17天前
NAME ID SIZE MODIFIED deepseek-r1:7b 755ced02ce7b 4.7 GB 41 hours ago
17天前
(e.response.text, e.response.status_code) from None ollama._types.ResponseError: model 'llama3.2' not found (status code: 404)
17天前
Unable to submit request because function call `default_api:builtin_web_search` in the 8. content block is missing a `thought_signature`.
17天前
":{"code":"","message":"Invalid Token (request id: 20251210111150998485416DOrkbllI)","type":"rix_api_error"}}
17天前
("Error code: 400 - {'error': {'message': 'Function call is missing a thought_signature in functionCall parts. This is required for tools to work correctly, and missing thought_signature may lead to degraded model performance. Additional data, function call default_api:web_search', position 2. Please refer to https://***.dev/**/***/*** for more details. (tid: 2025121002493569746236171884066)', 'type': 'rix_api_error', 'param'. ",'code': None}}")Traceback (most recent call last): File "/root/lht/bootstrapping-your-agent_batch_gaiavalid/.venv/lib/python3.12/site packages/langchain_core/language_models/chat_models.py", line 1164, in agenerate_with cache result = await self._agenerate AAAAAAAAAAAAAAAAAAAAAA File "/root/lht/bootstrapping-your-agent_batch_gaiavalid/.venv/lib/python3.12/site packages/langchain_openai/chat_models/base.py", line 1456, in _agenerate raise e
17天前
': {'code': '429', 'type': 'rate_limit', 'message': 'Rate limit exceeded'}}
17天前
] 请求任务错误 ... 429 RESOURCE_EXHAUSTED. {'error': {'code': 429, 'message': 'You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/usage?tab=rate-limit. \n* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash\n* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash\n* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash\nPlease retry in 44.854037496s.', 'status': 'RESOURCE_EXHAUSTED', 'details': [{'@type': 'type.googleapis.com/google.rpc.Help', 'links': [{'description': 'Learn more about Gemini API quotas', 'url': 'https://ai.google.dev/gemini-api/docs/rate-limits'}]}, {'@type': 'type.googleapis.com/google.rpc.QuotaFailure', 'violations': [{'quotaMetric': 'generativelanguage.googleapis.com/generate_content_free_tier_input_token_count', 'quotaId': 'GenerateContentInputTokensPerModelPerMinute-FreeTier', 'quotaDimensions': {'location': 'global', 'model': 'gemini-2.0-flash'}}, {'quotaMetric': 'generativelanguage.googleapis.com/generate_content_free_tier_requests', 'quotaId': 'GenerateRequestsPerMinutePerProjectPerModel-FreeTier', 'quotaDimensions': {'location': 'global', 'model': 'gemini-2.0-flash'}}, {'quotaMetric': 'generativelanguage.googleapis.com/generate_content_free_tier_requests', 'quotaId': 'GenerateRequestsPerDayPerProjectPerModel-FreeTier', 'quotaDimensions': {'location': 'global', 'model': 'gemini-2.0-flash'}}]}, {'@type': 'type.googleapis.com/google.rpc.RetryInfo', 'retryDelay': '44s'}]}}
17天前
] 请求任务错误 ... 429 RESOURCE_EXHAUSTED. {'error': {'code': 429, 'message': 'You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/usage?tab=rate-limit. \n* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 20, model: gemini-2.5-flash\nPlease retry in 14.347159481s.', 'status': 'RESOURCE_EXHAUSTED', 'details': [{'@type': 'type.googleapis.com/google.rpc.Help', 'links': [{'description': 'Learn more about Gemini API quotas', 'url': 'https://ai.google.dev/gemini-api/docs/rate-limits'}]}, {'@type': 'type.googleapis.com/google.rpc.QuotaFailure', 'violations': [{'quotaMetric': 'generativelanguage.googleapis.com/generate_content_free_tier_requests', 'quotaId': 'GenerateRequestsPerDayPerProjectPerModel-FreeTier', 'quotaDimensions': {'location': 'global', 'model': 'gemini-2.5-flash'}, 'quotaValue': '20'}]}, {'@type': 'type.googleapis.com/google.rpc.RetryInfo', 'retryDelay': '14s'}]}}
17天前
{ "id": "chatcmpl-twx6HzyDzTUyMr5M2Vtw8w", "choices": [ { "finish_reason": null, "index": 0, "logprobs": null, "message": { "content": "", "refusal": null, "role": "assistant", "annotations": null, "audio": null, "function_call": null, "tool_calls": null, "reasoning_content": "" } } ], "created": 1765301421, "model": "", "object": "chat.completion", "service_tier": null, "system_fingerprint": null, "usage": { "completion_tokens": 0, "prompt_tokens": 1000, "total_tokens": 1000, "completion_tokens_details": null, "prompt_tokens_details": null } }
17天前
":{"code":"AccountOverdueError","message":"The request failed because your account has an overdue balance. Request id: 021765308486393a58640f8f093e136cbe2236822f07c73fe696b","param":"","type":"Forbidden"}}
17天前
","message":"Error from provider(yinli,claude-sonnet-4-5-20250929-thinking: 400): {\"error\":{\"type\":\"\\u003cnil\\u003e\",\"message\":\"{\\\"type\\\":\\\"error\\\",\\\"error\\\":{\\\"type\\\":\\\"invalid_request_error\\\",\\\"message\\\":\\\"tools.0: Input tag 'function' found using 'type' does not match any of the expected tags: 'bash_20250124', 'custom', 'text_editor_20250124', 'text_editor_20250429', 'text_editor_20250728', 'web_search_20250305'\\\"},\\\"request_id\\\":\\\"req_011CVqng5MNBPYKrZjbjBgnj\\\"}(traceid: 15f2c659ccae7e185fdb0d23f2f5cfa7) (request id: 20251207005816550951659KrbF7fnf) (request id: 20251207005816470295483dkigtDmo) (request id: 20251207005816290163220r2t6g8Qy)\"},\"type\":\"error\"}","stack":"Error: Error from provider(yinli,claude-sonnet-4-5-20250929-thinking: 400): {\"error\":{\"type\":\"\\u003cnil\\u003e\",\"message\":\"{\\\"type\\\":\\\"error\\\",\\\"error\\\":{\\\"type\\\":\\\"invalid_request_error\\\",\\\"message\\\":\\\"tools.0: Input tag 'function' found using 'type' does not match any of the expected tags: 'bash_20250124', 'custom', 'text_editor_20250124', 'text_editor_20250429', 'text_editor_20250728', 'web_search_20250305'\\\"},\\\"request_id\\\":\\\"req_011CVqng5MNBPYKrZjbjBgnj\\\"}(traceid: 15f2c659ccae7e185fdb0d23f2f5cfa7) (request id: 20251207005816550951659KrbF7fnf) (request id: 20251207005816470295483dkigtDmo) (request id: 20251207005816290163220r2t6g8Qy)\"},\"type\":\"error\"}\n at nt (/Users/renfenghuang/.nvm/versions/node/v22.17.1/lib/node_modules/@musistudio/claude-code-router/dist/cli.js:79940:11)\n at p0 (/Users/renfenghuang/.nvm/versions/node/v22.17.1/lib/node_modules/@musistudio/claude-code-router/dist/cli.js:80000:11)\n at process.processTicksAndRejections (node:internal/process/task_queues:105:5)\n at async f0 (/Users/renfenghuang/.nvm/versions/node/v22.17.1/lib/node_modules/@musistudio/claude-code-router/dist/cli.js:79965:96)","statusCode":400,"code":"provider_response_error"},"msg":"Error from provider(yinli,claude-sonnet-4-5-20250929-thinking: 400): {\"error\":{\"type\":\"\\u003cnil\\u003e\",\"message\":\"{\\\"type\\\":\\\"error\\\",\\\"error\\\":{\\\"type\\\":\\\"invalid_request_error\\\",\\\"message\\\":\\\"tools.0: Input tag 'function' found using 'type' does not match any of the expected tags: 'bash_20250124', 'custom', 'text_editor_20250124', 'text_editor_20250429', 'text_editor_20250728', 'web_search_20250305'\\\"},\\\"request_id\\\":\\\"req_011CVqng5MNBPYKrZjbjBgnj\\\"}(traceid: 15f2c659ccae7e185fdb0d23f2f5cfa7) (request id: 20251207005816550951659KrbF7fnf) (request id: 20251207005816470295483dkigtDmo) (request id: 20251207005816290163220r2t6g8Qy)\"},\"type\":\"error\"}"}
17天前