User
Write something
Agent Zero Community Call is happening in 4 days
Connecting AgentZero to Signal
Hi guys, I'm just strating with A0 and I wanted to know if there's a way to connected it to the signal cli. Even if it's not a native way inside of agent zero
Venice image models using A0 API KEY
hello can anyone tell me how to connect A0 to Venice.ai image and video models using the agent zero api_key? Like he is doing here:
Memory issue for 0.98
litellm.exceptions.BadRequestError: litellm.BadRequestError: Lm_studioException - Error code: 400 - {'error': 'Context size has been exceeded.'} Traceback (most recent call last): Traceback (most recent call last): File "/opt/venv-a0/lib/python3.12/site-packages/litellm/llms/openai/openai.py", line 823, in acompletion headers, response = await self.make_openai_chat_completion_request( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/opt/venv-a0/lib/python3.12/site-packages/litellm/litellm_core_utils/logging_utils.py", line 190, in async_wrapper result = await func(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/opt/venv-a0/lib/python3.12/site-packages/litellm/llms/openai/openai.py", line 454, in make_openai_chat_completion_request raise e File "/opt/venv-a0/lib/python3.12/site-packages/litellm/llms/openai/openai.py", line 436, in make_openai_chat_completion_request await openai_aclient.chat.completions.with_raw_response.create( File "/opt/venv-a0/lib/python3.12/site-packages/openai/_legacy_response.py", line 381, in wrapped return cast(LegacyAPIResponse[R], await func(*args, **kwargs)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/opt/venv-a0/lib/python3.12/site-packages/openai/resources/chat/completions/completions.py", line 2589, in create return await self._post( ^^^^^^^^^^^^^^^^^ File "/opt/venv-a0/lib/python3.12/site-packages/openai/_base_client.py", line 1794, in post return await self.request(cast_to, opts, stream=stream, stream_cls=stream_cls) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/opt/venv-a0/lib/python3.12/site-packages/openai/_base_client.py", line 1594, in request raise self._make_status_error_from_response(err.response) from None openai.BadRequestError: Error code: 400 - {'error': 'Context size has been exceeded.'}
A0 Docekr Ports or LiteLLM Issues?
Hello and Good Afternoon Everyone, Please forgive my wall of text. I will state that I'm not 100% familiar with docker, as I've only used it on my desktop and not in the wild. Also, when discussing technicals on AI products I do believe that "context" is important for transparency. I have installed A0 on my Debian VPS. I followed the instructions found here -> https://www.agent-zero.ai/p/docs/get-started/#installation Docker components would not install as stated. Had to first add docker source and key, then I was able to install. Followed everything else; was able to pull image and start with persistance just fine. For added layer of security I have installed Tailscale - private vpn from my home -> A0 host (for free). I was able to get to the WebUI just fine by exposing port 50080 to tailscale (tailscale serve) which gives me an https URL only available on the tailscale network. Now for my nftables firewall - I am only allowing all connections through tailscale inteface; everything else gets denied by policy. I can ssh in, ping out, etc. Secure two-way communitcation established at the host/network level. After I installed A0 and able to login to the WebUI - I went to settings as instructed to set API keys. I found the area for Embedding Model. I set my model provider (openrouter), set my model (as suggested in documentation openai/text-embedding-3-small), set my API key, and set the embedding model base URL. Click "Save". I go to open "new chat' and type 'hello' -- starts "Initializing memory database...". It continues to do this for some time. I look in the logs (docker logs --tail 100 -f agent-zero) and I see this: WARNING [huggingface_hub.utils._http] Retrying in 8s [Retry 5/5]. '(MaxRetryError('HTTPSConnectionPool(host=\'huggingface.co\', port=443): Max retries exceeded with url: /sentence-transformers/all-MiniLM-L6-v2/resolve/main/tokenizer_config.json (Caused by NameResolutionError("HTTPSConnection(host=\'huggingface.co\', port=443): Failed to resolve \'huggingface.co\' ([Errno -3] Temporary failure in name resolution)"))'), '(Request ID: f16d5bf3-43e9-493a-803a-d30b16d0f987)')' thrown while requesting HEAD https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2/resolve/main/tokenizer_config.json
Agent Zero utility model crashes on history compression with Ollama — litellm 405 "method not allowed" on /api/generate
Setup: Agent Zero latest image (agent0ai/agent-zero:latest), litellm 1.79.3 (bundled in container), Chat model: Anthropic via OpenRouter (works fine), Utility model: switched to local Ollama (qwen3:32b on RTX 5090), Ollama runs on host, reachable from container via http://host.docker.internal:11434, The problem: Every time A0's conversation history gets long enough to trigger compression (_90_organize_history_wait.py → compress_attention → call_utility_model), it crashes with: litellm.APIConnectionError: OllamaException - 405 method not allowed URL: http://host.docker.internal:11434/api/generate The crash is fatal — kills the entire agent session. What I've verified: Ollama is healthy — curl -X POST http://host.docker.internal:11434/api/generate with a valid payload works perfectly from inside the A0 container, qwen3:32b is loaded and running on GPU (27GB VRAM, not CPU fallback), settings.json is correct:, "util_model_provider": "ollama", "util_model_name": "qwen3:32b", "util_model_api_base": "http://host.docker.internal:11434" Short tasks complete fine — the crash only happens when history is long enough to trigger the summarization/compression path, :thumbsup: Tried ollama as the provider(per the setup guide) -- same 405, Tried ollama_chat as provider — same 405, What I think is happening: litellm 1.79.3's Ollama provider is sending a malformed request to /api/generate — either wrong HTTP method, wrong content-type, or wrong payload format. Ollama's /api/generate endpoint returns 405 which means it's receiving a request it doesn't accept (likely a GET instead of POST, or missing required fields). The main chat model (Anthropic/OpenRouter) works because it never touches this code path. Only the utility model goes through the Ollama provider, and only during compression.
1-30 of 108
Agent Zero
skool.com/agent-zero
Agent Zero AI framework
Leaderboard (30-day)
Powered by