mirror of
https://github.com/NousResearch/hermes-agent.git
synced 2026-04-28 06:51:16 +08:00
WSL2 / Clash-style setups often export ALL_PROXY=socks://127.0.0.1:PORT. httpx and the Anthropic SDK reject that alias and expect socks5://, so agent startup failed early with "Unknown scheme for proxy URL" before any provider request could proceed. Add shared normalize_proxy_url()/normalize_proxy_env_vars() helpers in utils.py and route all proxy entry points through them: - run_agent._get_proxy_from_env - agent.auxiliary_client._validate_proxy_env_urls - agent.anthropic_adapter.build_anthropic_client - gateway.platforms.base.resolve_proxy_url Regression coverage: - run_agent proxy env resolution - auxiliary proxy env normalization - gateway proxy URL resolution Verified with: PYTEST_DISABLE_PLUGIN_AUTOLOAD=1 /home/nonlinear/.hermes/hermes-agent/venv/bin/pytest -o addopts='' -p pytest_asyncio.plugin tests/run_agent/test_create_openai_client_proxy_env.py tests/agent/test_proxy_and_url_validation.py tests/gateway/test_proxy_mode.py 39 passed.
146 lines
5.6 KiB
Python
146 lines
5.6 KiB
Python
"""Regression guard: _create_openai_client must honor HTTP(S)_PROXY env vars.
|
|
|
|
When #11277 re-landed TCP keepalives, ``_create_openai_client`` began passing
|
|
a custom ``transport=httpx.HTTPTransport(...)`` to ``httpx.Client``. httpx only
|
|
auto-reads ``HTTP_PROXY`` / ``HTTPS_PROXY`` / ``ALL_PROXY`` when
|
|
``transport is None`` (see ``Client.__init__``:
|
|
``allow_env_proxies = trust_env and transport is None``). As a result, proxy
|
|
env vars were silently ignored for the primary chat client, causing requests
|
|
to bypass local proxies (Clash, corporate egress, etc.) and hit upstream
|
|
directly from the raw interface.
|
|
|
|
For users on WSL2 + Clash TUN this surfaced as Cloudflare ``cf-mitigated:
|
|
challenge`` 403s against ``chatgpt.com/backend-api/codex`` once they upgraded
|
|
past #11277. The fix forwards the proxy URL explicitly to ``httpx.Client``
|
|
while keeping the keepalive-enabled transport in place.
|
|
|
|
This test pins that the constructed ``httpx.Client`` mounts an ``HTTPProxy``
|
|
pool when a proxy env var is set, AND that the socket-level keepalive
|
|
transport is still installed on the no-proxy default path.
|
|
"""
|
|
from unittest.mock import patch
|
|
|
|
import httpx
|
|
|
|
from run_agent import AIAgent, _get_proxy_from_env
|
|
|
|
|
|
def _make_agent():
|
|
return AIAgent(
|
|
api_key="test-key",
|
|
base_url="https://chatgpt.com/backend-api/codex",
|
|
provider="openai-codex",
|
|
model="gpt-5.4",
|
|
quiet_mode=True,
|
|
skip_context_files=True,
|
|
skip_memory=True,
|
|
)
|
|
|
|
|
|
def _extract_http_client(client_kwargs: dict):
|
|
"""_create_openai_client calls ``OpenAI(**client_kwargs)``; grab the injected client."""
|
|
return client_kwargs.get("http_client")
|
|
|
|
|
|
def test_get_proxy_from_env_prefers_https_then_http_then_all(monkeypatch):
|
|
for key in ("HTTPS_PROXY", "HTTP_PROXY", "ALL_PROXY",
|
|
"https_proxy", "http_proxy", "all_proxy"):
|
|
monkeypatch.delenv(key, raising=False)
|
|
assert _get_proxy_from_env() is None
|
|
|
|
monkeypatch.setenv("ALL_PROXY", "http://all:1")
|
|
assert _get_proxy_from_env() == "http://all:1"
|
|
|
|
monkeypatch.setenv("HTTP_PROXY", "http://http:2")
|
|
assert _get_proxy_from_env() == "http://http:2"
|
|
|
|
monkeypatch.setenv("HTTPS_PROXY", "http://https:3")
|
|
assert _get_proxy_from_env() == "http://https:3"
|
|
|
|
|
|
def test_get_proxy_from_env_ignores_blank_values(monkeypatch):
|
|
for key in ("HTTPS_PROXY", "HTTP_PROXY", "ALL_PROXY",
|
|
"https_proxy", "http_proxy", "all_proxy"):
|
|
monkeypatch.delenv(key, raising=False)
|
|
monkeypatch.setenv("HTTPS_PROXY", " ")
|
|
monkeypatch.setenv("HTTP_PROXY", "http://real-proxy:8080")
|
|
assert _get_proxy_from_env() == "http://real-proxy:8080"
|
|
|
|
|
|
def test_get_proxy_from_env_normalizes_socks_alias(monkeypatch):
|
|
for key in ("HTTPS_PROXY", "HTTP_PROXY", "ALL_PROXY",
|
|
"https_proxy", "http_proxy", "all_proxy"):
|
|
monkeypatch.delenv(key, raising=False)
|
|
monkeypatch.setenv("ALL_PROXY", "socks://127.0.0.1:1080/")
|
|
assert _get_proxy_from_env() == "socks5://127.0.0.1:1080/"
|
|
|
|
|
|
@patch("run_agent.OpenAI")
|
|
def test_create_openai_client_routes_via_proxy_when_env_set(mock_openai, monkeypatch):
|
|
"""With HTTPS_PROXY set, the custom httpx.Client must mount an HTTPProxy pool.
|
|
|
|
This is the WSL2 + Clash / corporate-egress case. Before the fix, the custom
|
|
transport suppressed httpx's env-proxy auto-detection, so requests bypassed
|
|
the proxy entirely.
|
|
"""
|
|
for key in ("HTTPS_PROXY", "HTTP_PROXY", "ALL_PROXY",
|
|
"https_proxy", "http_proxy", "all_proxy"):
|
|
monkeypatch.delenv(key, raising=False)
|
|
monkeypatch.setenv("HTTPS_PROXY", "http://127.0.0.1:7897")
|
|
|
|
agent = _make_agent()
|
|
kwargs = {
|
|
"api_key": "test-key",
|
|
"base_url": "https://chatgpt.com/backend-api/codex",
|
|
}
|
|
agent._create_openai_client(kwargs, reason="test", shared=False)
|
|
|
|
forwarded = mock_openai.call_args.kwargs
|
|
http_client = _extract_http_client(forwarded)
|
|
assert isinstance(http_client, httpx.Client), (
|
|
"Expected _create_openai_client to inject a keepalive-enabled "
|
|
"httpx.Client; got %r" % (http_client,)
|
|
)
|
|
# Verify a proxy mount exists. httpx Client(proxy=...) rewrites _mounts so
|
|
# the proxied pool (HTTPProxy) sits alongside the base transport.
|
|
proxied_pools = [
|
|
type(mount._pool).__name__
|
|
for mount in http_client._mounts.values()
|
|
if mount is not None and hasattr(mount, "_pool")
|
|
]
|
|
assert "HTTPProxy" in proxied_pools, (
|
|
"Expected httpx.Client to route through HTTPProxy when HTTPS_PROXY is "
|
|
"set; found pools: %r" % (proxied_pools,)
|
|
)
|
|
http_client.close()
|
|
|
|
|
|
@patch("run_agent.OpenAI")
|
|
def test_create_openai_client_no_proxy_when_env_unset(mock_openai, monkeypatch):
|
|
"""Without proxy env vars, the keepalive transport must still be installed
|
|
and no HTTPProxy mount should exist."""
|
|
for key in ("HTTPS_PROXY", "HTTP_PROXY", "ALL_PROXY",
|
|
"https_proxy", "http_proxy", "all_proxy"):
|
|
monkeypatch.delenv(key, raising=False)
|
|
|
|
agent = _make_agent()
|
|
kwargs = {
|
|
"api_key": "test-key",
|
|
"base_url": "https://chatgpt.com/backend-api/codex",
|
|
}
|
|
agent._create_openai_client(kwargs, reason="test", shared=False)
|
|
|
|
forwarded = mock_openai.call_args.kwargs
|
|
http_client = _extract_http_client(forwarded)
|
|
assert isinstance(http_client, httpx.Client)
|
|
pool_types = [
|
|
type(mount._pool).__name__
|
|
for mount in http_client._mounts.values()
|
|
if mount is not None and hasattr(mount, "_pool")
|
|
]
|
|
assert "HTTPProxy" not in pool_types, (
|
|
"No proxy env set but httpx.Client still mounted HTTPProxy; "
|
|
"pools were %r" % (pool_types,)
|
|
)
|
|
http_client.close()
|