"""Profile-path parity tests: verify profile path produces identical output to legacy flags.

Each test calls build_kwargs twice — once with legacy flags, once with provider_profile —
and asserts the output is identical. This catches any behavioral drift between the two paths.
"""

import pytest
from agent.transports.chat_completions import ChatCompletionsTransport
from providers import get_provider_profile


@pytest.fixture
def transport():
    return ChatCompletionsTransport()


def _msgs():
    return [{"role": "user", "content": "hello"}]


def _max_tokens_fn(n):
    return {"max_completion_tokens": n}


class TestNvidiaProfileParity:
    def test_max_tokens_match(self, transport):
        """NVIDIA profile sets max_tokens=16384; legacy flag is removed."""
        profile = transport.build_kwargs(
            model="nvidia/nemotron", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("nvidia"),
            max_tokens_param_fn=_max_tokens_fn,
        )
        assert profile["max_completion_tokens"] == 16384


class TestKimiProfileParity:
    def test_temperature_omitted(self, transport):
        legacy = transport.build_kwargs(
            model="kimi-k2", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("kimi-coding"), omit_temperature=True,
        )
        profile = transport.build_kwargs(
            model="kimi-k2", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("kimi"),
        )
        assert "temperature" not in legacy
        assert "temperature" not in profile

    def test_max_tokens(self, transport):
        legacy = transport.build_kwargs(
            model="kimi-k2", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("kimi-coding"), max_tokens_param_fn=_max_tokens_fn,
        )
        profile = transport.build_kwargs(
            model="kimi-k2", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("kimi"),
            max_tokens_param_fn=_max_tokens_fn,
        )
        assert profile["max_completion_tokens"] == legacy["max_completion_tokens"] == 32000

    def test_thinking_enabled(self, transport):
        rc = {"enabled": True, "effort": "high"}
        legacy = transport.build_kwargs(
            model="kimi-k2", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("kimi-coding"), reasoning_config=rc,
        )
        profile = transport.build_kwargs(
            model="kimi-k2", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("kimi"),
            reasoning_config=rc,
        )
        assert profile["extra_body"]["thinking"] == legacy["extra_body"]["thinking"]
        assert profile["reasoning_effort"] == legacy["reasoning_effort"] == "high"

    def test_thinking_disabled(self, transport):
        rc = {"enabled": False}
        legacy = transport.build_kwargs(
            model="kimi-k2", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("kimi-coding"), reasoning_config=rc,
        )
        profile = transport.build_kwargs(
            model="kimi-k2", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("kimi"),
            reasoning_config=rc,
        )
        assert profile["extra_body"]["thinking"] == legacy["extra_body"]["thinking"]
        assert profile["extra_body"]["thinking"]["type"] == "disabled"
        assert "reasoning_effort" not in profile
        assert "reasoning_effort" not in legacy

    def test_reasoning_effort_default(self, transport):
        rc = {"enabled": True}
        legacy = transport.build_kwargs(
            model="kimi-k2", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("kimi-coding"), reasoning_config=rc,
        )
        profile = transport.build_kwargs(
            model="kimi-k2", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("kimi"),
            reasoning_config=rc,
        )
        assert profile["reasoning_effort"] == legacy["reasoning_effort"] == "medium"


class TestOpenRouterProfileParity:
    def test_provider_preferences(self, transport):
        prefs = {"allow": ["anthropic"]}
        legacy = transport.build_kwargs(
            model="anthropic/claude-sonnet-4.6", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("openrouter"), provider_preferences=prefs,
        )
        profile = transport.build_kwargs(
            model="anthropic/claude-sonnet-4.6", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("openrouter"),
            provider_preferences=prefs,
        )
        assert profile["extra_body"]["provider"] == legacy["extra_body"]["provider"]

    def test_reasoning_full_config(self, transport):
        rc = {"enabled": True, "effort": "high"}
        legacy = transport.build_kwargs(
            model="anthropic/claude-sonnet-4.6", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("openrouter"), supports_reasoning=True, reasoning_config=rc,
        )
        profile = transport.build_kwargs(
            model="anthropic/claude-sonnet-4.6", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("openrouter"),
            supports_reasoning=True, reasoning_config=rc,
        )
        assert profile["extra_body"]["reasoning"] == legacy["extra_body"]["reasoning"]

    def test_default_reasoning(self, transport):
        legacy = transport.build_kwargs(
            model="anthropic/claude-sonnet-4.6", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("openrouter"), supports_reasoning=True,
        )
        profile = transport.build_kwargs(
            model="anthropic/claude-sonnet-4.6", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("openrouter"),
            supports_reasoning=True,
        )
        assert profile["extra_body"]["reasoning"] == legacy["extra_body"]["reasoning"]


class TestNousProfileParity:
    def test_tags(self, transport):
        legacy = transport.build_kwargs(
            model="hermes-3", messages=_msgs(), tools=None, provider_profile=get_provider_profile("nous"),
        )
        profile = transport.build_kwargs(
            model="hermes-3", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("nous"),
        )
        assert profile["extra_body"]["tags"] == legacy["extra_body"]["tags"]

    def test_reasoning_omitted_when_disabled(self, transport):
        rc = {"enabled": False}
        legacy = transport.build_kwargs(
            model="hermes-3", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("nous"), supports_reasoning=True, reasoning_config=rc,
        )
        profile = transport.build_kwargs(
            model="hermes-3", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("nous"),
            supports_reasoning=True, reasoning_config=rc,
        )
        assert "reasoning" not in legacy.get("extra_body", {})
        assert "reasoning" not in profile.get("extra_body", {})


class TestQwenProfileParity:
    def test_max_tokens(self, transport):
        legacy = transport.build_kwargs(
            model="qwen3.5", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("qwen-oauth"), max_tokens_param_fn=_max_tokens_fn,
        )
        profile = transport.build_kwargs(
            model="qwen3.5", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("qwen"),
            max_tokens_param_fn=_max_tokens_fn,
        )
        assert profile["max_completion_tokens"] == legacy["max_completion_tokens"] == 65536

    def test_vl_high_resolution(self, transport):
        legacy = transport.build_kwargs(
            model="qwen3.5", messages=_msgs(), tools=None, provider_profile=get_provider_profile("qwen-oauth"),
        )
        profile = transport.build_kwargs(
            model="qwen3.5", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("qwen"),
        )
        assert profile["extra_body"]["vl_high_resolution_images"] == legacy["extra_body"]["vl_high_resolution_images"]

    def test_metadata_top_level(self, transport):
        meta = {"sessionId": "s123", "promptId": "p456"}
        legacy = transport.build_kwargs(
            model="qwen3.5", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("qwen-oauth"), qwen_session_metadata=meta,
        )
        profile = transport.build_kwargs(
            model="qwen3.5", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("qwen"),
            qwen_session_metadata=meta,
        )
        assert profile["metadata"] == legacy["metadata"] == meta
        assert "metadata" not in profile.get("extra_body", {})

    def test_message_preprocessing(self, transport):
        """Qwen profile normalizes string content to list-of-parts."""
        msgs = [
            {"role": "system", "content": "You are helpful."},
            {"role": "user", "content": "hello"},
        ]
        profile = transport.build_kwargs(
            model="qwen3.5", messages=msgs, tools=None,
            provider_profile=get_provider_profile("qwen"),
        )
        out_msgs = profile["messages"]
        # System message content normalized + cache_control injected
        assert isinstance(out_msgs[0]["content"], list)
        assert out_msgs[0]["content"][0]["type"] == "text"
        assert "cache_control" in out_msgs[0]["content"][-1]
        # User message content normalized
        assert isinstance(out_msgs[1]["content"], list)
        assert out_msgs[1]["content"][0] == {"type": "text", "text": "hello"}


class TestDeveloperRoleParity:
    """Developer role swap must work on BOTH legacy and profile paths."""

    def test_legacy_path_swaps_for_gpt5(self, transport):
        msgs = [{"role": "system", "content": "Be helpful"}, {"role": "user", "content": "hi"}]
        kw = transport.build_kwargs(
            model="gpt-5.4", messages=msgs, tools=None,
        )
        assert kw["messages"][0]["role"] == "developer"

    def test_profile_path_swaps_for_gpt5(self, transport):
        msgs = [{"role": "system", "content": "Be helpful"}, {"role": "user", "content": "hi"}]
        kw = transport.build_kwargs(
            model="gpt-5.4", messages=msgs, tools=None,
            provider_profile=get_provider_profile("openrouter"),
        )
        assert kw["messages"][0]["role"] == "developer"

    def test_profile_path_no_swap_for_claude(self, transport):
        msgs = [{"role": "system", "content": "Be helpful"}, {"role": "user", "content": "hi"}]
        kw = transport.build_kwargs(
            model="anthropic/claude-sonnet-4.6", messages=msgs, tools=None,
            provider_profile=get_provider_profile("openrouter"),
        )
        assert kw["messages"][0]["role"] == "system"


class TestRequestOverridesParity:
    """request_overrides with extra_body must merge identically on both paths."""

    def test_extra_body_override_legacy(self, transport):
        kw = transport.build_kwargs(
            model="gpt-5.4", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("openrouter"),
            request_overrides={"extra_body": {"custom_key": "custom_val"}},
        )
        assert kw["extra_body"]["custom_key"] == "custom_val"

    def test_extra_body_override_profile(self, transport):
        kw = transport.build_kwargs(
            model="gpt-5.4", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("openrouter"),
            request_overrides={"extra_body": {"custom_key": "custom_val"}},
        )
        assert kw["extra_body"]["custom_key"] == "custom_val"

    def test_extra_body_override_merges_with_provider_body(self, transport):
        """Override extra_body merges WITH provider extra_body, not replaces."""
        kw = transport.build_kwargs(
            model="hermes-3", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("nous"),
            request_overrides={"extra_body": {"custom": True}},
        )
        assert kw["extra_body"]["tags"] == ["product=hermes-agent"]  # from profile
        assert kw["extra_body"]["custom"] is True  # from override

    def test_top_level_override(self, transport):
        kw = transport.build_kwargs(
            model="gpt-5.4", messages=_msgs(), tools=None,
            provider_profile=get_provider_profile("openrouter"),
            request_overrides={"top_p": 0.9},
        )
        assert kw["top_p"] == 0.9
