diff --git a/python/packages/autogen-agentchat/tests/test_group_chat.py b/python/packages/autogen-agentchat/tests/test_group_chat.py index 4676c7a5fee5..25194d6049af 100644 --- a/python/packages/autogen-agentchat/tests/test_group_chat.py +++ b/python/packages/autogen-agentchat/tests/test_group_chat.py @@ -34,10 +34,10 @@ from autogen_agentchat.teams._group_chat._swarm_group_chat import SwarmGroupChatManager from autogen_agentchat.ui import Console from autogen_core import AgentId, CancellationToken -from autogen_core.models import ReplayChatCompletionClient from autogen_core.tools import FunctionTool from autogen_ext.code_executors.local import LocalCommandLineCodeExecutor from autogen_ext.models.openai import OpenAIChatCompletionClient +from autogen_ext.models.replay import ReplayChatCompletionClient from openai.resources.chat.completions import AsyncCompletions from openai.types.chat.chat_completion import ChatCompletion, Choice from openai.types.chat.chat_completion_chunk import ChatCompletionChunk diff --git a/python/packages/autogen-agentchat/tests/test_magentic_one_group_chat.py b/python/packages/autogen-agentchat/tests/test_magentic_one_group_chat.py index 4b6df430fee7..4a2f831aeba6 100644 --- a/python/packages/autogen-agentchat/tests/test_magentic_one_group_chat.py +++ b/python/packages/autogen-agentchat/tests/test_magentic_one_group_chat.py @@ -18,7 +18,7 @@ ) from autogen_agentchat.teams._group_chat._magentic_one._magentic_one_orchestrator import MagenticOneOrchestrator from autogen_core import AgentId, CancellationToken -from autogen_core.models import ReplayChatCompletionClient +from autogen_ext.models.replay import ReplayChatCompletionClient from utils import FileLogHandler logger = logging.getLogger(EVENT_LOGGER_NAME) diff --git a/python/packages/autogen-core/docs/src/reference/index.md b/python/packages/autogen-core/docs/src/reference/index.md index 357aa58cf5cc..608866ce1a14 100644 --- a/python/packages/autogen-core/docs/src/reference/index.md +++ b/python/packages/autogen-core/docs/src/reference/index.md @@ -48,12 +48,16 @@ python/autogen_ext.agents.video_surfer python/autogen_ext.agents.video_surfer.tools python/autogen_ext.auth.azure python/autogen_ext.teams.magentic_one +python/autogen_ext.models.cache python/autogen_ext.models.openai +python/autogen_ext.models.replay python/autogen_ext.tools.langchain python/autogen_ext.tools.graphrag python/autogen_ext.tools.code_execution python/autogen_ext.code_executors.local python/autogen_ext.code_executors.docker python/autogen_ext.code_executors.azure +python/autogen_ext.cache_store.diskcache +python/autogen_ext.cache_store.redis python/autogen_ext.runtimes.grpc ``` diff --git a/python/packages/autogen-core/docs/src/reference/python/autogen_ext.cache_store.diskcache.rst b/python/packages/autogen-core/docs/src/reference/python/autogen_ext.cache_store.diskcache.rst new file mode 100644 index 000000000000..5fbc4c8b35ac --- /dev/null +++ b/python/packages/autogen-core/docs/src/reference/python/autogen_ext.cache_store.diskcache.rst @@ -0,0 +1,8 @@ +autogen\_ext.cache_store.diskcache +================================== + + +.. automodule:: autogen_ext.cache_store.diskcache + :members: + :undoc-members: + :show-inheritance: \ No newline at end of file diff --git a/python/packages/autogen-core/docs/src/reference/python/autogen_ext.cache_store.redis.rst b/python/packages/autogen-core/docs/src/reference/python/autogen_ext.cache_store.redis.rst new file mode 100644 index 000000000000..fab1b46d520a --- /dev/null +++ b/python/packages/autogen-core/docs/src/reference/python/autogen_ext.cache_store.redis.rst @@ -0,0 +1,8 @@ +autogen\_ext.cache_store.redis +============================== + + +.. automodule:: autogen_ext.cache_store.redis + :members: + :undoc-members: + :show-inheritance: \ No newline at end of file diff --git a/python/packages/autogen-core/docs/src/reference/python/autogen_ext.models.cache.rst b/python/packages/autogen-core/docs/src/reference/python/autogen_ext.models.cache.rst new file mode 100644 index 000000000000..48956ace16e2 --- /dev/null +++ b/python/packages/autogen-core/docs/src/reference/python/autogen_ext.models.cache.rst @@ -0,0 +1,8 @@ +autogen\_ext.models.cache +========================= + + +.. automodule:: autogen_ext.models.cache + :members: + :undoc-members: + :show-inheritance: \ No newline at end of file diff --git a/python/packages/autogen-core/docs/src/reference/python/autogen_ext.models.replay.rst b/python/packages/autogen-core/docs/src/reference/python/autogen_ext.models.replay.rst new file mode 100644 index 000000000000..4fc9aefbfb3d --- /dev/null +++ b/python/packages/autogen-core/docs/src/reference/python/autogen_ext.models.replay.rst @@ -0,0 +1,8 @@ +autogen\_ext.models.replay +========================== + + +.. automodule:: autogen_ext.models.replay + :members: + :undoc-members: + :show-inheritance: \ No newline at end of file diff --git a/python/packages/autogen-core/docs/src/user-guide/agentchat-user-guide/tutorial/models.ipynb b/python/packages/autogen-core/docs/src/user-guide/agentchat-user-guide/tutorial/models.ipynb index f71c14455338..a3a423553160 100644 --- a/python/packages/autogen-core/docs/src/user-guide/agentchat-user-guide/tutorial/models.ipynb +++ b/python/packages/autogen-core/docs/src/user-guide/agentchat-user-guide/tutorial/models.ipynb @@ -8,7 +8,9 @@ "\n", "In many cases, agents need access to LLM model services such as OpenAI, Azure OpenAI, or local models. Since there are many different providers with different APIs, `autogen-core` implements a protocol for [model clients](../../core-user-guide/framework/model-clients.ipynb) and `autogen-ext` implements a set of model clients for popular model services. AgentChat can use these model clients to interact with model services. \n", "\n", - "**NOTE:** See {py:class}`~autogen_core.models.ChatCompletionCache` for a caching wrapper to use with the following clients." + "```{note}\n", + "See {py:class}`~autogen_ext.models.cache.ChatCompletionCache` for a caching wrapper to use with the following clients.\n", + "```" ] }, { diff --git a/python/packages/autogen-core/docs/src/user-guide/core-user-guide/framework/model-clients.ipynb b/python/packages/autogen-core/docs/src/user-guide/core-user-guide/framework/model-clients.ipynb index 9bb8f79be571..38fd13195c6f 100644 --- a/python/packages/autogen-core/docs/src/user-guide/core-user-guide/framework/model-clients.ipynb +++ b/python/packages/autogen-core/docs/src/user-guide/core-user-guide/framework/model-clients.ipynb @@ -327,9 +327,20 @@ "source": [ "## Caching Wrapper\n", "\n", - "`autogen_core` implements a {py:class}`~autogen_core.models.ChatCompletionCache` that can wrap any {py:class}`~autogen_core.models.ChatCompletionClient`. Using this wrapper avoids incurring token usage when querying the underlying client with the same prompt multiple times. \n", + "`autogen_ext` implements {py:class}`~autogen_ext.models.cache.ChatCompletionCache` that can wrap any {py:class}`~autogen_core.models.ChatCompletionClient`. Using this wrapper avoids incurring token usage when querying the underlying client with the same prompt multiple times.\n", "\n", - "{py:class}`~autogen_core.models.ChatCompletionCache` uses a {py:class}`~autogen_core.CacheStore` protocol to allow duck-typing any storage object that has a pair of `get` & `set` methods (such as `redis.Redis` or `diskcache.Cache`). Here's an example of using `diskcache` for local caching:" + "{py:class}`~autogen_core.models.ChatCompletionCache` uses a {py:class}`~autogen_core.CacheStore` protocol. We have implemented some useful variants of {py:class}`~autogen_core.CacheStore` including {py:class}`~autogen_ext.cache_store.diskcache.DiskCacheStore` and {py:class}`~autogen_ext.cache_store.redis.RedisStore`.\n", + "\n", + "Here's an example of using `diskcache` for local caching:" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": {}, + "outputs": [], + "source": [ + "# pip install -U \"autogen-ext[openai, diskcache]\"" ] }, { @@ -346,18 +357,37 @@ } ], "source": [ - "from typing import Any, Dict, Optional\n", + "import asyncio\n", + "import tempfile\n", "\n", - "from autogen_core.models import ChatCompletionCache\n", + "from autogen_core.models import UserMessage\n", + "from autogen_ext.cache_store.diskcache import DiskCacheStore\n", + "from autogen_ext.models.cache import CHAT_CACHE_VALUE_TYPE, ChatCompletionCache\n", + "from autogen_ext.models.openai import OpenAIChatCompletionClient\n", "from diskcache import Cache\n", "\n", - "diskcache_client = Cache(\"/tmp/diskcache\")\n", "\n", - "cached_client = ChatCompletionCache(model_client, diskcache_client)\n", - "response = await cached_client.create(messages=messages)\n", + "async def main() -> None:\n", + " with tempfile.TemporaryDirectory() as tmpdirname:\n", + " # Initialize the original client\n", + " openai_model_client = OpenAIChatCompletionClient(model=\"gpt-4o\")\n", + "\n", + " # Then initialize the CacheStore, in this case with diskcache.Cache.\n", + " # You can also use redis like:\n", + " # from autogen_ext.cache_store.redis import RedisStore\n", + " # import redis\n", + " # redis_instance = redis.Redis()\n", + " # cache_store = RedisCacheStore[CHAT_CACHE_VALUE_TYPE](redis_instance)\n", + " cache_store = DiskCacheStore[CHAT_CACHE_VALUE_TYPE](Cache(tmpdirname))\n", + " cache_client = ChatCompletionCache(openai_model_client, cache_store)\n", + "\n", + " response = await cache_client.create([UserMessage(content=\"Hello, how are you?\", source=\"user\")])\n", + " print(response) # Should print response from OpenAI\n", + " response = await cache_client.create([UserMessage(content=\"Hello, how are you?\", source=\"user\")])\n", + " print(response) # Should print cached response\n", + "\n", "\n", - "cached_response = await cached_client.create(messages=messages)\n", - "print(cached_response.cached)" + "asyncio.run(main())" ] }, { diff --git a/python/packages/autogen-core/src/autogen_core/__init__.py b/python/packages/autogen-core/src/autogen_core/__init__.py index 1283f6d5f397..0198544ca61e 100644 --- a/python/packages/autogen-core/src/autogen_core/__init__.py +++ b/python/packages/autogen-core/src/autogen_core/__init__.py @@ -10,7 +10,7 @@ from ._agent_runtime import AgentRuntime from ._agent_type import AgentType from ._base_agent import BaseAgent -from ._cache_store import CacheStore +from ._cache_store import CacheStore, InMemoryStore from ._cancellation_token import CancellationToken from ._closure_agent import ClosureAgent, ClosureContext from ._component_config import ( @@ -87,6 +87,7 @@ "AgentRuntime", "BaseAgent", "CacheStore", + "InMemoryStore", "CancellationToken", "AgentInstantiationContext", "TopicId", diff --git a/python/packages/autogen-core/src/autogen_core/_cache_store.py b/python/packages/autogen-core/src/autogen_core/_cache_store.py index 92bafde1d02a..7f3a12185e6c 100644 --- a/python/packages/autogen-core/src/autogen_core/_cache_store.py +++ b/python/packages/autogen-core/src/autogen_core/_cache_store.py @@ -1,15 +1,16 @@ -from typing import Any, Optional, Protocol +from typing import Any, Dict, Generic, Optional, Protocol, TypeVar, cast +T = TypeVar("T") -class CacheStore(Protocol): + +class CacheStore(Protocol, Generic[T]): """ This protocol defines the basic interface for store/cache operations. - Allows duck-typing with any object that implements the get and set methods, - such as redis or diskcache interfaces. + Sub-classes should handle the lifecycle of underlying storage. """ - def get(self, key: str, default: Optional[Any] = None) -> Optional[Any]: + def get(self, key: str, default: Optional[T] = None) -> Optional[T]: """ Retrieve an item from the store. @@ -23,7 +24,7 @@ def get(self, key: str, default: Optional[Any] = None) -> Optional[Any]: """ ... - def set(self, key: str, value: Any) -> Optional[Any]: + def set(self, key: str, value: T) -> None: """ Set an item in the store. @@ -32,3 +33,14 @@ def set(self, key: str, value: Any) -> Optional[Any]: value: The value to be stored in the store. """ ... + + +class InMemoryStore(CacheStore[T]): + def __init__(self) -> None: + self.store: Dict[str, Any] = {} + + def get(self, key: str, default: Optional[T] = None) -> Optional[T]: + return cast(Optional[T], self.store.get(key, default)) + + def set(self, key: str, value: T) -> None: + self.store[key] = value diff --git a/python/packages/autogen-core/src/autogen_core/models/__init__.py b/python/packages/autogen-core/src/autogen_core/models/__init__.py index 9c958a540721..c9fa23ffa110 100644 --- a/python/packages/autogen-core/src/autogen_core/models/__init__.py +++ b/python/packages/autogen-core/src/autogen_core/models/__init__.py @@ -1,6 +1,4 @@ -from ._cache import ChatCompletionCache from ._model_client import ChatCompletionClient, ModelCapabilities, ModelFamily, ModelInfo # type: ignore -from ._replay_chat_completion_client import ReplayChatCompletionClient from ._types import ( AssistantMessage, ChatCompletionTokenLogprob, @@ -17,7 +15,6 @@ __all__ = [ "ModelCapabilities", - "ChatCompletionCache", "ChatCompletionClient", "SystemMessage", "UserMessage", @@ -32,5 +29,4 @@ "ChatCompletionTokenLogprob", "ModelFamily", "ModelInfo", - "ReplayChatCompletionClient", ] diff --git a/python/packages/autogen-core/tests/test_cache_store.py b/python/packages/autogen-core/tests/test_cache_store.py index cbd7ae4aa2fc..3caf058af053 100644 --- a/python/packages/autogen-core/tests/test_cache_store.py +++ b/python/packages/autogen-core/tests/test_cache_store.py @@ -1,6 +1,6 @@ from unittest.mock import Mock -from autogen_core import CacheStore +from autogen_core import CacheStore, InMemoryStore def test_set_and_get_object_key_value() -> None: @@ -30,3 +30,19 @@ def test_set_overwrite_existing_key() -> None: mock_store.get.return_value = new_value mock_store.set.assert_called_with(key, new_value) assert mock_store.get(key) == new_value + + +def test_inmemory_store() -> None: + store = InMemoryStore[int]() + test_key = "test_key" + test_value = 42 + store.set(test_key, test_value) + assert store.get(test_key) == test_value + + new_value = 2 + store.set(test_key, new_value) + assert store.get(test_key) == new_value + + key = "non_existent_key" + default_value = 99 + assert store.get(key, default_value) == default_value diff --git a/python/packages/autogen-ext/pyproject.toml b/python/packages/autogen-ext/pyproject.toml index 5fa2ce54379c..e0f05425b487 100644 --- a/python/packages/autogen-ext/pyproject.toml +++ b/python/packages/autogen-ext/pyproject.toml @@ -46,6 +46,12 @@ video-surfer = [ "ffmpeg-python", "openai-whisper", ] +diskcache = [ + "diskcache>=5.6.3" +] +redis = [ + "redis>=5.2.1" +] grpc = [ "grpcio~=1.62.0", # TODO: update this once we have a stable version. diff --git a/python/packages/autogen-ext/src/autogen_ext/cache_store/__init__.py b/python/packages/autogen-ext/src/autogen_ext/cache_store/__init__.py new file mode 100644 index 000000000000..e69de29bb2d1 diff --git a/python/packages/autogen-ext/src/autogen_ext/cache_store/diskcache.py b/python/packages/autogen-ext/src/autogen_ext/cache_store/diskcache.py new file mode 100644 index 000000000000..afb1db224253 --- /dev/null +++ b/python/packages/autogen-ext/src/autogen_ext/cache_store/diskcache.py @@ -0,0 +1,26 @@ +from typing import Any, Optional, TypeVar, cast + +import diskcache +from autogen_core import CacheStore + +T = TypeVar("T") + + +class DiskCacheStore(CacheStore[T]): + """ + A typed CacheStore implementation that uses diskcache as the underlying storage. + See :class:`~autogen_ext.models.cache.ChatCompletionCache` for an example of usage. + + Args: + cache_instance: An instance of diskcache.Cache. + The user is responsible for managing the DiskCache instance's lifetime. + """ + + def __init__(self, cache_instance: diskcache.Cache): # type: ignore[no-any-unimported] + self.cache = cache_instance + + def get(self, key: str, default: Optional[T] = None) -> Optional[T]: + return cast(Optional[T], self.cache.get(key, default)) # type: ignore[reportUnknownMemberType] + + def set(self, key: str, value: T) -> None: + self.cache.set(key, cast(Any, value)) # type: ignore[reportUnknownMemberType] diff --git a/python/packages/autogen-ext/src/autogen_ext/cache_store/redis.py b/python/packages/autogen-ext/src/autogen_ext/cache_store/redis.py new file mode 100644 index 000000000000..e751f418082c --- /dev/null +++ b/python/packages/autogen-ext/src/autogen_ext/cache_store/redis.py @@ -0,0 +1,29 @@ +from typing import Any, Optional, TypeVar, cast + +import redis +from autogen_core import CacheStore + +T = TypeVar("T") + + +class RedisStore(CacheStore[T]): + """ + A typed CacheStore implementation that uses redis as the underlying storage. + See :class:`~autogen_ext.models.cache.ChatCompletionCache` for an example of usage. + + Args: + cache_instance: An instance of `redis.Redis`. + The user is responsible for managing the Redis instance's lifetime. + """ + + def __init__(self, redis_instance: redis.Redis): + self.cache = redis_instance + + def get(self, key: str, default: Optional[T] = None) -> Optional[T]: + value = cast(Optional[T], self.cache.get(key)) + if value is None: + return default + return value + + def set(self, key: str, value: T) -> None: + self.cache.set(key, cast(Any, value)) diff --git a/python/packages/autogen-ext/src/autogen_ext/models/cache/__init__.py b/python/packages/autogen-ext/src/autogen_ext/models/cache/__init__.py new file mode 100644 index 000000000000..333d2b737a53 --- /dev/null +++ b/python/packages/autogen-ext/src/autogen_ext/models/cache/__init__.py @@ -0,0 +1,6 @@ +from ._chat_completion_cache import CHAT_CACHE_VALUE_TYPE, ChatCompletionCache + +__all__ = [ + "CHAT_CACHE_VALUE_TYPE", + "ChatCompletionCache", +] diff --git a/python/packages/autogen-core/src/autogen_core/models/_cache.py b/python/packages/autogen-ext/src/autogen_ext/models/cache/_chat_completion_cache.py similarity index 67% rename from python/packages/autogen-core/src/autogen_core/models/_cache.py rename to python/packages/autogen-ext/src/autogen_ext/models/cache/_chat_completion_cache.py index fe85381c6c79..79ed5f1660a0 100644 --- a/python/packages/autogen-core/src/autogen_core/models/_cache.py +++ b/python/packages/autogen-ext/src/autogen_ext/models/cache/_chat_completion_cache.py @@ -3,74 +3,80 @@ import warnings from typing import Any, AsyncGenerator, List, Mapping, Optional, Sequence, Union, cast -from .._cache_store import CacheStore -from .._cancellation_token import CancellationToken -from ..tools import Tool, ToolSchema -from ._model_client import ( +from autogen_core import CacheStore, CancellationToken +from autogen_core.models import ( ChatCompletionClient, - ModelCapabilities, # type: ignore - ModelInfo, -) -from ._types import ( CreateResult, LLMMessage, + ModelCapabilities, # type: ignore + ModelInfo, RequestUsage, ) +from autogen_core.tools import Tool, ToolSchema + +CHAT_CACHE_VALUE_TYPE = Union[CreateResult, List[Union[str, CreateResult]]] class ChatCompletionCache(ChatCompletionClient): """ - A wrapper around a ChatCompletionClient that caches creation results from an underlying client. + A wrapper around a :class:`~autogen_ext.models.cache.ChatCompletionClient` that caches + creation results from an underlying client. Cache hits do not contribute to token usage of the original client. Typical Usage: - Lets use caching with `openai` as an example: + Lets use caching on disk with `openai` client as an example. + First install `autogen-ext` with the required packages: - .. code-block:: bash + .. code-block:: bash - pip install "autogen-ext[openai]==0.4.0.dev13" + pip install -U "autogen-ext[openai, diskcache]" - And use it as: + And use it as: - .. code-block:: python + .. code-block:: python - # Initialize the original client - from autogen_ext.models.openai import OpenAIChatCompletionClient + import asyncio + import tempfile - openai_client = OpenAIChatCompletionClient( - model="gpt-4o-2024-08-06", - # api_key="sk-...", # Optional if you have an OPENAI_API_KEY environment variable set. - ) + from autogen_core.models import UserMessage + from autogen_ext.models.openai import OpenAIChatCompletionClient + from autogen_ext.models.cache import ChatCompletionCache, CHAT_CACHE_VALUE_TYPE + from autogen_ext.cache_store.diskcache import DiskCacheStore + from diskcache import Cache - # Then initialize the CacheStore. Either a Redis store: - import redis - redis_client = redis.Redis(host="localhost", port=6379, db=0) + async def main(): + with tempfile.TemporaryDirectory() as tmpdirname: + # Initialize the original client + openai_model_client = OpenAIChatCompletionClient(model="gpt-4o") - # or diskcache: - from diskcache import Cache + # Then initialize the CacheStore, in this case with diskcache.Cache. + # You can also use redis like: + # from autogen_ext.cache_store.redis import RedisStore + # import redis + # redis_instance = redis.Redis() + # cache_store = RedisCacheStore[CHAT_CACHE_VALUE_TYPE](redis_instance) + cache_store = DiskCacheStore[CHAT_CACHE_VALUE_TYPE](Cache(tmpdirname)) + cache_client = ChatCompletionCache(openai_model_client, cache_store) - diskcache_client = Cache("/tmp/diskcache") + response = await cache_client.create([UserMessage(content="Hello, how are you?", source="user")]) + print(response) # Should print response from OpenAI + response = await cache_client.create([UserMessage(content="Hello, how are you?", source="user")]) + print(response) # Should print cached response - # Then initialize the ChatCompletionCache with the store: - from autogen_core.models import ChatCompletionCache - # Cached client - cached_client = ChatCompletionCache(openai_client, diskcache_client) + asyncio.run(main()) - You can now use the `cached_client` as you would the original client, but with caching enabled. - """ + You can now use the `cached_client` as you would the original client, but with caching enabled. - def __init__(self, client: ChatCompletionClient, store: CacheStore): - """ - Initialize a new ChatCompletionCache. + Args: + client (ChatCompletionClient): The original ChatCompletionClient to wrap. + store (CacheStore): A store object that implements get and set methods. + The user is responsible for managing the store's lifecycle & clearing it (if needed). + """ - Args: - client (ChatCompletionClient): The original ChatCompletionClient to wrap. - store (CacheStore): A store object that implements get and set methods. - The user is responsible for managing the store's lifecycle & clearing it (if needed). - """ + def __init__(self, client: ChatCompletionClient, store: CacheStore[CHAT_CACHE_VALUE_TYPE]): self.client = client self.store = store diff --git a/python/packages/autogen-ext/src/autogen_ext/models/replay/__init__.py b/python/packages/autogen-ext/src/autogen_ext/models/replay/__init__.py new file mode 100644 index 000000000000..6e6da6f0a910 --- /dev/null +++ b/python/packages/autogen-ext/src/autogen_ext/models/replay/__init__.py @@ -0,0 +1,5 @@ +from ._replay_chat_completion_client import ReplayChatCompletionClient + +__all__ = [ + "ReplayChatCompletionClient", +] diff --git a/python/packages/autogen-core/src/autogen_core/models/_replay_chat_completion_client.py b/python/packages/autogen-ext/src/autogen_ext/models/replay/_replay_chat_completion_client.py similarity index 94% rename from python/packages/autogen-core/src/autogen_core/models/_replay_chat_completion_client.py rename to python/packages/autogen-ext/src/autogen_ext/models/replay/_replay_chat_completion_client.py index f6ee5ac2aca6..5ae7b6b665eb 100644 --- a/python/packages/autogen-core/src/autogen_core/models/_replay_chat_completion_client.py +++ b/python/packages/autogen-ext/src/autogen_ext/models/replay/_replay_chat_completion_client.py @@ -4,20 +4,17 @@ import warnings from typing import Any, AsyncGenerator, List, Mapping, Optional, Sequence, Union -from .. import EVENT_LOGGER_NAME -from .._cancellation_token import CancellationToken -from ..tools import Tool, ToolSchema -from ._model_client import ( +from autogen_core import EVENT_LOGGER_NAME, CancellationToken +from autogen_core.models import ( ChatCompletionClient, + CreateResult, + LLMMessage, ModelCapabilities, # type: ignore ModelFamily, ModelInfo, -) -from ._types import ( - CreateResult, - LLMMessage, RequestUsage, ) +from autogen_core.tools import Tool, ToolSchema logger = logging.getLogger(EVENT_LOGGER_NAME) @@ -43,7 +40,8 @@ class ReplayChatCompletionClient(ChatCompletionClient): .. code-block:: python - from autogen_core.models import ReplayChatCompletionClient, UserMessage + from autogen_core.models import UserMessage + from autogen_ext.models.replay import ReplayChatCompletionClient async def example(): @@ -62,7 +60,8 @@ async def example(): .. code-block:: python import asyncio - from autogen_core.models import ReplayChatCompletionClient, UserMessage + from autogen_core.models import UserMessage + from autogen_ext.models.replay import ReplayChatCompletionClient async def example(): @@ -87,7 +86,8 @@ async def example(): .. code-block:: python import asyncio - from autogen_core.models import ReplayChatCompletionClient, UserMessage + from autogen_core.models import UserMessage + from autogen_ext.models.replay import ReplayChatCompletionClient async def example(): diff --git a/python/packages/autogen-ext/tests/cache_store/test_diskcache_store.py b/python/packages/autogen-ext/tests/cache_store/test_diskcache_store.py new file mode 100644 index 000000000000..ddca0b82cdcc --- /dev/null +++ b/python/packages/autogen-ext/tests/cache_store/test_diskcache_store.py @@ -0,0 +1,48 @@ +import tempfile + +import pytest + +diskcache = pytest.importorskip("diskcache") + + +def test_diskcache_store_basic() -> None: + from autogen_ext.cache_store.diskcache import DiskCacheStore + from diskcache import Cache + + with tempfile.TemporaryDirectory() as temp_dir: + cache = Cache(temp_dir) + store = DiskCacheStore[int](cache) + test_key = "test_key" + test_value = 42 + store.set(test_key, test_value) + assert store.get(test_key) == test_value + + new_value = 2 + store.set(test_key, new_value) + assert store.get(test_key) == new_value + + key = "non_existent_key" + default_value = 99 + assert store.get(key, default_value) == default_value + + +def test_diskcache_with_different_instances() -> None: + from autogen_ext.cache_store.diskcache import DiskCacheStore + from diskcache import Cache + + with tempfile.TemporaryDirectory() as temp_dir_1, tempfile.TemporaryDirectory() as temp_dir_2: + cache_1 = Cache(temp_dir_1) + cache_2 = Cache(temp_dir_2) + + store_1 = DiskCacheStore[int](cache_1) + store_2 = DiskCacheStore[int](cache_2) + + test_key = "test_key" + test_value_1 = 5 + test_value_2 = 6 + + store_1.set(test_key, test_value_1) + assert store_1.get(test_key) == test_value_1 + + store_2.set(test_key, test_value_2) + assert store_2.get(test_key) == test_value_2 diff --git a/python/packages/autogen-ext/tests/cache_store/test_redis_store.py b/python/packages/autogen-ext/tests/cache_store/test_redis_store.py new file mode 100644 index 000000000000..111f38a4fffd --- /dev/null +++ b/python/packages/autogen-ext/tests/cache_store/test_redis_store.py @@ -0,0 +1,53 @@ +from unittest.mock import MagicMock + +import pytest + +redis = pytest.importorskip("redis") + + +def test_redis_store_basic() -> None: + from autogen_ext.cache_store.redis import RedisStore + + redis_instance = MagicMock() + store = RedisStore[int](redis_instance) + test_key = "test_key" + test_value = 42 + store.set(test_key, test_value) + redis_instance.set.assert_called_with(test_key, test_value) + redis_instance.get.return_value = test_value + assert store.get(test_key) == test_value + + new_value = 2 + store.set(test_key, new_value) + redis_instance.set.assert_called_with(test_key, new_value) + redis_instance.get.return_value = new_value + assert store.get(test_key) == new_value + + key = "non_existent_key" + default_value = 99 + redis_instance.get.return_value = None + assert store.get(key, default_value) == default_value + + +def test_redis_with_different_instances() -> None: + from autogen_ext.cache_store.redis import RedisStore + + redis_instance_1 = MagicMock() + redis_instance_2 = MagicMock() + + store_1 = RedisStore[int](redis_instance_1) + store_2 = RedisStore[int](redis_instance_2) + + test_key = "test_key" + test_value_1 = 5 + test_value_2 = 6 + + store_1.set(test_key, test_value_1) + redis_instance_1.set.assert_called_with(test_key, test_value_1) + redis_instance_1.get.return_value = test_value_1 + assert store_1.get(test_key) == test_value_1 + + store_2.set(test_key, test_value_2) + redis_instance_2.set.assert_called_with(test_key, test_value_2) + redis_instance_2.get.return_value = test_value_2 + assert store_2.get(test_key) == test_value_2 diff --git a/python/packages/autogen-core/tests/test_chat_completion_cache.py b/python/packages/autogen-ext/tests/models/test_chat_completion_cache.py similarity index 90% rename from python/packages/autogen-core/tests/test_chat_completion_cache.py rename to python/packages/autogen-ext/tests/models/test_chat_completion_cache.py index c5e2f46cb9da..ceb4d9a9f72a 100644 --- a/python/packages/autogen-core/tests/test_chat_completion_cache.py +++ b/python/packages/autogen-ext/tests/models/test_chat_completion_cache.py @@ -1,29 +1,17 @@ import copy -from typing import Any, List, Optional, Tuple, Union +from typing import List, Tuple, Union import pytest -from autogen_core import CacheStore +from autogen_core import InMemoryStore from autogen_core.models import ( - ChatCompletionCache, ChatCompletionClient, CreateResult, LLMMessage, - ReplayChatCompletionClient, SystemMessage, UserMessage, ) - - -class DictStore(CacheStore): - def __init__(self) -> None: - self._store: dict[str, Any] = {} - - def get(self, key: str, default: Optional[Any] = None) -> Optional[Any]: - return self._store.get(key, default) - - def set(self, key: str, value: Any) -> Optional[Any]: - self._store[key] = value - return None +from autogen_ext.models.cache import CHAT_CACHE_VALUE_TYPE, ChatCompletionCache +from autogen_ext.models.replay import ReplayChatCompletionClient def get_test_data() -> Tuple[list[str], list[str], SystemMessage, ChatCompletionClient, ChatCompletionCache]: @@ -33,7 +21,8 @@ def get_test_data() -> Tuple[list[str], list[str], SystemMessage, ChatCompletion system_prompt = SystemMessage(content="This is a system prompt") replay_client = ReplayChatCompletionClient(responses) replay_client.set_cached_bool_value(False) - cached_client = ChatCompletionCache(replay_client, store=DictStore()) + store = InMemoryStore[CHAT_CACHE_VALUE_TYPE]() + cached_client = ChatCompletionCache(replay_client, store) return responses, prompts, system_prompt, replay_client, cached_client diff --git a/python/packages/autogen-ext/tests/models/test_reply_chat_completion_client.py b/python/packages/autogen-ext/tests/models/test_reply_chat_completion_client.py index db878d9ad089..7c3fe584b656 100644 --- a/python/packages/autogen-ext/tests/models/test_reply_chat_completion_client.py +++ b/python/packages/autogen-ext/tests/models/test_reply_chat_completion_client.py @@ -12,13 +12,8 @@ default_subscription, message_handler, ) -from autogen_core.models import ( - ChatCompletionClient, - CreateResult, - ReplayChatCompletionClient, - SystemMessage, - UserMessage, -) +from autogen_core.models import ChatCompletionClient, CreateResult, SystemMessage, UserMessage +from autogen_ext.models.replay import ReplayChatCompletionClient @dataclass diff --git a/python/uv.lock b/python/uv.lock index 07319f962dfb..f408c1cae420 100644 --- a/python/uv.lock +++ b/python/uv.lock @@ -130,7 +130,7 @@ source = { registry = "https://pypi.org/simple" } dependencies = [ { name = "aiohappyeyeballs" }, { name = "aiosignal" }, - { name = "async-timeout", marker = "python_full_version < '3.11'" }, + { name = "async-timeout", version = "4.0.3", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version < '3.11'" }, { name = "attrs" }, { name = "frozenlist" }, { name = "multidict" }, @@ -317,11 +317,30 @@ wheels = [ name = "async-timeout" version = "4.0.3" source = { registry = "https://pypi.org/simple" } +resolution-markers = [ + "python_full_version < '3.11' and sys_platform == 'darwin'", + "python_full_version < '3.11' and platform_machine == 'aarch64' and sys_platform == 'linux'", + "(python_full_version < '3.11' and platform_machine != 'aarch64' and sys_platform == 'linux') or (python_full_version < '3.11' and sys_platform != 'darwin' and sys_platform != 'linux')", +] sdist = { url = "https://files.pythonhosted.org/packages/87/d6/21b30a550dafea84b1b8eee21b5e23fa16d010ae006011221f33dcd8d7f8/async-timeout-4.0.3.tar.gz", hash = "sha256:4640d96be84d82d02ed59ea2b7105a0f7b33abe8703703cd0ab0bf87c427522f", size = 8345 } wheels = [ { url = "https://files.pythonhosted.org/packages/a7/fa/e01228c2938de91d47b307831c62ab9e4001e747789d0b05baf779a6488c/async_timeout-4.0.3-py3-none-any.whl", hash = "sha256:7405140ff1230c310e51dc27b3145b9092d659ce68ff733fb0cefe3ee42be028", size = 5721 }, ] +[[package]] +name = "async-timeout" +version = "5.0.1" +source = { registry = "https://pypi.org/simple" } +resolution-markers = [ + "python_full_version == '3.11.*' and sys_platform == 'darwin'", + "python_full_version == '3.11.*' and platform_machine == 'aarch64' and sys_platform == 'linux'", + "(python_full_version == '3.11.*' and platform_machine != 'aarch64' and sys_platform == 'linux') or (python_full_version == '3.11.*' and sys_platform != 'darwin' and sys_platform != 'linux')", +] +sdist = { url = "https://files.pythonhosted.org/packages/a5/ae/136395dfbfe00dfc94da3f3e136d0b13f394cba8f4841120e34226265780/async_timeout-5.0.1.tar.gz", hash = "sha256:d9321a7a3d5a6a5e187e824d2fa0793ce379a202935782d555d6e9d2735677d3", size = 9274 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/fe/ba/e2081de779ca30d473f21f5b30e0e737c438205440784c7dfc81efc2b029/async_timeout-5.0.1-py3-none-any.whl", hash = "sha256:39e3809566ff85354557ec2398b55e096c8364bacac9405a7a1fa429e77fe76c", size = 6233 }, +] + [[package]] name = "asyncer" version = "0.0.7" @@ -508,6 +527,9 @@ azure = [ { name = "azure-core" }, { name = "azure-identity" }, ] +diskcache = [ + { name = "diskcache" }, +] docker = [ { name = "docker" }, ] @@ -535,6 +557,9 @@ openai = [ { name = "openai" }, { name = "tiktoken" }, ] +redis = [ + { name = "redis" }, +] video-surfer = [ { name = "autogen-agentchat" }, { name = "ffmpeg-python" }, @@ -565,6 +590,7 @@ requires-dist = [ { name = "autogen-core", editable = "packages/autogen-core" }, { name = "azure-core", marker = "extra == 'azure'" }, { name = "azure-identity", marker = "extra == 'azure'" }, + { name = "diskcache", marker = "extra == 'diskcache'", specifier = ">=5.6.3" }, { name = "docker", marker = "extra == 'docker'", specifier = "~=7.0" }, { name = "ffmpeg-python", marker = "extra == 'video-surfer'" }, { name = "graphrag", marker = "extra == 'graphrag'", specifier = ">=1.0.1" }, @@ -580,6 +606,7 @@ requires-dist = [ { name = "pillow", marker = "extra == 'web-surfer'", specifier = ">=11.0.0" }, { name = "playwright", marker = "extra == 'magentic-one'", specifier = ">=1.48.0" }, { name = "playwright", marker = "extra == 'web-surfer'", specifier = ">=1.48.0" }, + { name = "redis", marker = "extra == 'redis'", specifier = ">=5.2.1" }, { name = "tiktoken", marker = "extra == 'openai'", specifier = ">=0.8.0" }, ] @@ -2358,7 +2385,7 @@ version = "0.3.14" source = { registry = "https://pypi.org/simple" } dependencies = [ { name = "aiohttp" }, - { name = "async-timeout", marker = "python_full_version < '3.11'" }, + { name = "async-timeout", version = "4.0.3", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version < '3.11'" }, { name = "langchain-core" }, { name = "langchain-text-splitters" }, { name = "langsmith" }, @@ -4913,7 +4940,8 @@ name = "redis" version = "5.2.1" source = { registry = "https://pypi.org/simple" } dependencies = [ - { name = "async-timeout", marker = "python_full_version < '3.11.3'" }, + { name = "async-timeout", version = "4.0.3", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version < '3.11'" }, + { name = "async-timeout", version = "5.0.1", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version >= '3.11' and python_full_version < '3.11.3'" }, ] sdist = { url = "https://files.pythonhosted.org/packages/47/da/d283a37303a995cd36f8b92db85135153dc4f7a8e4441aa827721b442cfb/redis-5.2.1.tar.gz", hash = "sha256:16f2e22dff21d5125e8481515e386711a34cbec50f0e44413dd7d9c060a54e0f", size = 4608355 } wheels = [ @@ -5942,7 +5970,7 @@ name = "triton" version = "3.1.0" source = { registry = "https://pypi.org/simple" } dependencies = [ - { name = "filelock", marker = "(platform_machine != 'aarch64' and sys_platform == 'linux') or (sys_platform != 'darwin' and sys_platform != 'linux')" }, + { name = "filelock" }, ] wheels = [ { url = "https://files.pythonhosted.org/packages/98/29/69aa56dc0b2eb2602b553881e34243475ea2afd9699be042316842788ff5/triton-3.1.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6b0dd10a925263abbe9fa37dcde67a5e9b2383fc269fdf59f5657cac38c5d1d8", size = 209460013 },