@@ -209,7 +209,6 @@ async def get_working_memory(
209
209
session_id : str ,
210
210
user_id : str | None = None ,
211
211
namespace : str | None = None ,
212
- window_size : int | None = None ,
213
212
model_name : ModelNameLiteral | None = None ,
214
213
context_window_max : int | None = None ,
215
214
) -> WorkingMemoryResponse :
@@ -220,7 +219,6 @@ async def get_working_memory(
220
219
session_id: The session ID to retrieve working memory for
221
220
user_id: The user ID to retrieve working memory for
222
221
namespace: Optional namespace for the session
223
- window_size: Optional number of messages to include
224
222
model_name: Optional model name to determine context window size
225
223
context_window_max: Optional direct specification of context window tokens
226
224
@@ -241,9 +239,6 @@ async def get_working_memory(
241
239
elif self .config .default_namespace is not None :
242
240
params ["namespace" ] = self .config .default_namespace
243
241
244
- if window_size is not None :
245
- params ["window_size" ] = str (window_size )
246
-
247
242
# Use provided model_name or fall back to config default
248
243
effective_model_name = model_name or self .config .default_model_name
249
244
if effective_model_name is not None :
@@ -2139,7 +2134,6 @@ async def memory_prompt(
2139
2134
query : str ,
2140
2135
session_id : str | None = None ,
2141
2136
namespace : str | None = None ,
2142
- window_size : int | None = None ,
2143
2137
model_name : str | None = None ,
2144
2138
context_window_max : int | None = None ,
2145
2139
long_term_search : dict [str , Any ] | None = None ,
@@ -2154,7 +2148,6 @@ async def memory_prompt(
2154
2148
query: The input text to find relevant context for
2155
2149
session_id: Optional session ID to include session messages
2156
2150
namespace: Optional namespace for the session
2157
- window_size: Optional number of messages to include
2158
2151
model_name: Optional model name to determine context window size
2159
2152
context_window_max: Optional direct specification of context window tokens
2160
2153
long_term_search: Optional search parameters for long-term memory
@@ -2169,7 +2162,6 @@ async def memory_prompt(
2169
2162
prompt = await client.memory_prompt(
2170
2163
query="What are my UI preferences?",
2171
2164
session_id="current_session",
2172
- window_size=10,
2173
2165
long_term_search={
2174
2166
"topics": {"any": ["preferences", "ui"]},
2175
2167
"limit": 5
@@ -2190,8 +2182,6 @@ async def memory_prompt(
2190
2182
session_params ["namespace" ] = namespace
2191
2183
elif self .config .default_namespace is not None :
2192
2184
session_params ["namespace" ] = self .config .default_namespace
2193
- if window_size is not None :
2194
- session_params ["window_size" ] = str (window_size )
2195
2185
# Use provided model_name or fall back to config default
2196
2186
effective_model_name = model_name or self .config .default_model_name
2197
2187
if effective_model_name is not None :
0 commit comments