Fix platform.complete() to use correct LLM router API
- Import `complete` not `generate` from agentserver.llm - Pass `model` parameter required by router - Return response.content (string) not response object Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>
This commit is contained in:
parent
987c6aa214
commit
ad081d412c
1 changed files with 8 additions and 4 deletions
|
|
@ -118,9 +118,13 @@ async def complete(
|
||||||
|
|
||||||
# Make LLM call via router
|
# Make LLM call via router
|
||||||
try:
|
try:
|
||||||
from agentserver.llm import generate
|
from agentserver.llm import complete as llm_complete
|
||||||
|
|
||||||
response = await generate(
|
# Use model from kwargs or default
|
||||||
|
model = kwargs.pop("model", "grok-3-mini-beta")
|
||||||
|
|
||||||
|
response = await llm_complete(
|
||||||
|
model=model,
|
||||||
messages=messages,
|
messages=messages,
|
||||||
temperature=temperature,
|
temperature=temperature,
|
||||||
max_tokens=max_tokens,
|
max_tokens=max_tokens,
|
||||||
|
|
@ -129,10 +133,10 @@ async def complete(
|
||||||
|
|
||||||
logger.debug(
|
logger.debug(
|
||||||
f"platform.complete: agent={agent_name} thread={thread_id[:8]}... "
|
f"platform.complete: agent={agent_name} thread={thread_id[:8]}... "
|
||||||
f"messages={len(messages)} response_len={len(response)}"
|
f"messages={len(messages)} response_len={len(response.content)}"
|
||||||
)
|
)
|
||||||
|
|
||||||
return response
|
return response.content
|
||||||
|
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.error(f"LLM call failed for {agent_name}: {e}")
|
logger.error(f"LLM call failed for {agent_name}: {e}")
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue