-
Notifications
You must be signed in to change notification settings - Fork 266
Description
I’m exploring a research direction that extends optillm from per-request optimization to cross-session decision reuse via persistent agent memory (i.e., turning past trajectories into reusable policies, rather than a policy-search setup like OpenEvolve or LLM4AD).
I wish to check whether the current architecture supports the following:
1. Persistent State
I may need a mechanism to accumulate knowledge across sessions (e.g., successful traces, failure→fix mappings, routing policies).
- Is there any built-in shared/global state?
- If not, is an external memory layer (KV / vector DB) the intended pattern?
- Where should it integrate: strategy layer, middleware, or outside the proxy?
2. Feedback Loops
My setup depends on post-hoc signals (e.g., execution results, tests, errors) to inform future decisions.
Is there any support (or recommended pattern) for closing the loop:
response → execution → feedback → memory update → future decisions?
Are there hooks for post-response feedback ingestion?
3. Hybrid Decision Layer
I’m considering a tiered setup:
- fast path: local logic / small model for known patterns
- slow path: LLM via optillm
Would this be better implemented inside optillm as a custom strategy,
or as an external pre-routing layer before requests enter the proxy?
Any expected bottlenecks with a high-frequency local decision layer?
4. Scope Alignment
Does adding persistent agent memory and decision reuse across sessions align with optillm’s design direction, or is it outside the intended scope?
If aligned, I'd be interested in contributing a reusable “agent memory” extension. Thanks!