Skip to content

Persistent Agent Memory & Cross-Session Decision Reuse in optillm #299

@yuxuan-z19

Description

@yuxuan-z19

I’m exploring a research direction that extends optillm from per-request optimization to cross-session decision reuse via persistent agent memory (i.e., turning past trajectories into reusable policies, rather than a policy-search setup like OpenEvolve or LLM4AD).

I wish to check whether the current architecture supports the following:

1. Persistent State

I may need a mechanism to accumulate knowledge across sessions (e.g., successful traces, failure→fix mappings, routing policies).

  • Is there any built-in shared/global state?
  • If not, is an external memory layer (KV / vector DB) the intended pattern?
  • Where should it integrate: strategy layer, middleware, or outside the proxy?

2. Feedback Loops

My setup depends on post-hoc signals (e.g., execution results, tests, errors) to inform future decisions.

Is there any support (or recommended pattern) for closing the loop:
response → execution → feedback → memory update → future decisions?

Are there hooks for post-response feedback ingestion?

3. Hybrid Decision Layer

I’m considering a tiered setup:

  • fast path: local logic / small model for known patterns
  • slow path: LLM via optillm

Would this be better implemented inside optillm as a custom strategy,
or as an external pre-routing layer before requests enter the proxy?

Any expected bottlenecks with a high-frequency local decision layer?

4. Scope Alignment

Does adding persistent agent memory and decision reuse across sessions align with optillm’s design direction, or is it outside the intended scope?


If aligned, I'd be interested in contributing a reusable “agent memory” extension. Thanks!

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions