Skip to content

Reproducible Structured Memory for LLMs

License

openscilab/memor

Memor Logo

Memor: Reproducible Structured Memory for LLMs


PyPI version built with Python3 GitHub repo size Discord Channel

Overview

With Memor, users can store their LLM conversation history using an intuitive and structured data format. It abstracts user prompts and model responses into a "Session", a sequence of message exchanges. In addition to the content, it includes details like decoding temperature and token count of each message. Therefore users could create comprehensive and reproducible logs of their interactions. Because of the model-agnostic design, users can begin a conversation with one LLM and switch to another keeping the context the same. For example, they might use a retrieval-augmented model (like RAG) to gather relevant context for a math problem, and then switch to a model better suited for reasoning to solve the problem based on the retrieved information presented by Memor.

Memor also lets users select, filter, and then share the specific parts of the past conversations across different models. This means users are not only able to reproduce and review previous chats through structured logs, but can also flexibly transfer the content of their conversations between LLMs. In a nutshell, Memor makes it easy and effective to manage and reuse conversations with large language models.

PyPI Counter
Github Stars
Branch main dev
CI
Code Quality CodeFactor

Installation

PyPI

Source code

Usage

Memor provides Prompt, Response, and Session as abstractions by which you can save your conversation history much structured. You can set a Session object before starting a conversation, make a Prompt object from your prompt and a Response object from LLM's response. Then adding them to the created Session can keep the conversation history.

from memor import Session, Prompt, Response
from memor import RenderFormat
from mistralai import Mistral

client = Mistral(api_key="YOUR_MISTRAL_API")
session = Session()
while True:
    user_input = input(">> You: ")
    prompt = Prompt(message=user_input)
    session.add_message(prompt) # Add user input to session
    response = client.chat.complete(
        model="mistral-large-latest",
        messages=session.render(RenderFormat.OPENAI)  # Render the whole session history
    )
    print("<< MistralAI:", response.choices[0].message.content)
    response = Response(message=response.choices[0].message.content)
    session.add_message(response) # Add model response to session

Your conversations would carry the past interactions and LLM remembers your session's information:

>> You: Imagine you have 3 apples. You eat one of them. How many apples remain?
<< MistralAI: If you start with 3 apples and you eat one of them, you will have 2 apples remaining.
>> You: How about starting from 2 apples?
<< MistralAI: If you start with 2 apples and you eat one of them, you will have 1 apple remaining. Here's the simple math:
2 apples - 1 apple = 1 apple

In the following, we detail different abstraction levels Memor provides for the conversation artifacts.

Prompt

The Prompt class is a core abstraction in Memor, representing a user prompt. The prompt can be associated with one or more responses from an LLM, with the first one being the most confident usually. It encapsulates not just the prompt text but also metadata, a template for rendering into the API endpoint, and serialization capabilities that enable saving and reusing prompts.

from memor import Prompt, Response, PresetPromptTemplate
prompt = Prompt(
    message="Hello, how are you?",
    responses=[
        Response(message="I'm fine."),
        Response(message="I'm not fine."),
    ],
    template=PresetPromptTemplate.BASIC.PROMPT_RESPONSE_STANDARD
)
prompt.render()
# Prompt: Hello, how are you?
# Response: I'm fine.

Parameters

Name Type Description
message str The core prompt message content
responses List[Response] List of associated responses
role Role Role of the message sender (USER, SYSTEM, etc.)
tokens int Token count
template PromptTemplate | PresetPromptTemplate Template used to format the prompt
file_path str Path to load a prompt from a JSON file
init_check bool Whether to verify template rendering during initialization

Methods

Method Description
add_response Add a new response (append or insert)
remove_response Remove the response at specified index
select_response Mark a specific response as selected to be included in memory
update_template Update the rendering template
update_responses Replace all responses
update_message Update the prompt text
update_role Change the prompt role
update_tokens Set a custom token count
to_json / from_json Serialize or deserialize the prompt data
to_dict Convert the object to a Python dictionary
save / load Save or load prompt from file
render Render the prompt in a specified format
check_render Validate if the current prompt setup can render
estimate_tokens Estimate the token usage for the prompt
get_size Return prompt size in bytes (JSON-encoded)
copy Clone the prompt
regenerate_id Reset the unique identifier of the prompt
contains_xml Check if the prompt contains any XML tags
set_size_warning / reset_size_warning Set or reset size warning

Response

The Response class represents an answer or a completion generated by a model given a prompt. It encapsulates metadata such as score, temperature, model, tokens, inference time, and more. It also provides utilities for JSON serialization, rendering in multiple formats, and import/export functionality.

from memor import Response, Role, LLMModel
response = Response(
    message="Sure! Here's a summary.",
    score=0.94,
    temperature=0.7,
    model=LLMModel.GPT_4,
    inference_time=0.3
)
response.render()
# Sure! Here's a summary.

Parameters

Name Type Description
message str The content of the response
score float Evaluation score representing the response quality
role Role Role of the message sender (USER, SYSTEM, etc.)
temperature float Sampling temperature
top_k int k in top-k sampling method
top_p float p in top-p (nucleus) sampling
tokens int Number of tokens in the response
inference_time float Time spent generating the response (seconds)
model LLMModel | str Model used
gpu str GPU model used
date datetime.datetime Timestamp of the creation
file_path str Path to load a saved response

Methods

Method Description
update_score Update the response score
update_temperature Set the generation temperature
update_top_k Set the top-k value
update_top_p Set the top-p value
update_model Set the model name or enum
update_gpu Set the GPU model identifier
update_inference_time Set the inference time in seconds
update_message Update the response message
update_role Update the sender role
update_tokens Set the number of tokens
to_json / from_json Serialize or deserialize to/from JSON
to_dict Convert the object to a Python dictionary
save / load Save or load the response to/from a file
render Render the response in a specific format
check_render Validate if the current response setup can render
estimate_tokens Estimate the token usage for the response
get_size Return response size in bytes (JSON-encoded)
copy Clone the response
regenerate_id Reset the unique identifier of the response
contains_xml Check if the response contains any XML tags
set_size_warning / reset_size_warning Set or reset size warning

Prompt Templates

The PromptTemplate class provides a structured interface for managing, storing, and customizing text prompt templates used in prompt engineering tasks. This class supports template versioning, metadata tracking, file-based persistence, and integration with preset template formats. It is a core component of the memor library, designed to facilitate reproducible and organized prompt workflows for LLMs.

from memor import Prompt, PromptTemplate
template = PromptTemplate(content="{instruction}, {prompt[message]}", custom_map={"instruction": "Hi"})
prompt = Prompt(message="How are you?", template=template)
prompt.render()
'Hi, How are you?'

Parameters

Name Type Description
title str The template name
content str The template content string with placeholders
custom_map Dict[str, str] A dictionary of custom variables used in the template
file_path str Path to a JSON file to load the template from

Methods

Method Description
update_title Update the template title
update_content Update the template content
update_map Update the custom variable map
get_size Return the size (in bytes) of the JSON representation
save / load Save or load the template to/from a file
to_json / from_json Serialize or deserialize to/from JSON
to_dict Convert the template to a plain Python dictionary
copy Return a shallow copy of the template instance

Preset Templates

Memor provides a variety of pre-defined PromptTemplates to control how prompts and responses are rendered. Each template is prefixed by an optional instruction string and includes variations for different formatting styles. Following are different variants of parameters:

  • INSTRUCTION1: "I'm providing you with a history of a previous conversation. Please consider this context when responding to my new question."
  • INSTRUCTION2: "Here is the context from a prior conversation. Please learn from this information and use it to provide a thoughtful and context-aware response to my next questions."
  • INSTRUCTION3: "I am sharing a record of a previous discussion. Use this information to provide a consistent and relevant answer to my next query."
Template Title Description
PROMPT Only includes the prompt message
RESPONSE Only includes the response message
RESPONSE0 to RESPONSE3 Include specific responses from a list of multiple responses
PROMPT_WITH_LABEL Prompt with a "Prompt: " prefix
RESPONSE_WITH_LABEL Response with a "Response: " prefix
RESPONSE0_WITH_LABEL to RESPONSE3_WITH_LABEL Labeled response for the i-th response
PROMPT_RESPONSE_STANDARD Includes both labeled prompt and response on a single line
PROMPT_RESPONSE_FULL A detailed multi-line representation including role, date, model, etc

You can access them using:

from memor import PresetPromptTemplate
template = PresetPromptTemplate.INSTRUCTION1.PROMPT_RESPONSE_STANDARD

Session

The Session class represents a conversation session composed of Prompt and Response messages. It supports creation, modification, saving, loading, searching, rendering, and token estimation — offering a structured way to manage LLM interaction histories. Each session tracks metadata such as title, creation/modification time, render count, and message activation (masking) status.

from memor import Session, Prompt, Response
session = Session(title="Q&A Session", messages=[
            Prompt(message="What is the capital of France?"),
            Response(message="The capital of France is Paris.")
            ])
session.add_message(Prompt(message="What is the population of Paris?"))
print(session.render())
# What is the capital of France?
# The capital of France is Paris.
# What is the population of Paris?

results = session.search("Paris")
print("Found at indices:", results)
# Found at indices: [1, 2]

tokens = session.estimate_tokens()
print("Estimated tokens:", tokens)
# Estimated tokens: 35

Parameters

Parameter Type Description
title str The title of the session
messages List[Prompt or Response] The list of initial messages
init_check bool Whether to check rendering at initialization
file_path str The Path to a saved session file

Methods

Method Description
add_message Add a Prompt or Response to the session
remove_message Remove a message by index or ID
remove_message_by_index Remove a message by numeric index
remove_message_by_id Remove a message by its unique ID
update_title Update the title of the session
update_messages Replace all messages and optionally update their status list
update_messages_status Update the message status without changing the content
clear_messages Remove all messages from the session
get_message Retrieve a message by index, slice, or ID
get_message_by_index Get a message by integer index or slice
get_message_by_id Get a message by its unique ID
enable_message Mark the message at the given index as active
disable_message Mark the message as inactive (masked)
mask_message Alias for disable_message()
unmask_message Alias for enable_message()
search Search for a string or regex pattern in the messages
save / load Save or load the session to/from a file
to_json / from_json Serialize or deserialize the session to/from JSON
to_dict Return a Python dict representation of the session
render Render the session in the specified format
check_render Return True if the session renders without error
get_size Return session size in bytes (JSON-encoded)
copy Return a shallow copy of the session
estimate_tokens Estimate the token count of the session content
set_size_warning / reset_size_warning Set or reset size warning

Examples

You can find more real-world usage of Memor in the examples directory. This directory includes concise and practical Python scripts that demonstrate key features of Memor library.

Issues & bug reports

Just fill an issue and describe it. We'll check it ASAP! or send an email to memor@openscilab.com.

  • Please complete the issue template

You can also join our discord server

Discord Channel

Show your support

Star this repo

Give a ⭐️ if this project helped you!

Donate to our project

If you do like our project and we hope that you do, can you please support us? Our project is not and is never going to be working for profit. We need the money just so we can continue doing what we do ;-) .

Memor Donation