With Memor, users can store their LLM conversation history using an intuitive and structured data format. It abstracts user prompts and model responses into a "Session", a sequence of message exchanges. In addition to the content, it includes details like decoding temperature and token count of each message. Therefore users could create comprehensive and reproducible logs of their interactions. Because of the model-agnostic design, users can begin a conversation with one LLM and switch to another keeping the context the same. For example, they might use a retrieval-augmented model (like RAG) to gather relevant context for a math problem, and then switch to a model better suited for reasoning to solve the problem based on the retrieved information presented by Memor.
Memor also lets users select, filter, and then share the specific parts of the past conversations across different models. This means users are not only able to reproduce and review previous chats through structured logs, but can also flexibly transfer the content of their conversations between LLMs. In a nutshell, Memor makes it easy and effective to manage and reuse conversations with large language models.
| PyPI Counter |
|
| Github Stars |
|
| Branch | main | dev |
| CI |
|
|
| Code Quality |
- Check Python Packaging User Guide
- Run
pip install memor==0.9
- Download Version 0.9 or Latest Source
- Run
pip install .
Memor provides Prompt, Response, and Session as abstractions by which you can save your conversation history much structured. You can set a Session object before starting a conversation, make a Prompt object from your prompt and a Response object from LLM's response. Then adding them to the created Session can keep the conversation history.
from memor import Session, Prompt, Response
from memor import RenderFormat
from mistralai import Mistral
client = Mistral(api_key="YOUR_MISTRAL_API")
session = Session()
while True:
user_input = input(">> You: ")
prompt = Prompt(message=user_input)
session.add_message(prompt) # Add user input to session
response = client.chat.complete(
model="mistral-large-latest",
messages=session.render(RenderFormat.OPENAI) # Render the whole session history
)
print("<< MistralAI:", response.choices[0].message.content)
response = Response(message=response.choices[0].message.content)
session.add_message(response) # Add model response to sessionYour conversations would carry the past interactions and LLM remembers your session's information:
>> You: Imagine you have 3 apples. You eat one of them. How many apples remain?
<< MistralAI: If you start with 3 apples and you eat one of them, you will have 2 apples remaining.
>> You: How about starting from 2 apples?
<< MistralAI: If you start with 2 apples and you eat one of them, you will have 1 apple remaining. Here's the simple math:
2 apples - 1 apple = 1 apple
In the following, we detail different abstraction levels Memor provides for the conversation artifacts.
The Prompt class is a core abstraction in Memor, representing a user prompt. The prompt can be associated with one or more responses from an LLM, with the first one being the most confident usually. It encapsulates not just the prompt text but also metadata, a template for rendering into the API endpoint, and serialization capabilities that enable saving and reusing prompts.
from memor import Prompt, Response, PresetPromptTemplate
prompt = Prompt(
message="Hello, how are you?",
responses=[
Response(message="I'm fine."),
Response(message="I'm not fine."),
],
template=PresetPromptTemplate.BASIC.PROMPT_RESPONSE_STANDARD
)
prompt.render()
# Prompt: Hello, how are you?
# Response: I'm fine.| Name | Type | Description |
|---|---|---|
message |
str |
The core prompt message content |
responses |
List[Response] |
List of associated responses |
role |
Role |
Role of the message sender (USER, SYSTEM, etc.) |
tokens |
int |
Token count |
template |
PromptTemplate | PresetPromptTemplate |
Template used to format the prompt |
file_path |
str |
Path to load a prompt from a JSON file |
init_check |
bool |
Whether to verify template rendering during initialization |
| Method | Description |
|---|---|
add_response |
Add a new response (append or insert) |
remove_response |
Remove the response at specified index |
select_response |
Mark a specific response as selected to be included in memory |
update_template |
Update the rendering template |
update_responses |
Replace all responses |
update_message |
Update the prompt text |
update_role |
Change the prompt role |
update_tokens |
Set a custom token count |
to_json / from_json |
Serialize or deserialize the prompt data |
to_dict |
Convert the object to a Python dictionary |
save / load |
Save or load prompt from file |
render |
Render the prompt in a specified format |
check_render |
Validate if the current prompt setup can render |
estimate_tokens |
Estimate the token usage for the prompt |
get_size |
Return prompt size in bytes (JSON-encoded) |
copy |
Clone the prompt |
regenerate_id |
Reset the unique identifier of the prompt |
contains_xml |
Check if the prompt contains any XML tags |
set_size_warning / reset_size_warning |
Set or reset size warning |
The Response class represents an answer or a completion generated by a model given a prompt. It encapsulates metadata such as score, temperature, model, tokens, inference time, and more. It also provides utilities for JSON serialization, rendering in multiple formats, and import/export functionality.
from memor import Response, Role, LLMModel
response = Response(
message="Sure! Here's a summary.",
score=0.94,
temperature=0.7,
model=LLMModel.GPT_4,
inference_time=0.3
)
response.render()
# Sure! Here's a summary.| Name | Type | Description |
|---|---|---|
message |
str |
The content of the response |
score |
float |
Evaluation score representing the response quality |
role |
Role |
Role of the message sender (USER, SYSTEM, etc.) |
temperature |
float |
Sampling temperature |
top_k |
int |
k in top-k sampling method |
top_p |
float |
p in top-p (nucleus) sampling |
tokens |
int |
Number of tokens in the response |
inference_time |
float |
Time spent generating the response (seconds) |
model |
LLMModel | str |
Model used |
gpu |
str |
GPU model used |
date |
datetime.datetime |
Timestamp of the creation |
file_path |
str |
Path to load a saved response |
| Method | Description |
|---|---|
update_score |
Update the response score |
update_temperature |
Set the generation temperature |
update_top_k |
Set the top-k value |
update_top_p |
Set the top-p value |
update_model |
Set the model name or enum |
update_gpu |
Set the GPU model identifier |
update_inference_time |
Set the inference time in seconds |
update_message |
Update the response message |
update_role |
Update the sender role |
update_tokens |
Set the number of tokens |
to_json / from_json |
Serialize or deserialize to/from JSON |
to_dict |
Convert the object to a Python dictionary |
save / load |
Save or load the response to/from a file |
render |
Render the response in a specific format |
check_render |
Validate if the current response setup can render |
estimate_tokens |
Estimate the token usage for the response |
get_size |
Return response size in bytes (JSON-encoded) |
copy |
Clone the response |
regenerate_id |
Reset the unique identifier of the response |
contains_xml |
Check if the response contains any XML tags |
set_size_warning / reset_size_warning |
Set or reset size warning |
The PromptTemplate class provides a structured interface for managing, storing, and customizing text prompt templates used in prompt engineering tasks. This class supports template versioning, metadata tracking, file-based persistence, and integration with preset template formats. It is a core component of the memor library, designed to facilitate reproducible and organized prompt workflows for LLMs.
from memor import Prompt, PromptTemplate
template = PromptTemplate(content="{instruction}, {prompt[message]}", custom_map={"instruction": "Hi"})
prompt = Prompt(message="How are you?", template=template)
prompt.render()
'Hi, How are you?'| Name | Type | Description |
|---|---|---|
title |
str |
The template name |
content |
str |
The template content string with placeholders |
custom_map |
Dict[str, str] |
A dictionary of custom variables used in the template |
file_path |
str |
Path to a JSON file to load the template from |
| Method | Description |
|---|---|
update_title |
Update the template title |
update_content |
Update the template content |
update_map |
Update the custom variable map |
get_size |
Return the size (in bytes) of the JSON representation |
save / load |
Save or load the template to/from a file |
to_json / from_json |
Serialize or deserialize to/from JSON |
to_dict |
Convert the template to a plain Python dictionary |
copy |
Return a shallow copy of the template instance |
Memor provides a variety of pre-defined PromptTemplates to control how prompts and responses are rendered. Each template is prefixed by an optional instruction string and includes variations for different formatting styles. Following are different variants of parameters:
INSTRUCTION1: "I'm providing you with a history of a previous conversation. Please consider this context when responding to my new question."INSTRUCTION2: "Here is the context from a prior conversation. Please learn from this information and use it to provide a thoughtful and context-aware response to my next questions."INSTRUCTION3: "I am sharing a record of a previous discussion. Use this information to provide a consistent and relevant answer to my next query."
| Template Title | Description |
|---|---|
PROMPT |
Only includes the prompt message |
RESPONSE |
Only includes the response message |
RESPONSE0 to RESPONSE3 |
Include specific responses from a list of multiple responses |
PROMPT_WITH_LABEL |
Prompt with a "Prompt: " prefix |
RESPONSE_WITH_LABEL |
Response with a "Response: " prefix |
RESPONSE0_WITH_LABEL to RESPONSE3_WITH_LABEL |
Labeled response for the i-th response |
PROMPT_RESPONSE_STANDARD |
Includes both labeled prompt and response on a single line |
PROMPT_RESPONSE_FULL |
A detailed multi-line representation including role, date, model, etc |
You can access them using:
from memor import PresetPromptTemplate
template = PresetPromptTemplate.INSTRUCTION1.PROMPT_RESPONSE_STANDARDThe Session class represents a conversation session composed of Prompt and Response messages. It supports creation, modification, saving, loading, searching, rendering, and token estimation — offering a structured way to manage LLM interaction histories. Each session tracks metadata such as title, creation/modification time, render count, and message activation (masking) status.
from memor import Session, Prompt, Response
session = Session(title="Q&A Session", messages=[
Prompt(message="What is the capital of France?"),
Response(message="The capital of France is Paris.")
])
session.add_message(Prompt(message="What is the population of Paris?"))
print(session.render())
# What is the capital of France?
# The capital of France is Paris.
# What is the population of Paris?
results = session.search("Paris")
print("Found at indices:", results)
# Found at indices: [1, 2]
tokens = session.estimate_tokens()
print("Estimated tokens:", tokens)
# Estimated tokens: 35| Parameter | Type | Description |
|---|---|---|
title |
str |
The title of the session |
messages |
List[Prompt or Response] |
The list of initial messages |
init_check |
bool |
Whether to check rendering at initialization |
file_path |
str |
The Path to a saved session file |
| Method | Description |
|---|---|
add_message |
Add a Prompt or Response to the session |
remove_message |
Remove a message by index or ID |
remove_message_by_index |
Remove a message by numeric index |
remove_message_by_id |
Remove a message by its unique ID |
update_title |
Update the title of the session |
update_messages |
Replace all messages and optionally update their status list |
update_messages_status |
Update the message status without changing the content |
clear_messages |
Remove all messages from the session |
get_message |
Retrieve a message by index, slice, or ID |
get_message_by_index |
Get a message by integer index or slice |
get_message_by_id |
Get a message by its unique ID |
enable_message |
Mark the message at the given index as active |
disable_message |
Mark the message as inactive (masked) |
mask_message |
Alias for disable_message() |
unmask_message |
Alias for enable_message() |
search |
Search for a string or regex pattern in the messages |
save / load |
Save or load the session to/from a file |
to_json / from_json |
Serialize or deserialize the session to/from JSON |
to_dict |
Return a Python dict representation of the session |
render |
Render the session in the specified format |
check_render |
Return True if the session renders without error |
get_size |
Return session size in bytes (JSON-encoded) |
copy |
Return a shallow copy of the session |
estimate_tokens |
Estimate the token count of the session content |
set_size_warning / reset_size_warning |
Set or reset size warning |
You can find more real-world usage of Memor in the examples directory.
This directory includes concise and practical Python scripts that demonstrate key features of Memor library.
Just fill an issue and describe it. We'll check it ASAP! or send an email to memor@openscilab.com.
- Please complete the issue template
You can also join our discord server
Give a ⭐️ if this project helped you!
If you do like our project and we hope that you do, can you please support us? Our project is not and is never going to be working for profit. We need the money just so we can continue doing what we do ;-) .

