3 steps to fix app memory leaks

How to detect if your Streamlit app leaks memory and identify faulty code

Posted in Announcement, April 14 2022

Does your Streamlit app crash after long use or constantly runs out of memory? Chances are, it has a memory leak.

In this post, you’ll learn how to find and fix memory leaks in three simple steps:

  1. Identify the memory leak
  2. Identify the leaking objects
  3. Identify the code that's allocating the leaking objects

NOTE: Streamlit is used in a variety of settings, from short-lived research projects to live company dashboards. This post is primarily aimed at developers deploying Streamlit apps with very long uptimes. We at Streamlit strive to keep the Streamlit framework itself free of leaks. Developers with long-running apps also need to be mindful of memory usage over time. Even a tiny memory leak can compound. Eventually, even the beefiest machine will fall over if memory usage isn’t kept in check.

Let’s dive right in.

1. Identify the memory leak

A leak happens when your app acquires memory resources and never releases them. It just consumes more and more memory. Since it's a finite resource, eventually it gets exhausted, and the OS terminates your application.

As an example, we'll use a Streamlit library memory leak that we investigated and patched as part of this pull request. After we found the faulty code and released a fix, apps started using one-tenth to one-twentieth of the memory used before!

To find out if your app leaks memory, use a memory profiler (we used mprof ):

# make sure mprof is installed
pip install memory_profiler
# find the pid of the Streamlit app (the next number after the user)
ps aux | grep "streamlit run" | grep -v grep
# start profiling the memory of the app
mprof run --attach <pid>

Once the profiler is attached, use your app to simulate a higher load (multiple sessions, more complex operations, etc.). After a few minutes, plot a memory graph:

mprof plot
# if the above gives you an error related to GUI you can try fixing it with
pip install PyQt5

If the memory usage doesn't plateau, your app is leaking memory:

2. Identify leaking objects

Now identify which objects are allocated and never released. Use the tracemalloc Python library. Take snapshots between executions (after forcing the garbage collection).

If an object persists in a snapshot, it means it's never collected:

import tracemalloc, json
import streamlit as st
import gc
def init_tracking_object():
  return {
    "runs": 0,
    "tracebacks": {}
_TRACES = init_tracking_object()
def traceback_exclude_filter(patterns, tracebackList):
    Returns False if any provided pattern exists in the filename of the traceback,
    Returns True otherwise.
    for t in tracebackList:
        for p in patterns:
            if p in t.filename:
                return False
        return True
def traceback_include_filter(patterns, tracebackList):
    Returns True if any provided pattern exists in the filename of the traceback,
    Returns False otherwise.
    for t in tracebackList:
        for p in patterns:
            if p in t.filename:
                return True
    return False
def check_for_leaks(diff):
    Checks if the same traceback appears consistently after multiple runs.
    diff - The object returned by tracemalloc#snapshot.compare_to
    _TRACES["runs"] = _TRACES["runs"] + 1
    tracebacks = set()
    for sd in diff:
        for t in sd.traceback:
    if "tracebacks" not in _TRACES or len(_TRACES["tracebacks"]) == 0:
        for t in tracebacks:
            _TRACES["tracebacks"][t] = 1
        oldTracebacks = _TRACES["tracebacks"].keys()
        intersection = tracebacks.intersection(oldTracebacks)
        evictions = set()
        for t in _TRACES["tracebacks"]:
            if t not in intersection:
                _TRACES["tracebacks"][t] = _TRACES["tracebacks"][t] + 1
        for t in evictions:
            del _TRACES["tracebacks"][t]
    if _TRACES["runs"] > 1:
        st.write(f'After {_TRACES["runs"]} runs the following traces were collected.')
        prettyPrint = {}
        for t in _TRACES["tracebacks"]:
            prettyPrint[str(t)] = _TRACES["tracebacks"][t]
        st.write(json.dumps(prettyPrint, sort_keys=True, indent=4))
def compare_snapshots():
    Compares two consecutive snapshots and tracks if the same traceback can be found
    in the diff. If a traceback consistently appears during runs, it's a good indicator
    for a memory leak.
    snapshot = tracemalloc.take_snapshot()
    if "snapshot" in _TRACES:
        diff = snapshot.compare_to(_TRACES["snapshot"], "lineno")
        diff = [d for d in diff if
                d.count_diff > 0 and traceback_exclude_filter(["tornado"], d.traceback)
                and traceback_include_filter(["streamlit"], d.traceback)
    _TRACES["snapshot"] = snapshot

NOTE: Call compare_snapshots() always after the gc collection is forced, to make sure you track only the objects for which memory can’t be reclaimed.

Run the above in the Streamlit app, version lower than 1.0, and you'll get the following output:

    "<attrs generated init streamlit.state.session_state.SessionState>:17": 22,

SessionState is the object that leaked.

Now let's identify the part of the code that allocates the SessionState and never releases it.

3. Identify the code that's allocating the leaking objects

Track which object is not releasing the memory using the objgraph library. To install it, run pip install objgraph. Track the holder of the SessionState after the gc collection is forced:

import gc
import objgraph
for o in gc.get_objects():
    if 'session_state.SessionState' in str(type(o)) and o is not st.session_state:
        filename = f'/tmp/session_state_{hex(id(o))}.png'
        st.write("SessionState reference retained by: ", type(o))

In our case, when we run on Streamlit versions below 1.0.0, you can see that the SessionState object is held by the Signal class:

The allocation of SessionState happens in the streamlit.config module. The object holding the resource is Signal. It holds the resources that need to be released (check out our pull request).

After we applied the fix and updated the app, the memory usage plateaued:

Wrapping up

Now you know how to detect if your Streamlit app is leaking memory and how to fix it! If you have any questions, please let us know in the comments below or on the forum. We'd be happy to help! ❤️

This is a companion discussion topic for the original entry at https://blog.streamlit.io/3-steps-to-fix-app-memory-leaks/