Fixing Memory Exhaustion Bugs in My Golang Web App

Earlier this year, I created an open-source app called PicoShare. It’s a simple Golang web app for sharing files. I use it to send files that are too large to be email attachments, but I don’t want the recipient to deal with Dropbox or Google Drive.

Animated demo of uploading a video file to PicoShare and streaming it in another browser window

A few months ago, I started seeing my PicoShare server die every few days. When I checked the logs, I saw an out of memory error:

Out of memory: Killed process 515 (picoshare)

I didn’t have time to debug the crash, so I just increased the server’s memory from 512 MB to 1 GB. And then I kept seeing crashes, so I increased it again to 2 GB.

It’s unsatisfying to fix a crash by just throwing more RAM at the problems, so for the past two weeks, I’ve been debugging the crashes and sharing my progress on Twitter.

At this point, I’ve fixed all the issues that were causing crashes and learned some useful lessons along the way about Go, SQLite, and debugging.

If you want to see the story as it unfolded, check out the Twitter thread. If you’d like a cleaned up, condensed version of what I learned, read on.

Contents ๐Ÿ”—︎

Preface: I use SQLite strangely ๐Ÿ”—︎

One of the strange architecture decisions I made with PicoShare was to store all file data in SQLite. This is an unusual choice, as web applications typically store file uploads directly on the filesystem, not in a database. This especially true when the uploads can be arbitrarily large.

The advantage of writing file data to SQLite is that all of PicoShare’s application state is in a single database. That’s not so special itself, but I designed PicoShare to integrate with Litestream, a tool for replicating SQLite databases to cloud storage. Litestream essentially gives PicoShare backup and restore “for free.” I can completely blow away a server and then redeploy it anywhere (even another cloud hosting provider), and PicoShare will wake up with the exact same state, serving all the same files.

The debugging process ๐Ÿ”—︎

Reproducing the error ๐Ÿ”—︎

I only saw PicoShare crash every few days, so my first step was to find a way to force the crash more quickly.

I managed to reproduce the error by deploying PicoShare on a Fly instance with only 256 MB of RAM and then uploading large files. I used high-resolution versions of the short film Big Buck Bunny with sizes ranging from 269 MB to 618 MB.

Still from Big Buck Bunny short film

I used the short film Big Buck Bunny as my test file as it was large enough to test big uploads.

Uploading two copies of the 618 MB version in parallel consistently caused PicoShare to die with an out of memory error within a minute or so.

Using profiling tools to identify RAM bloat ๐Ÿ”—︎

The first break in the investigation came from Ben Johnson, the author of Litestream and a recent addition to the Fly team. Ben created a detailed pull request explaining how a single line of code was causing PicoShare to consume large amounts of RAM.

Ben has extensive experience with profiling, so he was able to reproduce the issue by creating a new unit test and profiling memory after the test. Ben later showed me an easier way to get the same information, so I’m going to show that, but you can find Ben’s original technique in his pull request.

It turns out the Go standard library comes with a magical tool for debugging issues in web applications. All you have to do is add this line to your imports:

_ "net/http/pprof"

Now, when you run your app, there will be a /debug/pprof/ route with lots of useful debugging information.

Debug interface at http://ps:4001/debug/pprof

I was surprised at how easy this was to add. There’s a lot of interesting data in this web interface, but the one that I used was heap. To use it, I uploaded a large file to PicoShare and then ran this the following command:

go tool pprof \
  -http=:8081 \
  -alloc_space \
  call_tree \
  http://localhost:4001/debug/pprof/heap

That popped up a web interface and rendered this graph:

Graph showing all memory allocations

At the bottom, you can see a large red block labeled bytes makeSlice 63.99 MB, meaning that 64 MB of PicoShare’s allocated RAM came from Go’s makeSlice function.

makeSlice is in the Go standard library, not my code. To find what code in PicoShare caused this memory allocation, I traced up the graph until I found a PicoShare function:

Zoom in on graph showing call from fileFromRequest to ParseMultipartForm

The last PicoShare function in this chain is handlers.fileFromRequest, which calls the Go standard library function *Request.ParseMultipartForm. That function is responsible for parsing multipart HTTP data, which is how PicoShare accepts file uploads.

ParseMultipartForm accepts a maxMemory parameter, documented as follows:

The whole request body is parsed and up to a total of maxMemory bytes of its file parts are stored in memory, with the remainder stored on disk in temporary files.

PicoShare’s call looked like this:

r.ParseMultipartForm(32 << 20) // 32 MB

Even though we were specifying a limit of 32 MB, Go was allocating 64 MB of RAM.

Ben tried reducing the maxMemory parameter to 1 << 20 (1 MB), and the RAM usage from ParseMultipartForm dropped to only 2.5 MB:

Graph showing 2572.91kB in makeSlice after the fix

This was a huge reduction in memory, so I thought for sure Ben had solved it.

Unfortunately, I deployed a test version with Ben’s fix, and it still crashed.

After Ben’s fix, PicoShare could withstand more load before crashing, so it did seem to make a difference. Still, when I uploaded three large files in parallel, the server died with the same out of memory error.

Freeing resources after calling ParseMultipartForm ๐Ÿ”—︎

From Googling, I discovered another gotcha with ParseMultipartForm.

The documentation doesn’t warn the reader, but the caller is responsible for calling r.MultipartForm.RemoveAll() to free the resources Go allocated during ParseMultipartForm. So, I was leaking memory every time I called ParseMultipartForm.

Update (2022-08-11): Damien Neil points out in the comments that Go should clean up these resources automatically. On Go’s HTTP/1 implementation, it automatically cleans up resources, and Damien has submitted a bugfix to make Go’s HTTP/2 implementation behave consistently.

To fix the leak, I rewrote my code to clean up the multipart resources:

multipartMaxMemory := 1 << 20 // 1 MiB
if err := r.ParseMultipartForm(multipartMaxMemory); err != nil {
  return err
}
// Free form resources before returning from function.
defer func() {
  if err := r.MultipartForm.RemoveAll(); err != nil {
    log.Printf("failed to free multipart form resources: %v", err)
  }
}()

This fix looked promising, as I saw huge reductions in RAM usage on Fly after freeing resources explicitly:

Fly graph showing memory increase when I call ParseMultipartForm and decrease when I call r.MultipartForm.RemoveAll

Sadly, even with this fix, the crashes continued.

Optimizing downloads ๐Ÿ”—︎

At this point, Dan Wilhelm started following the Twitter thread. Even though he’s never used Go a day in his life, he rolled up his sleeves and started experimenting with the code on his development machine.

Dan noticed that RAM usage shot up when he downloaded files. That was strange, as downloads shouldn’t consume much RAM. Parsing a multipart form is complex, so there are many factors that could be bloating RAM, but serving a file is pretty straightforward.

PicoShare stores all of its file data in SQLite in 328 KB chunks. That shouldn’t be RAM intensive because we should be able to just read some chunks into RAM, send them to the client, then free the memory.

Dan found a bug in the code responsible for reading PicoShare’s file data from the database. See if you can spot it:

func (fr *fileReader) populateBuffer() error {
  if fr.offset == int64(fr.fileLength) {
    return io.EOF
  }

  startChunk := fr.offset / int64(fr.chunkSize)
  stmt, err := fr.db.Prepare(`
      SELECT
        chunk
      FROM
        entries_data
      WHERE
        id=? AND
        chunk_index>=?
      ORDER BY
        chunk_index ASC
      `)
  if err != nil {
    log.Printf("reading chunk failed: %v", err)
    return err
  }
  defer stmt.Close()

  var chunk []byte
  err = stmt.QueryRow(fr.entryID, startChunk).Scan(&chunk)
  if err != nil {
    return err
  }

  // Move the start index to the position in the chunk we want to read.
  readStart := fr.offset % int64(fr.chunkSize)

  fr.buf = bytes.NewBuffer(chunk[readStart:])
  fr.offset += int64(len(chunk)) - readStart

  return nil
}

The bug is in the WHERE clause of the SQL query:

WHERE
    id=? AND
    chunk_index>=?

The query is only supposed to retrieve a single chunk of file data. Instead, it reads the target chunk and everything after.

The fix was simply to change the >= to =:

WHERE
    id=? AND
    chunk_index=?
Sidenote: Reading this code, I also realized I was using prepared statements when I didn’t need to, though I don’t think this affected RAM.

Dan’s change was on the download side, so I didn’t expect it to fix crashes I saw during upload. And indeed it didn’t, but there was a drastic performance improvement in serving downloads. Especially with streaming content like videos or audio, PicoShare was much more responsive when I jumped to different positions in the file.

Removing SQLite transactions ๐Ÿ”—︎

Within the Twitter thread, several people suggested that PicoShare’s SQLite transactions were likely bloating RAM.

When PicoShare wrote file data into SQLite, I did it within a transaction. The purpose was to ensure the database was always in a consistent state.

By using transactions, SQLite guaranteed that I would never reach a state where only part of the file was in the database if some writes failed. It also ensured that I couldn’t accidentally write the file metadata without writing the file contents and vice-versa.

Some Googling indicated that large SQLite transactions can be a source of memory bloat, so I figured it was worth trying. I tried committing changes to SQLite immediately instead of using transactions, but it still bloated RAM. It didn’t seem like transactions were making any difference.

RAM bloat is fine, but crashes are not ๐Ÿ”—︎

At this point, I was measuring RAM usage from three different angles that all disagreed with each other:

  • Go’s debug metrics for how much memory it had allocated
  • htop within in the VM
  • Fly’s RAM metrics from the VM host

In particular, Fly’s metrics would frequently show RAM maxed out when Go and htop showed barely any usage. It was frustrating to debug because the further I drilled down, the further RAM measurements diverged from the crash behavior I was observing.

The game-changing insight came from Andrew Ayer, who pointed out that RAM bloat was likely a red herring:

Kurt Mackey, Fly’s CEO, popped into the thread to confirm Andrew’s hypothesis:

This is the page cache usage for your -dbg app over the last 3 hours. Page cache shows as usage in our UI, but it's almost the same as free memory. It should be evicted when there's memory pressure.

So, Fly’s memory metrics included the page cache, but the VM should reclaim that RAM if running applications needed it.

This was a huge realization. Because of the difficulty of causing out of memory crashes, I’d used RAM bloat as an approximation for the crashes. But RAM bloat is fine as long as the VM still has enough memory to keep my processes running.

I had to reevaluate everything now. When I dismissed other fixes, had it been because they caused harmless RAM bloat? Or did I observe actual crashes?

Re-examining SQLite transactions ๐Ÿ”—︎

Given what Andrew Ayer said about RAM bloat, I revisited PicoShare’s SQLite transactions. When I tried the implementation with no transactions, did I see crashes or just RAM bloat? I couldn’t remember.

I tried running the transactionless implementation again. Sure enough, RAM bloated but PicoShare kept running. I uploaded three 618 MB files in parallel, and every upload succeeded with PicoShare continuing to serve HTTP requests.

Screenshot of three parallel PicoShare uploads succeeding without crashes

It worked! I’d finally gotten to the bottom of the performance issues.

Or so I thought…

I left my server running overnight, and when I checked it the next morning, it had failed with the same out of memory crash.

Screenshot of log showing 'Process appears to have been OOM killed!'

Eliminating SQLite vacuuming ๐Ÿ”—︎

I immediately suspected that the overnight crash was related to the SQLite VACUUM command, which compresses the database file to reclaim unused disk space.

Nobody was using the PicoShare server when it crashed, but it did line up with PicoShare’s scheduled database maintenance. Every seven hours, PicoShare removes expired entries from the database and performs a VACUUM to reclaim unused disk space.

I tested running the VACUUM command on my server and saw that it did indeed reduce the size of my main .db file, but it was increasing the size of the SQLite write-ahead log.

store.db-wal increasing in size by 310 MB after each call to sqlite3 /data/store.db 'VACUUM'

At this point, Ben asked me why I need to VACUUM at all:

Yeah, why am I doing that?

When I first launched PicoShare, users complained that it wasn’t giving back disk space after they deleted files. It didn’t affect me because I run PicoShare on a Fly VM with a fixed disk volume, so it doesn’t matter how much of the disk I use. But it was easy enough to add in the periodic VACUUM, so I did.

After thinking it over, I decided to change PicoShare’s behavior so that VACUUM is off by default, but users can enable it with a command-line flag.

dbPath := flag.String("db", "data/store.db", "path to database")
vacuumDb := flag.Bool("vacuum", false, "vacuum database periodically to reclaim disk space")
flag.Parse()

Success: PicoShare running on 256 MB of RAM ๐Ÿ”—︎

With VACUUM disabled by default and my other performance fixes in place, PicoShare was finally running stable with low RAM.

I ran PicoShare for 24 hours without any crashes on a Fly VM with just 256 MB of RAM.

Fly dashboard showing PicoShare has 256 RAM
Uptime checks showing 100% availability

100% uptime over the last 24 hours

Other lessons learned ๐Ÿ”—︎

In addition to what I learned above, I also picked up some useful side lessons in this debugging quest.

Optimize your build-test loop ๐Ÿ”—︎

One thing I wish I’d done earlier was to optimize my build-test loop. To test any hypothesis, my process was:

  1. Deploy my changes to Fly (2-3 minutes)
  2. Upload a large file (1-2 minutes)
  3. Wait for Fly’s RAM metrics to catch up (30-60 seconds)

So, that’s up to six minutes just to test any change, and each step required manual work. And that’s not even counting time to write the code changes.

I had originally tried running PicoShare in a Docker container with limited RAM, but it never crashed.

RAM_LIMIT="64m"
PORT=3001
PS_SHARED_SECRET="somesecretpass"

docker run \
  --memory "${RAM_LIMIT}" \
  --env "PORT=${PORT}" \
  --env "PS_SHARED_SECRET=${PS_SHARED_SECRET}" \
  --publish "${PORT}:${PORT}/tcp" \
  --name picoshare \
  mtlynch/picoshare:1.1.7
$ docker stats
CONTAINER ID   NAME        CPU %     MEM USAGE / LIMIT   MEM %     NET I/O         BLOCK I/O        PIDS
1ababd398113   picoshare   3.82%     63.68MiB / 64MiB    99.50%    278MB / 208MB   6.09GB / 7.1MB   21

I still don’t understand why PicoShare behaves differently under Docker than a real VM, but my best guess is that Docker isn’t really restricting RAM usage as tightly as a VM.

When Dan Wilhelm reported how much progress he’d made by running PicoShare locally and observing RAM usage, it made me realize how much time I was wasting by deploying to Fly for every change. I tried running PicoShare on my home VM server, but it never crashed or bloated RAM the way it did on Fly.

What eventually worked was creating my own development environment on Fly. I wrote a Dockerfile that had the PicoShare source and some dev tools and deployed that to Fly. From there, I could use fly ssh console to open a shell on my server and then test code changes quickly.

It still wasn’t super fast because there’s about 30 seconds of latency before Fly’s RAM metrics update, but it was a big improvement over having to deploy each change from scratch.

Use descriptive git branches and commit messages to record notes ๐Ÿ”—︎

One useful technique I discovered during this investigation was to test each hypothesis in its own git branch and then record the results with a commit message:

Branch repro-no-tx has commit name 'Working - no OOM crashes with 3x parallel 600 MB uploads'

With so many different hypotheses flying around, it was difficult to remember what state the code was in when I tested each idea. For example, at one point, I was seeing crashes due to a new bug I had introduced while debugging:

Having a record of what state the code was in and what I did to test it helped me organize my thoughts and avoid duplicating effort.

Go’s measurement tools can’t see memory allocations in cgo ๐Ÿ”—︎

One of my earliest debugging steps was adding a page to PicoShare that showed some of the RAM metrics from runtime.ReadMemStats (I later realized that net/http/pprof did this better).

PicoShare debug page showing Alloc: 96.47 MB, TotalAlloc: 395.47 MB

James Tucker pointed out that this measurement would exclude any resources I allocated through cgo:

are you using sqlite via cgo, or a pure go implementation? a cgo version would allocate memory unknown to the go heap statistics.

I was indeed using SQLite via cgo. PicoShare uses mattn/go-sqlite3, the most popular SQLite library for Go.

And it makes sense that using cgo prevents Go from showing accurate performance metrics. If you’re using Go to call external C code, Go can’t track resources in the external code.

To work around this, I tried using modernc.org/sqlite, a pure Go implementation of SQLite. But for whatever reason, I couldn’t see resource leaks with pure Go code either.

Fly has a crazy fast disk performance ๐Ÿ”—︎

At one point, commenters on Twitter suggested that I might be exhausting RAM with disk writes. If PicoShare was writing to the Fly VM’s disk faster than the disk could write the data to physical media, the data would get queued in RAM.

To test this theory, I used the fio disk benchmarking utility, which I’d never used before. I thought I was using the tool wrong, because it reported write speeds of 3353 MB/s, which seemed impossibly fast for a cloud VM. For context, that’s about 30 times faster than when I tried the same test on my home NAS server.

Kurt Mackey confirmed that the measurements were likely correct because Fly’s local disks are Enterprise NVMe drives:

Dead ends ๐Ÿ”—︎

As much as I wish my investigation was an exercise of strictly increasing progress, I took lots of wrong turns and followed hypotheses that led nowhere. Here are some of those dead ends.

Blaming Litestream ๐Ÿ”—︎

The first rule of debugging is to assume the problem is in your code. But I broke that rule here, partially because I dreaded how much work it would be to chase down these bugs.

That said, there were legitimate reasons to suspect Litestream. Even though PicoShare uses SQLite in a strange way, storing 1 GB of data in SQLite isn’t that strange. Litestream is relatively new and uses SQLite in novel ways, so it wasn’t too big a leap to imagine the crashes were coming from Litestream.

Even though I suspected Litestream, I didn’t want to create more work for Ben Johnson, Litestream’s maintainer. I knew PicoShare was an unusual use case for Litestream, and I didn’t have a simple repro to isolate the problem.

But then in May, Fly acquired Litestream and hired Ben to maintain it. Now seemed like the perfect time to bother Ben with this as it concerned both Litestream and Fly!

I filed a bug on Litestream explaining what I’d tried and why I thought the problem was related to Litestream. And then 15 minutes later, I managed to crash PicoShare with Litestream disabled, so I closed the bug.

That said, filing the issue against Litestream was useful because it forced me to approach the problem rigorously enough to write a detailed bug report. And it also piqued Ben’s curiosity, leading him to offer lots of useful advice even after it was clear Litestream wasn’t the cause.

Blaming Fly ๐Ÿ”—︎

When I couldn’t reproduce the crashes on my local VMs or under Docker, I started to suspect that the problem was on Fly’s end. It seemed unlikely because I wasn’t doing anything very exotic, so it would be strange if none of Fly’s other users had noticed their deployments dying from RAM starvation.

Still, I wanted to eliminate Fly as a possibility. I deployed PicoShare to Lightsail, Amazon’s managed Docker container service. They don’t have a 256 MB RAM option, so I deployed to a 512 MB instance. Within a few minutes, I was able to reproduce the crash there, eliminating Fly as the culprit:

/tmp is not a RAMdisk ๐Ÿ”—︎

In the Twitter thread, a few commenters suggested that Fly might be mounting its temporary directory as a RAM disk. Go’s ParseMultipartForm function keeps file uploads in the temporary directory, so if Fly mounted the temp directory as a RAM disk, that would explain why regular file uploads exhausted RAM.

But when I ran lblk, I didn’t see any indications that /tmp or any other temp directory was a RAM disk. The system just seemed to have regular disks:

$ lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINTS
vda  254:0   0  128M  0 disk
vdb  254:16  0    8G  0 disk /
$ du -h /tmp/*
149.3M /tmp/multipart-338465586
7.66 /tmp/test1

Writing a hand-crafted multipart form reader ๐Ÿ”—︎

It was a red flag that the Go standard library ParseMultipartForm function seeming to consume more memory than its documented limit. I also noticed that if I called ParseMultipartForm and discarded the bytes, it also maxed out RAM.

To see if I could work around issues in ParseMultipartForm, I tried writing my own artisanal hand-crafted multipart reader for my specific scenario.

Sadly, my multipart reader performed no better than the standard library, but it was fun to play around with multipart data at a lower level.

Throttling uploads ๐Ÿ”—︎

I was curious if uploading data more slowly would have any effect on RAM. I tried throttling uploads from the client end by setting Chrome to simulate 3G speeds, but PicoShare had the same behavior.

I also tried throttling at the server end to reduce speed of writing to disk, but that didn’t do anything either.

I did learn how easy it is to throttle I/O in Go. If you’re working with an io.Reader interface, you can just wrap a Reader with a throttled reader like this:

import "github.com/juju/ratelimit"

...

throttleRate := 1 << 20 // 1 MB
bucket := ratelimit.NewBucketWithRate(float64(throttleRate), throttleRate)
throttledReader := ratelimit.Reader(reader, bucket)

w := file.NewWriter(tx, metadata.ID, d.chunkSize)
if _, err := io.Copy(w, throttledReader); err != nil {
  return err
}

PicoShare 1.2.0 ๐Ÿ”—︎

Over the weekend, I published PicoShare’s 1.2.0 release, which includes fixes for all the performance issues I discovered through this investigation.

Acknowledgments ๐Ÿ”—︎

A big thanks to everyone who helped me investigate this issue, but a special thanks to a few people who went above and beyond: