Offloading Peertube Media to S3

My last post documented offloading Azuracast media to Digital Ocean’s Spaces, and this is a follow-up on the offloading theme. Only this time it was for my Peertube instance hosted at I’ll be honest, this could have been a very painful process if Taylor hadn’t  blogged how he did it with Backblaze last December, so in many ways thanks his work made the process a piece of cake, so thank you, Taylor!

I made all media public in the Digital Ocean Spaces bucket, and also created a CORS rule to give the domain full control over bucket. After that, I just got the keys, endpoint, region, and bucket name, and I was good to go:

  enabled: true
  endpoint: '' 
    public: 'public-read'
    private: 'private' 
    bucket_name: 'bavatv'
    prefix: 'videos/' 
    bucket_name: 'bavatv' 
    prefix: 'streaming-playlists/' 
    access_key_id: 'youraccesskey' 
    secret_access_key: 'yoursupersecretkeytothekingdom'

Once I replaced that with existing object_storage code at the bottom of  the production.yaml file (found in /home/peertube/docker-volume/config/) and restarted the container the site was offloading videos to my bucket in Spaces seamlessly. Now I just needed to offload all existing videos to that bucket, which this command did perfectly:

docker-compose exec -u peertube peertube npm run create-move-video-storage-job -- --to-object-storage --all-videos

After it was done running it removed close to 120GBs from the server and my videos loaded from the Spaces bucket. The videos load a bit slower, but that is to be expected, and the next piece will be finding a way to speed them up using a proxy of some kind. But for now I am offloading media like it’s my job, well, actually, it kind of is.

This entry was posted in PeerTube, s3 and tagged , , , , . Bookmark the permalink.

4 Responses to Offloading Peertube Media to S3

  1. Taylor Jadin says:

    At some point I need to try and use Cloudflare in front of a storage bucket to make things load even faster.
    Backblaze has this article, but I think something similar would work for lots of s3-style services:

    • Reverend says:

      Yeah, one thing that annoys me about DO Spaces is you have to run the DNS for a domain through them to get a alias like, which I would want. I may try SW and Cloudfront for the experience to see what that is like, but I also have to prepare myself for the complexity. The simplicity of DO is hard to beat, and that also seems to be the case with Backblaze.

  2. How do the costs compare?

    • Reverend says:

      Spaces is a flat $5 a month for 250GB, which is not bad, but I think Backblaze is even cheaper. If you are amassing storage in the Cloud, I think S3 is a no brainer simply for portability and longevity, but I also thin k it will also save you money in the end.

Leave a Reply

Your email address will not be published. Required fields are marked *

This site uses Akismet to reduce spam. Learn how your comment data is processed.