selfhosted-apps-docker/qbittorrent
DoTheEvo 3ecfc48706 update 2022-01-17 21:43:12 +01:00
..
readme.md update 2022-01-17 21:43:12 +01:00

readme.md

qBittorrent in docker

guide-by-example

logo

Purpose & Overview

WORK IN PROGRESS
WORK IN PROGRESS
WORK IN PROGRESS

Torrents downloader.

qBittorrent is a cross-platform free and open-source BitTorrent client.
The interface is clone of the popular uTorrent.
Major feature is the build in torrent search on various trackers.

Written mostly in C++, using libtorrent under the hood.

Files and directory structure

/mnt/
└── smalldisk/
    └── torrents/
/home/
└── ~/
    └── docker/
        └── qbittorrent/
            ├── qbittorrent-config/
            ├── .env
            └── docker-compose.yml
  • /mnt/bigdisk/... - a mounted storage share
  • jellyfin-config/ - configuration
  • .env - a file containing environment variables for docker compose
  • docker-compose.yml - a docker compose file, telling docker how to run the containers

You only need to provide the files.
The directories are created by docker compose on the first run.

docker-compose

Dockerhub linuxserver/bookstack example compose.

docker-compose.yml

services:

  qbittorrent:
    image: lscr.io/linuxserver/qbittorrent
    container_name: qbittorrent
    hostname: qbittorrent
    restart: unless-stopped
    env_file: .env
    volumes:
      - ./qbittorrent-config:/config
      - /mnt/smalldisk/torrents:/downloads
    expose:
      - 8080
    ports:
      - 6881:6881
      - 6881:6881/udp
      - 8080:8080

networks:
  default:
    name: $DOCKER_MY_NETWORK
    external: true

.env

# GENERAL
MY_DOMAIN=example.com
DOCKER_MY_NETWORK=caddy_net
TZ=Europe/Bratislava

WEBUI_PORT=8080
PUID=1000
PGID=1000

All containers must be on the same network.
Which is named in the .env file.
If one does not exist yet: docker network create caddy_net

Reverse proxy

Caddy is used, details here.

Caddyfile

q.{$MY_DOMAIN} {
    reverse_proxy qbittorrent:8080
}

First run

Default login: admin@admin.com // password


interface-pic

Specifics of my setup

  • no long term use yet

  • no gpu, so no experience with hw transcoding

  • media files are stored and shared on trunas scale VM and mounted to the docker host using systemd mounts, instead of fstab or autofs.

    /etc/systemd/system/mnt-bigdisk.mount

    [Unit]
    Description=12TB truenas mount
    
    [Mount]
    What=//10.0.19.19/Dataset-01
    Where=/mnt/bigdisk
    Type=cifs
    Options=ro,username=ja,password=qq,file_mode=0700,dir_mode=0700,uid=1000
    DirectoryMode=0700
    
    [Install]
    WantedBy=multi-user.target
    

    /etc/systemd/system/mnt-bigdisk.automount

    [Unit]
    Description=myshare automount
    
    [Automount]
    Where=/mnt/bigdisk
    
    [Install]
    WantedBy=multi-user.target
    

    automount on boot - sudo systemctl start mnt-bigdisk.automount

Update

Manual image update:

  • docker-compose pull
  • docker-compose up -d
  • docker image prune

Backup and restore

Backup

Using borg that makes daily snapshot of the entire directory.

Restore

  • down the bookstack containers docker-compose down
  • delete the entire bookstack directory
  • from the backup copy back the bookstack directory
  • start the containers docker-compose up -d

Backup of just user data

Users data daily export using the official procedure.
For bookstack it means database dump and backing up several directories containing user uploaded files.

Daily borg run takes care of backing up the directories. So only database dump is needed.
The created backup sqlite3 file is overwritten on every run of the script, but that's ok since borg is making daily snapshots.

Create a backup script

Placed inside bookstack directory on the host

bookstack-backup-script.sh

#!/bin/bash

# CREATE DATABASE DUMP, bash -c '...' IS USED OTHERWISE OUTPUT > WOULD TRY TO GO TO THE HOST
docker container exec bookstack-db bash -c 'mysqldump -u $MYSQL_USER -p$MYSQL_PASSWORD $MYSQL_DATABASE > $MYSQL_DIR/BACKUP.bookstack.database.sql'

the script must be executable - chmod +x bookstack-backup-script.sh

Cronjob

Running on the host, so that the script will be periodically run.

  • su - switch to root
  • crontab -e - add new cron job
  • 0 22 * * * /home/bastard/docker/bookstack/bookstack-backup-script.sh
    runs it every day at 22:00
  • crontab -l - list cronjobs to check

Restore the user data

Assuming clean start, first restore the database before running the app container.

  • start only the database container: docker-compose up -d bookstack-db
  • copy BACKUP.bookstack.database.sql in bookstack/bookstack-db-data/
  • restore the database inside the container
    docker container exec --workdir /config bookstack-db bash -c 'mysql -u $MYSQL_USER -p$MYSQL_PASSWORD $MYSQL_DATABASE < BACKUP.bookstack.database.sql'
  • now start the app container: docker-compose up -d
  • let it run so it creates its file structure
  • down the containers docker-compose down
  • in bookstack/bookstack-data/www/
    replace directories files,images,uploads and the file .env
    with the ones from the BorgBackup repository
  • start the containers: docker-compose up -d
  • if there was a major version jump, exec in to the app container and run php artisan migrate
    docker container exec -it bookstack /bin/bash
    cd /var/www/html/
    php artisan migrate

Again, the above steps are based on the official procedure.