Hi I have jellyfin installed in a VM with 24 cores and 32GiB RAM (VM also used for Docker). Whenever I attempt to play higher quality files, jellyfin crashes after a few minutes. I haven’t seen it struggle with lower quality media.
Here are some logs: FFmpeg.Transcode-2024-03-24_16-11-43_d48825174d455ae3ff859d8b28582853_ce3f3ebf.log upload_org.jellyfin.androidtv_0.16.7_20240324161053_d2befd034e424a3490e7ea55af1fe1f2.log Fmpeg.Transcode-2024-03-24_16-11-38_d48825174d455ae3ff859d8b28582853_dafa4555.log FFmpeg.DirectStream-2024-03-24_16-08-12_d48825174d455ae3ff859d8b28582853_dac7115f.log
I cannot for the life of me figure out whats wrong. I’ve tried disabling plugins, different clients, hard resets etc, but it still crashes.
Can someone enlighten me?? :(
Why not in a VM? All my stuff runs in VM’s.
Doesn’t the vm produce an overhead? More power consumption and it can’t use the full capabilities of the machine?
There are many benefits to VMs. You can limit how much RAM is available to each one, so one app doesn’t eat all of your RAM. Same with CPU. Virtual Machines can be backed up, uploaded to remote storage, and restored. When it’s time to do a big update on your main machine (either changing OS or getting new hardware), restoring VM’s is super simple compared to the alternative.
You can limit them in docker as well
https://phoenixnap.com/kb/docker-memory-and-cpu-limit You can spin up the same container on another machine with one command.
Docker seems to be easier, not?
Docker is great, and I have it running in multiple VM’s. For me to restore everything without VM’s would be a little tedious.
I’d have to search for Docker install instructions, follow the instructions like importing keys, adding the repository, doing the post-install stuff (adding my user to the docker group), etc. Not a big deal, but it’s something.
Then I’d start copying data, making sure to keep the same folder structure so my compose files work. Then I start running all the commands to get all the containers back up and running (or in my case, creating 20+ stacks in Portainer).
This is all a bit tedious for me compared to opening the web interface for Proxmox, clicking on the backup (which is just “there” because it’s on a hard drive (or a ZFS pool in my case)), and clicking “restore”. Once restored it just boots up and my Docker stuff is good to go.
Can run multiple different os’s - I have truenas, windows server, Ubuntu server for docker, etc
This can be done with containers and you don’t get the overhead of virtualizing a whole operating system for every service/app you might be hosting.
This can also be done with containers in a more elegant way as there’s no need to back up any VM/OS data.
E.g. I have a docker compose file that can nearly immediately stand up a container with the right settings/image, point it to my restored data and be up and running in no time. The best part is i don’t need to back up the container/OS because that data is irrelevant.
With the alternative you just restore your data and run
docker-compose up -d
. Docker will handle the process of building, starting and managing the service.Simple example: Your minecraft server died but you have backups. You just restore the data to
/docker/minecraft
. Then (to keep things really simple) you just run:docker run -d -p 25565:25565 --name minecraft -e EULA=TRUE -v /docker/minecraft:/data itzg/minecraft-server
and in a few minutes your server is ready to go.
If that’s the way you’d prefer to do it, I highly recommend taking that approach.
VMs can also do something called https://en.wikipedia.org/wiki/Kernel_same-page_merging - meaning a single machine could run something like 52 instances of Windows XP on 1gb of RAM
So for example, say I needed to run a bunch of ARK servers - I could put them all in VMs, and since they’re all loading mostly the same assets, I could run a lot of them in the very tiny amount of memory.
Thats crazy
Why do you guys do it?