Hi I have jellyfin installed in a VM with 24 cores and 32GiB RAM (VM also used for Docker). Whenever I attempt to play higher quality files, jellyfin crashes after a few minutes. I haven’t seen it struggle with lower quality media.

Here are some logs: FFmpeg.Transcode-2024-03-24_16-11-43_d48825174d455ae3ff859d8b28582853_ce3f3ebf.log upload_org.jellyfin.androidtv_0.16.7_20240324161053_d2befd034e424a3490e7ea55af1fe1f2.log Fmpeg.Transcode-2024-03-24_16-11-38_d48825174d455ae3ff859d8b28582853_dafa4555.log FFmpeg.DirectStream-2024-03-24_16-08-12_d48825174d455ae3ff859d8b28582853_dac7115f.log

I cannot for the life of me figure out whats wrong. I’ve tried disabling plugins, different clients, hard resets etc, but it still crashes.

Can someone enlighten me??  :(

  • HumanPerson@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    4
    ·
    7 months ago

    It looks like your using nvidia for transcodes. Try disabling hardware acceleration temporarily to test if it is related.

  • stankmut@lemmy.world
    link
    fedilink
    English
    arrow-up
    4
    ·
    edit-2
    7 months ago

    I have this issue whenever I have to deal with transcoding. The stream will usually die after a few minutes of watching. I haven’t seen anything out of the ordinary in the logs. I’ve just resigned myself to lugging around the device I have that can direct stream the files.

  • Diplomjodler@feddit.de
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    1
    ·
    7 months ago

    This is pretty basic but have you checked how many cores and how much memory you have allocated to your VM?

  • AMillionMonkeys@lemmy.world
    link
    fedilink
    English
    arrow-up
    3
    ·
    7 months ago

    If the crashes are seemingly at random when transcoding I would suspect overheating hardware. Transcoding uses more energy and produces more heat than playing directly. If it were a software issue it would either work or not.

    • Sammeeeeeee@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      1
      ·
      edit-2
      7 months ago

      Pastebin wouldn’t accept it (said it was nsfw - maybe because of killing processes?) so I used paste.io, which I now see is blocked by adblocking. I am not using Hardware transcoding, I am unsure where to find the driver version.

    • Sammeeeeeee@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      6
      ·
      7 months ago

      I have a couple of vms on my server - truenas, pihole, etc. One is for my media - arrs in docker containers and Jellyfin installed (not at container as it wasn’t binding to the media dir properly)

        • waldenA
          link
          fedilink
          English
          arrow-up
          6
          ·
          7 months ago

          Why not in a VM? All my stuff runs in VM’s.

          • GravitySpoiled@lemmy.ml
            link
            fedilink
            English
            arrow-up
            4
            arrow-down
            2
            ·
            7 months ago

            Doesn’t the vm produce an overhead? More power consumption and it can’t use the full capabilities of the machine?

            • waldenA
              link
              fedilink
              English
              arrow-up
              5
              ·
              7 months ago

              There are many benefits to VMs. You can limit how much RAM is available to each one, so one app doesn’t eat all of your RAM. Same with CPU. Virtual Machines can be backed up, uploaded to remote storage, and restored. When it’s time to do a big update on your main machine (either changing OS or getting new hardware), restoring VM’s is super simple compared to the alternative.

                • waldenA
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  ·
                  7 months ago

                  Docker is great, and I have it running in multiple VM’s. For me to restore everything without VM’s would be a little tedious.

                  I’d have to search for Docker install instructions, follow the instructions like importing keys, adding the repository, doing the post-install stuff (adding my user to the docker group), etc. Not a big deal, but it’s something.

                  Then I’d start copying data, making sure to keep the same folder structure so my compose files work. Then I start running all the commands to get all the containers back up and running (or in my case, creating 20+ stacks in Portainer).

                  This is all a bit tedious for me compared to opening the web interface for Proxmox, clicking on the backup (which is just “there” because it’s on a hard drive (or a ZFS pool in my case)), and clicking “restore”. Once restored it just boots up and my Docker stuff is good to go.

                • Sammeeeeeee@lemmy.worldOP
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  7 months ago

                  Can run multiple different os’s - I have truenas, windows server, Ubuntu server for docker, etc

              • ᗪᗩᗰᑎ@lemmy.ml
                link
                fedilink
                English
                arrow-up
                2
                ·
                7 months ago

                You can limit how much RAM is available to each one, so one app doesn’t eat all of your RAM. Same with CPU.

                This can be done with containers and you don’t get the overhead of virtualizing a whole operating system for every service/app you might be hosting.

                Virtual Machines can be backed up, uploaded to remote storage, and restored.

                This can also be done with containers in a more elegant way as there’s no need to back up any VM/OS data.

                E.g. I have a docker compose file that can nearly immediately stand up a container with the right settings/image, point it to my restored data and be up and running in no time. The best part is i don’t need to back up the container/OS because that data is irrelevant.

                When it’s time to do a big update on your main machine (either changing OS or getting new hardware), restoring VM’s is super simple compared to the alternative.

                With the alternative you just restore your data and run docker-compose up -d. Docker will handle the process of building, starting and managing the service.

                Simple example: Your minecraft server died but you have backups. You just restore the data to /docker/minecraft. Then (to keep things really simple) you just run:

                docker run -d -p 25565:25565 --name minecraft -e EULA=TRUE -v /docker/minecraft:/data itzg/minecraft-server

                and in a few minutes your server is ready to go.

                • waldenA
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  7 months ago

                  If that’s the way you’d prefer to do it, I highly recommend taking that approach.

            • thantik@lemmy.world
              link
              fedilink
              English
              arrow-up
              2
              ·
              edit-2
              7 months ago

              VMs can also do something called https://en.wikipedia.org/wiki/Kernel_same-page_merging - meaning a single machine could run something like 52 instances of Windows XP on 1gb of RAM

              So for example, say I needed to run a bunch of ARK servers - I could put them all in VMs, and since they’re all loading mostly the same assets, I could run a lot of them in the very tiny amount of memory.

  • entropicdrift@lemmy.sdf.org
    link
    fedilink
    English
    arrow-up
    1
    ·
    7 months ago

    There’s a couple things that stick out to me in these logs:

    1. the androidtv client logs show the crash was due to a time-out while waiting for the server to respond to a request for over 3 seconds.

    2. you included a DirectStream log, that’s usually not too hard for even a Raspberry Pi to handle reasonably well

    3. the first log I looked at showed libx264 was using 36 threads

    So my guess is that, due to video encoding not scaling all that well across larger numbers of threads, your server is being bogged down with this transcoding and isn’t providing http responses fast enough to your client device(s).

    A simple way to troubleshoot this would be to explicitly set the transcoding thread count to something a bit lower than what your server has, say 16 or 20, and see if that does any better.

    An obvious potential fix would be to use hardware acceleration if it’s available to you. I run my Jellyfin server off of a little N100 mini PC and it can transcode 4K HDR 70 Mbps video (tonemapping included) at about 45 fps due to the hw acceleration. That said, I know it can be tricky to set up in a VM and you may not have the HW accel capability in your server CPU anyhow.

    • Sammeeeeeee@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      2
      ·
      7 months ago

      Wow, I never considered that too many cores may be the issue. I will for sure try that.

      I believe my cpu is compatible, but I believe this is not iGPU transcoding. Pretty sure my server does have an iGPU (although I’ve never actually plugged in a screen so I can’t be sure). I will research passing through either of those.

      I also read that it’s possible to use a different server for just transcoding. I have an old laptop I’ve been considering setting for that, but it’s network speeds are shite.