Hey everyone, I’m building a new server to run Jellyfin (with a few other services like Pi-hole) and I’m stuck on GPU or CPU transcoding.

My main concern is smooth 4K HDR transcoding for 1 stream. I’ve been reading mixed advice online – some people say a strong CPU with good single-core performance can handle it, while others recommend a dedicated GPU.

Should I focus my budget (~$1000AUD/$658USD) on a good CPU, or spend some of it on a dedicated GPU?

  • Possibly linux@lemmy.zip
    link
    fedilink
    English
    arrow-up
    2
    ·
    8 months ago

    I tried that with a cheap minipc I bought and it was CPU limited. The GPU was fine it was the overhead that killed me.

    • Faceman🇦🇺@discuss.tchncs.de
      link
      fedilink
      English
      arrow-up
      2
      ·
      8 months ago

      Was it an n100? They have a severely limited power budget of 6w compared to the n95 at 25w or so.

      I’m running jellyfin ontop of ubuntu desktop while also playing retro games. That all sits in a proxmox vm with other services running alongside it. It’s perfectly snappy.

        • Faceman🇦🇺@discuss.tchncs.de
          link
          fedilink
          English
          arrow-up
          1
          ·
          8 months ago

          N5095 ? lots of reports of that one not supporting everything it should based on other Jasper Lake chips, CPU getting hit for Decode when it shouldn’t for example. Also HDR to SDR cant be accelerated with VPP on that one as far as I know so the CPU gets smashed. I think you can do it with OpenCL though.