this post was submitted on 18 Oct 2023
24 points (87.5% liked)
Piracy: ꜱᴀɪʟ ᴛʜᴇ ʜɪɢʜ ꜱᴇᴀꜱ
54716 readers
336 users here now
⚓ Dedicated to the discussion of digital piracy, including ethical problems and legal advancements.
Rules • Full Version
1. Posts must be related to the discussion of digital piracy
2. Don't request invites, trade, sell, or self-promote
3. Don't request or link to specific pirated titles, including DMs
4. Don't submit low-quality posts, be entitled, or harass others
Loot, Pillage, & Plunder
📜 c/Piracy Wiki (Community Edition):
💰 Please help cover server costs.
Ko-fi | Liberapay |
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Yes it’s good, but with AV1 hanging about then you’re WAAAY better off using that over x265.
I re-encode all my stuff with AV1. It will take a 40GB x264 rip to 3-4GB. Where as with x265 It will be around 10-15GB.
It’s a significant difference in storage size and (as far as I can tell) no obvious difference in quality.
if u re getting 3 to 4Gb out of a 40G x264, u re definitely losing a lot of data... With proper settings, It should be around 16 - 18G. AV1 can't do miracles.
For sure!
What I’m talking about is perception of quality. If you know what you’re looking for then you’ll notice some of the artifacts. Especially in the darkest areas and when going from HDR to SDR.
It still looks better than steaming the same thing off of Netflix, Hulu, etc. So that’s all I need/want.
I fully realize there’s compromise there and if I want to view it in all its original glory I can bust out the bluray.
Sadly some clients (nvidia shield tv) does not support AV1 :( right now I'm encoding some AV1 content I have back to HEVC just because of that.
And most older gpu models also don’t support av1. So transcoding for these clients happens on the cpu. This is why I will continue to use x265 for now.
Only NVIDIA 3000-series cards and up support hardware AV1 decoding and only 4000-series cards support both hardware AV1 encoding and decoding, so only just about 3 year old cards! source
All Intel ARC cards support both decode and encode, (released October 2022) source
AMD 6000-series cards (November 2020) support AV1 hardware decoding and only 7000-series cards (December 2022) support both hardware AV1 encoding and decoding. source
Yes, you’re completely correct. There’s something to consider though.
CPU encoding gives the best results possible, in terms of quality and size. Decoding, unless you have a very weak CPU, isn’t necessarily the bottleneck it most transcoding applications eg plex, jellyfin, etc.
So you can do things to make the media as streamable as possible for instance encoding your media in AV1 using the mp4 container rather than mkv. If you make it web optimized aka ATOM upfront it makes playing the file much easier and less resource intensive. Now when a client that can’t use AV1 requests it your transcode can do SW decode and HW encode. Not as efficient as pure HW but IMHO it’s a worthwhile trade off for the storage space you get in return.
You can make things more efficient by disabling subtitles and/or burnin on the media server side. If you have people like myself who need subs in everything then you can burn them in while you’re encoding the media to AV1 or only using formats like UTF8 so you can pass through them as m4v/mp4 doesn’t support subs like mkv does.
That’s essentially what the optimized versions do on Plex. Only it sticks with x264 rather than AV1.
If your media is only 720p then none of this would really make a difference for you. If you’re using 1080p+ rips then this will make a SIGNIFICANT difference. It’s made such a difference that I’ve started redoing my rips in 4K.
Unless that is you got a SAN in your closet and free electricity that is…