Amazon VOD publishes they use a bit rate of 2.5 Mbps for their HD 720p content.
When I capture at 2.5 Mbps (MP4: h.264 and AAC) I can see obvious pixelation. If I bump up the bit rate to 5.0 Mbps, it is much cleaner.
I am trying to understand why the 2.5 is not adequate. Being a noob, I would expect it to be an exact replica of the stream.
I am guessing, re-encoding an already compressed stream is the culprit. If that is the case, what bit rate do I have to use in order to maintain 99% of the quality of the stream?
+ Reply to Thread
Results 1 to 14 of 14
Thread: Bit Rate Question
They have streaming to worry about, so they settled on a (relatively) low Mbps.
Perhaps they soften the image a little to get away with it.
What are you trying to do with this stream ?
"I am trying to understand why the 2.5 is not adequate" - adequate for what?
I am capturing the stream (MP4 - h.256 AAC).
When i configure the capture, setting the bit rate at 2.5 Mbps (which is Amazon's published bit rate) results in a much poorer quality than what the original stream looks like.
I would like to maintain the original quality of the stream. I want to know what he minimum bit rate is that I can do that with.
As a general noob question, if i re-encode a video with the exact same parameters as it was originally encoded with do I lose quality?
Every time video is re-encoded, there is some quality loss regardless of the settings.
How are you capturing the stream? Can you not capture it in its original form?
Isn't Amazons stream DRM protected? Please explain your process.
I am capturing output from my PS4 using an avermedia extreme cap u3 unit.
I have amazon app installed on PS4.
Not doing anything to bypass DRM, I plugged my PS4 into the avermedia unit and it just works.
Capture at a higher bitrate and then re-encode to 2.5mbps using 2 pass,you will get much better results than capturing at 2.5 mbps.I think,therefore i am a hamster.
It's difficult to recommend a specific bitrate, so much depends on the material being captured.
For example, the video complexity, sharpness, noise, etc,etc.
So mach better if the software , instead of a bitrate setting, had a quality setting.
For example, H.264, use Constant Rate Factor - then you get the specific quality you ask for, and the
resulting bitrate achieved is a secondary consideration.
Does your capture software offer a "quality" setting, Vs. hard coded bitrate?
Perhaps as an experiment you should consider capturing a small sample at a high bitrate (say 10mbps)
and then re-encoding in Ripbot264 or Handbrake using one of the presets. Would offer you many more options.
There may be some others in the forum who may have specific experience re: this kind of capture
and may offer some alternate suggestions.
1) Hardware AVC encoders aren't as good as software AVC encoders. So they'll need more bitrate right off the bat.
2) If you're capturing 720p from a game console it's 720p at 60 fps. Movies from Amazon are at 24 fps. More fps means more bitrate is required.
3) Different videos require different amounts of bitrate. The more motion, grain, etc., the more bitrate you need.
Thanks to all.
This unit actually is a software based encoder.
There is not a quality setting, however it can capture 1080p 60fps at 30 Mbps.
I will capture using max settings and then re-encode as suggested.
Realtime encoder encodes during the capture - time is of the essence; any delays = dropped frames.
Regular encoders, from a source already on your PC, can take as long as it wants, the next frame will always be there,
no matter how slow the encode goes.
Last edited by blimey; 14th Jan 2014 at 01:50.