Sorry if this has been asked before but I could not find a good answer
in the forums...
With respect to DVD encoding, at what bitrate cutoff would you choose
to do a fast CBR encode rather than a VBR encode? In other words, if
your bitrate calculator says to use a bitrate of XXXX Mbps, what
value of XXXX is high enough for you to just CBR encode it
instead of VBR encoding it? What would answer be if you used 9000Mbps
as the max VBR bitrate? and for max 8000 Mbps VBR bitrate?
+ Reply to Thread
Results 1 to 4 of 4
I would say, first encode a section that is of highest movements. Then,
see how high the bitrate went up with that bitrate, using bitrate.exe
Then, just ues that as your highest. However, if you want to go further
in your math, you could factor in w/ the bitrate calculator, (assuming you
are wanting to CBR your source)
The way I see it is like this, IF you are ceroius about highest quality
in your DV footage transfers, you want to match it, grain-for-grain, not
macro-blocks to macro-blocks If I see grain in my footage, I want
my footage to have that SAME grain. I don't want to see it "smoothed"
out or pixelated or macro-block'ulated. And, only a final burn to CD/DVD
and played on your TV, will you conclude w/ best judgements on your encodes.
There are other techniques to better quality DV to DVD transfer or projects,
but this is not the time here. You can have a look at a sample I did w/
my TRV-22 at my VHELP's Sample clips... thread for an idea or approach towards better quality.
I use 8000 mbit/s CBR and can fit one hour with WAV audio and good quality from DV to DVD-R. If the average bitrate is below 8000 then I set max to 8000 and use 2-pass VBR instead.Ronny