BBC Lowers HDTV Bitrate; Users Notice 412
aws910 writes "According to an article on the BBC website, BBC HD lowered the bitrate of their broadcasts by almost 50% and are surprised that users noticed. From the article: 'The replacement encoders work at a bitrate of 9.7Mbps (megabits per second), while their predecessors worked at 16Mbps, the standard for other broadcasters.' The BBC claims 'We did extensive testing on the new encoders which showed that they could produce pictures at the same or even better quality than the old encoders ...' I got a good laugh off of this, but is it really possible to get better quality from a lower bitrate?"
Focus group... (Score:2, Insightful)
...of blind retards.
Comment removed (Score:5, Funny)
Re:Focus group... (Score:5, Funny)
FTA: ""Even my wife can see a reduction in picture quality and she's got cataracts," wrote one. "
They must have a pretty big screen if she can see that difference from the kitchen.
Re: (Score:3, Funny)
That's just wrong and offensive. I can' not believe you would you would say something so neanderthall.
This is the modern age asshole, and we put TV's in the kitchen~
No TV in the kitchen, sheesh.
Re:Focus group... (Score:5, Funny)
Re:Focus group... (Score:5, Informative)
1) The alleged wife in the quote is purported to have cataracts. Cataracts typically reduce visual acuity due to the cloudiness they impart to the lens of the eye. How does a reduction of visual acuity translate to "just another racist characterization of women being incompetent with technology"?
2) If the quote had been ""Even my husband can see a reduction in picture quality and he's got cataracts," wrote one." would you have bothered to make your little rant post?
P.S. The term you were looking for is "sexist" not "racist".
Re: (Score:2, Insightful)
As a 43 year old father of a special needs child I find your comment retarded (And I deplore the use of the word). The salient point was that the wife of the commenter in the article has cataracts, not that she is female. If the comment had been "Even my husband can see a reduction in picture quality and he's got cataracts" it would have been no less salient.
Re: (Score:3, Insightful)
I deplore the use of the word
Really? I don't get that. The reason that people use the word as an insult is because it is a word used to describe people who have intellectual challenges. It doesn't matter what you do, the euphemism treadmill [wikipedia.org] will get you in the end. As evidence I submit this anecdote: I was spending some time with a couple of young gentlmen (kids of a friend of mine) who are 9 and 11 years old. Big kid was teasing little kid, so big kid said "you're a special needs kid!". Does this help you see the futility of rot
Re: (Score:3, Insightful)
Re: (Score:2, Offtopic)
Excellent troll. Nice haul on that one.
Re:Focus group... (Score:5, Funny)
As a 3 yo lesbian, father of seven, socialist COBOL programmer, I'm not sure which of your stated attributes qualify you to be racially offended.
Re:Focus group... (Score:5, Funny)
My cranium nearly exploded while attempting to parse
Re: (Score:3, Informative)
My cranium nearly exploded while attempting to parse
"3yo lesbian, father of seven"
Father of 7, then transgendered 3 years ago?
Re: (Score:3, Interesting)
In the UK, for tax purposes, the "father" in a lesbian relationship is the one that didn't get pregnant with the child in question.
Re: (Score:2)
"And do you know why? Because they're Scotts! Ha ha ha ha hah!"
Sigh...
Re:Focus group... (Score:5, Informative)
Yes, it IS possible to get higher picture quality out of a lower bitrate, but not with all else equal. For example, you can get higher quality with CPU-intensive settings using H.264 5.1 Profile than you can with H.264 4.1 (what Blu-Ray's/HD DVDs use), at the same bitrate. You're giving up CPU cycles in decoding for lower video size. This is why x264 can produce near-transparent encodes of Blu-Ray movies at about half the size. x264 uses much more demanding settings.
x264 at 20 Mbit which high-quality settings is far more demanding than a 40 Mbit H.264 stream from a Blu-Ray.
Re:Focus group... (Score:5, Interesting)
In the US, Comcast uses codex compression to squeeze HD on their cable systems. When people get to see native resolution at the TV store, then get the Comcast version when they plug in their shiny new HD TV, they wonder WTF? That the beeb would put their foot on the garden hose and expect no one to notice is ludicrous.
I wish the FCC would get involved in the US to force cable companies to limit the number of channels supported and broadcast them in the highest sustainable resolution-- or tell their users the truth about what's happening and why. Maybe we can start to get rid of the excess junk channels.
Re:Focus group... (Score:4, Interesting)
Adding to that, Comcast's programming is 720p, with much of it upscaled. The Blu-Ray source you see at the stores are often 1080p, or at least 1080i. You're comparing rotten wormy apples to nice juicy oranges, where Comcast's feeds are the rotten wormy apples.
Re:Focus group... (Score:4, Informative)
Theoretically, perhaps. In reality either one could look better given other factors.
Re: (Score:3, Informative)
I left comcast because their *digital* signals were worse than standard TV over the antennae.
Dish has gone the other way-- their signal *looks* crisp, but there is a lot more blockiness than there used to be. I used to have blocky outbreaks perhaps 1 or 2 times in 40 hours of viewing. Now I get blockiness 1 or 2 times per 10 hours of viewing.
Re: (Score:3, Interesting)
to limit the number of channels supported
Of course, as long as most storylines are low def, having a lot of channels might not be a bad idea; don't assume they'll necessarily kill off the channels you consider excess junk.
Personally I can't say I care much about resolution. As long as the format is progressive, digital and anywhere near SD increases in resolution are about the last thing I'll notice in any normal viewing situation. Improved contrast, lack of compression artifacts, sound, content, basically
Re:Focus group... (Score:4, Interesting)
A codec is a compressor/decompressor piece of code that's used in one of two circumstances-- lossy or non-lossy stream compression, usually (but not always) of audio/video information. The eye and ear can detect certain types lossy compression effects, and some people are better at detecting problems than others. Generally, more compression yields more information loss that is sensed by low quality video (jaggies, weird frame transitions, noise, fewer colors, or distorted sound of various kinds). But more compression means less bandwidth used, so that more streams can be handled per given bandwidth 'space'.
In the US, the current max horizontal by vertical HD TV resolution is 1080 pixels, and its data rate at full color value is about 16megabits/sec. There are two types, interlaced and progressive scans. Interlaced writes and holds information from frame to frame while progressive writes whole frames (a simple explanation) and progressive is preferred but requires more intelligent electronics to produce. The 1080p HD picture is preferred. An interim size, 720p, is often what cable companies send down the wires to your set. The native resolution refers to the uncompressed data rate, or one that's used with a non-lossy compressor (meaning that the decompressor can re-interpret the compressed stream to reproduce the original image 100%).
Re:Focus group... (Score:4, Informative)
Actually, the bandwidth at 8bits per channel not including the 5.1 sound is 16,588,800 bits per FRAME not per second, so at 60 FPS you get a 950 mb/s bandwidth requirement for the video alone, and that`s why we need to use a compressed distribution method...
Re: (Score:3, Interesting)
Codex is the plural of 'codec'. It could also be stated 'codecs'. It's an abbreviation of COmpressor/DECompressor, in the plural.
Saying "Comcast uses codex compression" without specifying any particular type of compression/decompression is rather awkward. "Comcast uses compression" is no less accurate, unless you specifically wanted to distinguish the type of compression that uses a compressor/decompressor from the type of compression that doesn't, if that's even possible.
Besides which, I don't believe you that codex is the plural of codec.
Re: (Score:3, Insightful)
No, those aren't the only differences. For one, you are limited in the number of reference frames you can use at a given resolution at 4.1. For example, at full 1920x1200 I don't think you can use more than 4 or 5 reference frames at 4.1, but I've seen 5.1 encodes that use 16 reference frames for animated films that achieve very high compression ratios while maintaining transparency.
4.1's maximum maximum allowed bitrate is not the constraint. Doom9 or Wikipedia can provide much more detailed information abo
Re: (Score:3, Informative)
These are the Reference frame limits in Level 4.1
Resolution | no. ref
-----------|---------
1280x544 | 12
1280x720 | 9
1920x800 | 5
1920x816 | 5
1920x1080 | 4
If none of the resolutions above match your source, use the following equation to work it out for yourself:
8388608
__________________
They suck at math too (Score:4, Insightful)
Re:They suck at math too (Score:4, Insightful)
To one significant figure, they are.
Re:They suck at math too (Score:4, Informative)
Technically speaking, they suck at "maths".
Re:They suck at math too (Score:5, Funny)
Technically speaking, they suck at "maths".
We Amurkins don't recognize no commie "maths." We want our math to grow up as individuals
Re: (Score:3, Informative)
One generally accepted practice is to put the punctuation inside the quote if the punctuation is part of the quotation, and outside the quote otherwise. According to that rule of thumb, his use of punctuation was correct.
Re: (Score:2)
"Math" is just as valid a truncation as "maths".
Re: (Score:2)
Mostly because it's the difference between singular and plural. While I might shorten Pointy-Haired Boss to PHB, I would also shorten Pointy-Haired Bosses to PHBs, and there is a fundamental difference between one PHB and many PHBs -- just ask Peter Gibbons.
Re: (Score:3, Insightful)
Re: (Score:3, Funny)
Re:They suck at math too (Score:5, Funny)
Re: (Score:2)
Yes (Score:2, Informative)
Re: (Score:3, Insightful)
In other words, lower bitrate can be better, but only if you compare to shitty and inefficient compression.
Re:Yes (Score:5, Insightful)
In other words, lower bitrate can be better, but only if you compare to shitty and inefficient compression.
And by this you mean compression that is state of the art two minutes ago, vs. today. Seriously, this field is moving pretty fast, and what you call shitty and inefficient was not long ago the best people could do. A few years ago when I was messing with the x264-svn libraries, stuff would get tweaked daily.
Not to mention there are other factors at play with regards to compression. A well-engineered system isn't necessarily going to go for the maximum compression rate for video quality. One has to look at other limitations, such as the decoding hardware, the method by which the video is being delievered, and even the viewing devices on the receiving end.
What is disheartening about the article is that it looks like the BBC are just in denial mode, and not really taking the complaints seriously. "Standard viewing equipment"? Seriously, what exactly are they getting at with that comment? On top of that it looks like they are trying to blame the quality of the source material, which certainly muddies the picture, but certainly the customers that are complaining would be used to these variations in quality before the change and not just suddenly notice it at the same time this equipment was rolled out.
I have respect for them sticking to their guns, but not when they are doing it with such lame excuses. Then again, the BBC spokesperson and reporter may not be the most tech savvy individuals, and its likely some of the message here is lost in translation. Lossy codec indeed.
Re: (Score:2)
It would also be quite remarkable to see better quality compared to any other modern encoding while reducing bitrate by 50%.
Yes (Score:5, Informative)
Sure, if you also switch to a better codec, such as using H.264 instead of MPEG-2. However, I don't think that's what's happening in this case.
Re:Yes (Score:5, Insightful)
You can also get better compression by specifying a more sophisticated compression method within the same codec, for example, since many codecs support various levels of compression.
Generally, "better" compression (fitting a higher resolution and/or framerate into a smaller size) requires a lot more power to encode and often some more power to decode. You can use less bitrate to get a quality signal there, but you need "smarter" coders and decoders at the respective ends of the transmission. So the BBC may have upgraded their compression engine to something that can do "better" compression, thereby fitting the same resolution and framerate into a 40% smaller stream. But their customers' television sets might not have the horsepower to decode it at full quality.
That could easily explain why the BBC's testing went so well but their consumers (with varying brands of TV sets probably mostly tested for British use with the old compression) can't keep up and render an inferior picture.
It's also possible that, by compressing the video stream into a denser compression method, signal loss is having a greater effect than it did with the old compression method. The viewers may be seeing artifacts that are the decoder's attempts to fill in the blanks. The old compression method might have allowed a certain amount of redundancy or error correction that the new one lacks, and the loss of part of the signal has a more visible effect on the new one.
Re: (Score:3, Insightful)
> can't keep up and render an inferior picture.
It's not like there's half-way here. This is the digital age - if it can't keep up you won't see anything!
Re: (Score:2, Informative)
On my TV at home I have changed the settings to turn off the "blue screen/bad signal screen". The TV does its best
Re: (Score:2)
These are two different things:
1. Partial decode due to missing data (i.e a picture is comprised of blocks, you can have some of the block but not all of them - hence partial decode)
2. A primitive decoder's inability to decipher/keep-real-time-rate-of an advanced encoding (i.e. trying to decompress something in realtime on a weak machine will at best result in frame drops -- or completely incompatible schemes of decoding: my decoder supports only mpeg-2 but i've got a h.264 movie)
You are talking about optio
Re: (Score:3, Informative)
Say what?
f(x) = 1 - x^2/2 + x^4/24 - x^6/720 + ... (where the constant diminishes rapidly, and 0<x<1)
If you know that the HOT affect the result less and less, you can drop them and still get a "good enough" though less perfect answer. You can keep dropping terms until the error is unacceptable, or in the case of something where the actual value is not critical (i.e. a block of pixels), you can keep dropping terms to reach a target number of operations and hope that the answer is sufficiently precise
So what you're saying is.. (Score:2)
Re: (Score:2)
I got a good laugh off of this, but is it really possible to get better quality from a lower bitrate?
Sure, if you also switch to a better codec, such as using H.264 instead of MPEG-2. However, I don't think that's what's happening in this case
Just to amplify what has been said here a few times, yes it is possible, and not only from changing codecs. H.264 supports many optional features that are not implemented in all decoders, and these features can have an effect on quality. Use of a variable bit rate vice a constant bitrate can also increase quality and decrease bandwidth needs, at the cost of requiring some bursting capability or buffering to accomodate the variation. Also, there are tricks that can be played with dark and light tone maski
Re: (Score:2)
Re: (Score:2)
H.264 = MPEG-4 Part 10 (http://en.wikipedia.org/wiki/H.264/MPEG-4_AVC)
If you mean MPEG-2, H.264 was designed as a replacement for this technology amongst others.
H.264 and VC-1 are currently the most efficient methods in terms of bandwidth to transmit video.
Re:Yes (Score:5, Funny)
Aren't you comparing x264 to oranges?
Of course it is possible (Score:3, Informative)
but is it really possible to get better quality from a lower bitrate?
If you are changing the compression algorithm of course it is possible. In H264, there are a lot of compression possibilities which are not used by the compression algorithm but which will be recognized by the decompression algorithm.
Re: (Score:2)
Re: (Score:2)
That second sentence makes absolutely no sense.
Yes, of course (Score:5, Informative)
Re:Yes, of course (Score:4, Interesting)
Re:Yes, of course (Score:4, Informative)
Re:Yes, of course (Score:5, Informative)
LAME was a pretty good example of this for MP3 - Eventually it was able to achieve (somewhat) better quality at (somewhat) lower bitrates than the reference encoders.
Vorbis, similarly, had the AoTUV tuning - This provided significant rate/distortion tradeoff improvements compared to a "vanilla" encoder, without changing the decoder.
However, 40% reduction in bitrate with an increase in quality is very difficult unless the original encoder was CRAP. (Which is actually a definite possibility for a realtime hardware encoder.) Also, it's far more likely to have such improvements with H.264 or MPEG-4 ASP, not nearly as likely with MPEG-2, which had a far less flexible encoding scheme.
Re: (Score:3, Informative)
Which is actually a definite possibility for a realtime hardware encoder.
Not just a realtime hardware encoder, but likely a first-generation encoder. Most compression standards are now designed with some headroom. When AAC was first introduced, Dolby provided two encoders, a consumer-grade and a professional encoder. The consumer-grade one was only slightly better than MP3, but ran much faster. The pro encoder was a lot slower but the quality was noticeably better. More recent encoders produce even better quality. A 40% decrease in bitrate is about what I'd expect going f
Re:Yes, of course (Score:4, Informative)
A 40% decrease in bitrate is about what I'd expect going from a single-pass to a two-pass H.264 encoder, and it's entirely possible that a newer single-pass encoder can do the same sort of thing just by using a longer window now that RAM is a lot cheaper.
No, there is no difference in compressibility between a single pass and a two pass encoder. The two pass encoder simply allows you to set the quantizer so as to very accurately hit a target average bitrate.
Re: (Score:3, Funny)
Re:Yes, of course (Score:4, Insightful)
You do understand though, that the lost information in your example is lost at the capture stage not the compression stage don't you?
Lossless compression is just that - lossless. Try compressing your copy of notepad.exe with WinZip, extract it and tell me if it still works. That's lossless compression. The result of compression then decompression is bitwise identical to the original. It has nothing to do with whether the original data is an accurate representation of what it claims to be.
Re:Yes, of course (Score:4, Funny)
I cmprsd ths pst wth jpg nd thn dcmprsd t, lssy ncdng s jst s gd s lsslss ncdng.
Re:Yes, of course (Score:4, Funny)
So, DVD vs. Blu-Ray is pointless if I'm using eye-buds?
Bitrate vs. Quality (Score:5, Insightful)
If your decode hardware is fixed (it's generic HDTV hardware), then there is much less room for improvement, and half the bitrate is an enormous drop. It's no surprise that the BBC viewers complained.
Re: (Score:2)
You're assuming they started with a halfway decent encoder to begin with. The difference between a good encoder and a crappy one is vastly more than 50%.
Added complexity need not be involved (though it certainly can help). A better Quantization table, for instance, wouldn't be any slower, and the reduced bitrate would speed-up encoding/decoding.
More relevant
Summary rounding error (Score:5, Informative)
Re: (Score:2, Interesting)
Depends on your definition of "almost" and what precision you're rounding to.
Re: (Score:3, Informative)
Yes, its almost 50% -- if you are, e.g., relating it to the nearest 25%. (Rounding it to the nearest 25% it would be just plain 50%, not "almost 50%".)
Its also almost 40% -- if you are, e.g., relating it to the nearest 10% (or 5% or 2%). And, in fact, 6.3/16 is also "almost 39.5%" if you are relating it to nearest 0.5%, and "just over 39%" if you are relating it to the nearest 1%.
"Almost" means you are giving an approximation (and the
Re: (Score:2)
When they say "news for nerds" they mean computer nerds, not math nerds.
Their new algorithm? (Score:5, Funny)
Re: (Score:3, Funny)
Re: (Score:2, Funny)
It is absolutely possible (Score:5, Informative)
Bitrate is only part of the equation -- the H.264 spec allows for a number of different ways to compress video, and it's up to the encoder to find out which is best for your video. Even in the same encoder, you can tweak dozens of settings in ways that dramatically change output quality -- usually a trade off between time and size.
x264 has beat every commercial encoder out there -- in some cases, on a level that would indeed render higher quality with half the bitrate.
Re: (Score:2)
x264 has beat every commercial encoder out there -- in some cases, on a level that would indeed render higher quality with half the bitrate.
Last I checked x264 was just on par or slightly below some commercial encoders with a standard profile. But x264 tends to have a bunch of OCD encoders who don't quit until they've tweaked it for just the right grain settings and tweaks for a given show or movie, which is usually what gives it the edge.
Re:It is absolutely possible (Score:5, Interesting)
It's always nice when free software solutions trash the commercial alternatives.
Crap HD Quality (Score:3, Interesting)
Try watching a football game here in the US and you will see what crap quality can be. The turf turns into squares of blur when the camera moves, then returns to blades of grass when the picture is stationary. As soon as you spot it you will hate it. If you don't see it then OK for you.
I used to have a friend who could spot the two little circles in the top right of a movie in the theater telling the projectionist to change the reel. Once he saw them the movies were never the same again.
Re:Crap HD Quality (Score:5, Informative)
I think you might want to talk to you cable company on that one. I know the effect you are seeing (it's by far the worst on local Public TV since they crammed 7 sub-channels into the same carrier), but network TV coverage of football in my area is pretty pristine for the most part. OTA is even better but cable is still awfully good.
Of course, by "talk to your cable company", I mean "do nothing" because talking to the cable company is a complete waste of time.
Brett
It depends on the material (Score:4, Informative)
Still a way to go... (Score:3, Informative)
iPlayer appears to use H.264 (Score:3, Insightful)
Re:iPlayer appears to use H.264 (Score:4, Informative)
BBC evil Jedi (Score:2, Funny)
BBC accountant: We provide the $ame or better picture quality with half the bitrate! Just think of the $aving$!
BBC IT decision maker: I $ee what you're $aying.... The$e picture$ look $uper!
Public: This looks like crap.
BBC rep: (waves hand) The$e aren't the compre$$ion artifact$ you're looking for. We can go about our bu$ine$$. There are no complaint$.
Re:BBC evil Jedi (Score:5, Funny)
Re: (Score:2)
Find one, refresh every 30 seconds and watch the incompetence. I've saw half a dozen edits in five minutes after the footy games last night.
Re: (Score:2)
Depends on the codec (Score:2, Redundant)
If they've switched from MPEG-2 to MPEG-4, then yes, you can get equal or better quality at a lower bitrate.
Test video (Score:5, Funny)
Quite a bit left out (Score:2)
Re: (Score:2)
BBC HD is a separate channel from normal BBC broadcasts, and uses H.264 rather than MPEG2.
The standard definition channels remain MPEG2 for compatibility.
Re: (Score:2)
Re: (Score:3, Informative)
Ummm .... (Score:2)
Is she saying that they've optimized their HD for people without HD screens, or am I just confused?
It really does sound like they're trying to sell something which isn't HD, but gets sold as if it is. Strange. "Now, to server you better, we are open fewer hours."
Cheers
A better model beats higher bitrate every time (Score:3, Insightful)
Lossy compression formats depend on an understanding of human perception. Nobody has a perfect model of the human brain, and nobody has a perfect algorithm for deciding what data to keep and what data to throw away.
If you have a better model of human perception than your competitors, then your encoder will yield higher quality output. If you spend 50% of your bits on things that nobody will notice, and I spend 25% of my bits on things that nobody will notice, then my 650kbps stream is going to look better than your 900kbps stream.
LAME did not win out as the MP3 encoder of choice just because it is free. It won out because its psychoacoustic model yields better-sounding MP3s at 128kbps than its competitors managed at 160kbps or even 192kbps.
spectrum (Score:2)
I suspect the move is connected with Freeview (a non-profit organisation, of which BBC is a member, that runs the free digital broadcasts) and the digital switchover. The BBC are probably thinking that if they cut their HD bitrate there can be more HD channels (and I assume more people would be able to get a good-enough signal). Or it just costs less.
Doctor Who hit hardest... (Score:3, Funny)
Now when we go to watch the Christmas special it will look like Cardboard sets, that wibble and wobble. The TARDIS will look utterly horrible and the Doctor will revert to a bloke from Liverpool with big eyes, big teeth, curly hair and a long scarf.
They also lowered the audio rate down to 16Kbps, so that rich orchestral music will sound like it came out of a cheap 1970's Moog.
Great, just when they updated the look of the show this will undo all of their work and it will look like the viewers were taken back to the 80's in an actual TARDIS.
Bravo BBC, Just Bravo......
better quality... (Score:3, Insightful)
Depending upon the configuration settings (frames per second, bit rate, I frame P frame structure etc.) it is most certainly possible to have a lower bit rate setting with better quality video than that of a higher bit rate setting. For example, if you drop the frame rate on a lower bitrate you often increase the quality of the video. So theoretically you can get easily the same quality at say 5 Mbps with a 15 fps as you can at 10 Mbps with 30 fps. I don't have specific numbers but subjectively (and empirically) it's quite possible.
There are definitely things that do make a difference here though, such as motion or high speed camera work. These types of video often suffer more noticeably than others so anyone watching sports, for example, will see the differences in quality more readily than someone watching a soap opera.
Re: (Score:2, Informative)
I'll toss FIOS under the bus too. Verizon's HD varies greatly. I'm not sure if its the channel companies themselves or Verizon doing it...
Either way, I hate watching fast motion movies or tv shows where the bitrate is too low.
Try watching "How its Made" on discovery HD and watch how compressed things look as fast moving manufactured parts pass through machinery.
Same for HBO films etc.
Re: (Score:2)
Try watching "How its Made" on discovery HD and watch how compressed things look as fast moving manufactured parts pass through machinery.
Well, there's your problem. Try watching The Woodwright's Shop [pbs.org] on PBS. Things move a bit slower there.