Chad wrote:

On 4/28/05, Chad <[EMAIL PROTECTED]> wrote:
I see posts and FAQ's and specs that display different numbers and I'm
wondering what one means over the other.

For example:

I have a Samsung DLP Monitor that has the following "specs":

It's a 61" 16x9 TV.

It has a resolution of 1280x720

It supports the following formats:  480i, 480p, 720p, 1080i


Any generic explaination would be really nice.

2 specific questions I have:
In the MythTV recording profiles, to 'get the most' out of my TV, what
resolution should I use to record?
For recording HDTV, you get the resolution in which the program is broadcast. Basically, HDTV tuners receive an MPEG-2 stream, so there's no processing done on that stream. This is ideal since it means the quality of the recording is the same as the quality of the broadcast. For a 720p broadcast, you get a 720p recording. For a 1080i broadcast, you get a 1080i recording. (Note that there are actually 18 formats defined in the ATSC specification, although the two just mentioned are the most commonly used. See http://www.hdtvprimer.com/ISSUES/what_is_ATSC.html for more info).

For recording SDTV, you have options. Since SDTV is an analog format, it specifies a number of lines (480 for NTSC). Each of these lines is analog (i.e. continuous), so there's no concept of horizontal resolution. However, to make the content digital, your tuner card must "sample" that analog line along its length, finding the color for a pixel at the position of the sample that most closely reproduces the original. As you might suspect, taking two samples along the horizontal axis would give you a very bad picture (i.e. trying to reproduce half the line with only one color would give you 2 pixels per line, so at 480 lines would be only 960 pixels--very bad quality). However, since the line is analog, we can sample it as many times as desired. The more samples, the better the quality--up to a certain point. Once we hit the ceiling, adding more pixels no longer improves picture quality (we basically plateau and quality remains the same--assuming we continue to increase the bitrate with the resolution). Therefore, the ideal resolution is the one at which you've hit the ceiling--it provides maximum quality at the minimum required bitrate.

The ceiling is different for different types of hardware. There are many factors involved in determining this maximum, but basically, you hit it when you push your hardware to its limit. For my PVR-{2,3}50's, I can't tell a difference in quality above 480x480 pixels resolution (assuming appropriate bitrate). However, I record at 720x480 simply because the DVD specification supports only 720x480, 704x480 (cut--i.e. no borders/rainbow edges on the video), 352x480, and 352x240 (for NTSC--the vertical numbers are 576 and 288 for PAL). Therefore, by recording at 720x480, I get the maximum picture quality (which I achieve way back at 480x480) and can burn to DVD without scaling/re-encoding. I have to use a slightly higher bitrate, but hard drives are cheap compared to the amount of time/effort required to scale the video after recording (even considering how few DVD's I make :).

What resolution should I have in my xorg.conf?
Ideally, since your TV has a native resolution of 1280x720, you would use that resolution. Then, your TV doesn't have to scale the video and you get a pixel-for-pixel representation of the output, giving you the highest quality possible. (Note that although some (older/cheaper) TV's have a native resolution of 1280x720, they may not be able to accept input at that resolution from a computer. If that's the case for yours, you would probably get the best results using the maximum resolution allowed. However, this is most likely not the case with your Samsung.)

Related question:
I have both VGA and DVI inputs on the TV.  I am wondering if running a
vga to vga cable from my videocard to the input will transfer an HD
stream, or if I HAVE to get a videocard with a DVI output (not that
it's that big of a deal, it'd just be nice to save ~30 bucks ;) ).
VGA cables can carry significantly higher resolutions than those required for HDTV output. However, VGA cables carry an analog signal. Since your computer's video card produces a digital image, it must use a digital to analog converter to send that image out the VGA port. In the case of a CRT, this is not a problem because the CRT also uses an analog signal, so there's one conversion that must happen anyway.

LCD monitors and digital TV's, on the other hand, need a digital signal. Therefore, if using a digital DVI connection (i.e. DVI-D or the digital side of DVI-I--which has both digital and analog signals), the video card outputs the image in digital and the display uses the exact image created by the video card. If, instead, you use a VGA connection to the LCD or digital TV, your video card does a digital to analog conversion, then the display does an analog to digital conversion. At this point, two conversions were performed on the signal--neither of which was required.

That being said, analog-to-digital and digital-to-analog conversions have been around for a long time, so the converters available today are typically very high quality (even the cheap ones). Therefore, it's quite possible that the image produced using a digital DVI connection will be indistinguishable from that produced using an analog connection (especially for relatively short cable runs). As a matter of fact, for this reason, some LCD/TV manufacturers--needing the ADC to support VGA connections, anyway--decided to use DVI-A (analog) or to use the analog side of a DVI-I connection to save money by providing a single internal path for the signal (whether from VGA or DVI inputs). The DVI-A signal is basically the same as the VGA signal--you can actually convert from one to the other using only a cable (for $20 or so)--whereas converting from DVI-D to VGA requires a digital to analog converter (DAC) like this one for $282 ( http://www.networktechinc.com/dvi-vga.html ).

Therefore, if your TV is using the analog side of the DVI input, there's absolutely no benefit to using DVI instead of VGA. However, if your TV is using the digital side of the DVI input, there may be a benefit to using DVI (although, that benefit may simply be the placebo effect... ;).

Basically, when you boil it down to basics, the best answer is, "If it looks good to you, it's good."

Oh, and one last bit of info. Since the DVI-D connection is digital, any relatively new TV with DVI-D connection almost definitely has HDCP (High-bandwidth Digital Content Protection) built into that connection. (But, since VGA connections are analog, they are not (yet) required to be protected with HDCP or other DRM.) While it probably doesn't matter now, I'll leave it to you to decide whether that's good or bad for you in the future... ;)

Thanks!!
Sorry, don't know how the original got lost ;)
Since I'm the one who gave you a hard time about the lost question, I figured I should at least take a stab at answering... :)

Thanks for the replies!

Hope it helps.

Mike
_______________________________________________
mythtv-users mailing list
[email protected]
http://mythtv.org/cgi-bin/mailman/listinfo/mythtv-users

Reply via email to