What is up with screen resolutions? [closed]
Clash Royale CLAN TAG#URR8PPP
up vote
17
down vote
favorite
I thought, it is determined by the horizontal pixel width, but why is 1920*1080 1080p but 1280*720 720p? It makes kinda no sense or is my source wrong?
display tv
closed as off-topic by Máté Juhász, JakeGould, robinCTS, Pimp Juice IT, bertieb Aug 20 at 10:12
This question appears to be off-topic. The users who voted to close gave this specific reason:
- "This question is not about computer hardware or software, within the scope defined in the help center." â Máté Juhász, JakeGould, robinCTS, Pimp Juice IT
 |Â
show 3 more comments
up vote
17
down vote
favorite
I thought, it is determined by the horizontal pixel width, but why is 1920*1080 1080p but 1280*720 720p? It makes kinda no sense or is my source wrong?
display tv
closed as off-topic by Máté Juhász, JakeGould, robinCTS, Pimp Juice IT, bertieb Aug 20 at 10:12
This question appears to be off-topic. The users who voted to close gave this specific reason:
- "This question is not about computer hardware or software, within the scope defined in the help center." â Máté Juhász, JakeGould, robinCTS, Pimp Juice IT
11
You thought what is determined by horizontal width? Screen resolution is width by height. Multiply one by the other to get "K" [in the broadest terms]. 720p & 1080p are rendering standards, not resolution standards. The 'p' means progressive scan, to differentiate from the inferior 'i' for interleaved scan.
â Tetsujin
Aug 18 at 16:53
23
It's an idiotic change made by manufacturers so that the latest jump to "4k" sounds like a huge deal rather than the on going incremental change that it is. Up to 4k we were naming the resolution by vertical resolution and then suddenly with 4k they decided to use the horizontal width instead.
â Mokubaiâ¦
Aug 18 at 17:00
4
Look at this answer I posted here.
â JakeGould
Aug 18 at 17:32
4
@Mokubai - yeah the next step is for them to go all hard drive on us and we'll lose the last 96 pixels too.
â davidbak
Aug 18 at 22:45
5
I'm very confused because I don't see a difference between the two examples of 1920*1080 and 1280*720. The "p" number is the second number in both cases. Maybe the "but" should be an "and", because, as written, it reads very odd to me.
â Greg Schmit
Aug 19 at 23:39
 |Â
show 3 more comments
up vote
17
down vote
favorite
up vote
17
down vote
favorite
I thought, it is determined by the horizontal pixel width, but why is 1920*1080 1080p but 1280*720 720p? It makes kinda no sense or is my source wrong?
display tv
I thought, it is determined by the horizontal pixel width, but why is 1920*1080 1080p but 1280*720 720p? It makes kinda no sense or is my source wrong?
display tv
edited Aug 19 at 20:58
WELZ
121128
121128
asked Aug 18 at 16:49
Maritn Ge
8313
8313
closed as off-topic by Máté Juhász, JakeGould, robinCTS, Pimp Juice IT, bertieb Aug 20 at 10:12
This question appears to be off-topic. The users who voted to close gave this specific reason:
- "This question is not about computer hardware or software, within the scope defined in the help center." â Máté Juhász, JakeGould, robinCTS, Pimp Juice IT
closed as off-topic by Máté Juhász, JakeGould, robinCTS, Pimp Juice IT, bertieb Aug 20 at 10:12
This question appears to be off-topic. The users who voted to close gave this specific reason:
- "This question is not about computer hardware or software, within the scope defined in the help center." â Máté Juhász, JakeGould, robinCTS, Pimp Juice IT
11
You thought what is determined by horizontal width? Screen resolution is width by height. Multiply one by the other to get "K" [in the broadest terms]. 720p & 1080p are rendering standards, not resolution standards. The 'p' means progressive scan, to differentiate from the inferior 'i' for interleaved scan.
â Tetsujin
Aug 18 at 16:53
23
It's an idiotic change made by manufacturers so that the latest jump to "4k" sounds like a huge deal rather than the on going incremental change that it is. Up to 4k we were naming the resolution by vertical resolution and then suddenly with 4k they decided to use the horizontal width instead.
â Mokubaiâ¦
Aug 18 at 17:00
4
Look at this answer I posted here.
â JakeGould
Aug 18 at 17:32
4
@Mokubai - yeah the next step is for them to go all hard drive on us and we'll lose the last 96 pixels too.
â davidbak
Aug 18 at 22:45
5
I'm very confused because I don't see a difference between the two examples of 1920*1080 and 1280*720. The "p" number is the second number in both cases. Maybe the "but" should be an "and", because, as written, it reads very odd to me.
â Greg Schmit
Aug 19 at 23:39
 |Â
show 3 more comments
11
You thought what is determined by horizontal width? Screen resolution is width by height. Multiply one by the other to get "K" [in the broadest terms]. 720p & 1080p are rendering standards, not resolution standards. The 'p' means progressive scan, to differentiate from the inferior 'i' for interleaved scan.
â Tetsujin
Aug 18 at 16:53
23
It's an idiotic change made by manufacturers so that the latest jump to "4k" sounds like a huge deal rather than the on going incremental change that it is. Up to 4k we were naming the resolution by vertical resolution and then suddenly with 4k they decided to use the horizontal width instead.
â Mokubaiâ¦
Aug 18 at 17:00
4
Look at this answer I posted here.
â JakeGould
Aug 18 at 17:32
4
@Mokubai - yeah the next step is for them to go all hard drive on us and we'll lose the last 96 pixels too.
â davidbak
Aug 18 at 22:45
5
I'm very confused because I don't see a difference between the two examples of 1920*1080 and 1280*720. The "p" number is the second number in both cases. Maybe the "but" should be an "and", because, as written, it reads very odd to me.
â Greg Schmit
Aug 19 at 23:39
11
11
You thought what is determined by horizontal width? Screen resolution is width by height. Multiply one by the other to get "K" [in the broadest terms]. 720p & 1080p are rendering standards, not resolution standards. The 'p' means progressive scan, to differentiate from the inferior 'i' for interleaved scan.
â Tetsujin
Aug 18 at 16:53
You thought what is determined by horizontal width? Screen resolution is width by height. Multiply one by the other to get "K" [in the broadest terms]. 720p & 1080p are rendering standards, not resolution standards. The 'p' means progressive scan, to differentiate from the inferior 'i' for interleaved scan.
â Tetsujin
Aug 18 at 16:53
23
23
It's an idiotic change made by manufacturers so that the latest jump to "4k" sounds like a huge deal rather than the on going incremental change that it is. Up to 4k we were naming the resolution by vertical resolution and then suddenly with 4k they decided to use the horizontal width instead.
â Mokubaiâ¦
Aug 18 at 17:00
It's an idiotic change made by manufacturers so that the latest jump to "4k" sounds like a huge deal rather than the on going incremental change that it is. Up to 4k we were naming the resolution by vertical resolution and then suddenly with 4k they decided to use the horizontal width instead.
â Mokubaiâ¦
Aug 18 at 17:00
4
4
Look at this answer I posted here.
â JakeGould
Aug 18 at 17:32
Look at this answer I posted here.
â JakeGould
Aug 18 at 17:32
4
4
@Mokubai - yeah the next step is for them to go all hard drive on us and we'll lose the last 96 pixels too.
â davidbak
Aug 18 at 22:45
@Mokubai - yeah the next step is for them to go all hard drive on us and we'll lose the last 96 pixels too.
â davidbak
Aug 18 at 22:45
5
5
I'm very confused because I don't see a difference between the two examples of 1920*1080 and 1280*720. The "p" number is the second number in both cases. Maybe the "but" should be an "and", because, as written, it reads very odd to me.
â Greg Schmit
Aug 19 at 23:39
I'm very confused because I don't see a difference between the two examples of 1920*1080 and 1280*720. The "p" number is the second number in both cases. Maybe the "but" should be an "and", because, as written, it reads very odd to me.
â Greg Schmit
Aug 19 at 23:39
 |Â
show 3 more comments
3 Answers
3
active
oldest
votes
up vote
44
down vote
In the "old days", TVs were analog devices - a CRT scans an electron beam across the front of the display, from left to right. In this sense, people have tried to argue that analog TV has "infinite" horizontal resolution, but they have an exact vertical resolution - the picture is formed of multiple horizontal lines.
Depending on where you are, this would have been either NTSC (525 lines) or PAL (625 lines).
Modern resolutions are still referred to by their "line count" - hence 1080 is the vertical resolution.
With such displays, the image was transmitted interlaced - i.e: the first field contains lines 1, 3, 5, etc... and the second field contains lines 2, 4, 6, etc...
With the advent of digital TV, the display technology changes - we now have discrete pixels instead of lines. A 2D image made from an array of pixels - the array having specific dimensions in the horizontal and vertical axes.
At this point, interlacing remains around only to reduce the bandwidth required for a video stream - in my opinion it's a horrible idea that is fundamentally incompatible with digital display systems.
As mentioned earlier, modern resolutions are still referred to by their "line count". But as shown above, we also have to identify whether we are referring to "interlaced" video (indicated by an i
), or "progressive" video (indicated by a p
).
The frame rate can also be specified, for example:
480i60
- 480 rows, interlaced, 60 Hz field rate (i.e: 30 Hz frame rate)1080p30
- 1080 rows, progressive, 30 Hz frame rate
"... okay, but where did 480 come from?"
The analog electronics involved in CRTs are imprecise, and a particular feature of early models was that as the set warmed up, or the capacitors and electronics aged, the image started to change shape. In addition to this, the electron beam has to be turned off and then redirected to the left of the screen for each line, and to the top for each new field/frame - this takes time, and is the reason for "blanking".
To account for this somewhat, only a handful of the scanned lines were intended / expected to be displayed. NTSC scans 525 lines, but only 483 if these are intended for display - each field shows ~241.5 lines to a viewer (with an additional 21 lines of "blanking"). ~240 lines per field (remember, this is interlaced) equates to 480 lines per frame (in the progressive world). Thus 480. Yay.
For the digital resolutions we follow a pattern... ish:
480 * 1.5 = 720
- "HD Ready"720 * 1.5 = 1080
- "Full HD"1080 * 2 = 2160
- "4k" or "Ultra HD"
So actually "4k" isn't following the * 1.5
pattern from before, and it's not really 4000 pixels in either dimension - it's 3840 ÃÂ 2160.
"4k" is actually "Four times the number of pixels as Full HD". 1920 * 2 = 3840
and 1080 * 2 = 2160
. Or lay out four 1080p displays in a 2 ÃÂ 2 grid - you get 3840 ÃÂ 2160.
Additionally, if we're using 1080p
as a resolution description, then really "4k" should be called 2160p
(which it is in the technical world).
In summary, in the consumer / broadcast space:
480
is because that's approximately the number of visible lines that NTSC would display720
is because that's 1.5ÃÂ480
1080
is because that's 1.5ÃÂ720
2160
is because that's 2ÃÂ1080
- "4k" is because it's a marketing thing - it isn't a technical specification, and I don't believe that there are any stipulations around the use of it...
Note: I've been taking about consumer / broadcast...
Within Cinema there is DCI 2K (capital K
, 2048 ÃÂ 1080) and DCI 4K (capital K
, 4096 ÃÂ 2160), where the 'K' presumably refers to kibi and the horizontal resolution. "DCI 4K" predates consumer "4k".
The aspect ratio of DCI 4K is not 16:9, but a slightly wider 256â¶135... to bring the resolution back in line with 16:9, you can increase the vertical resolution, or decrease the horizontal... But I'm not entirely convinced that the cinema and broadcast standards evolved this closely.
Cinema evolved from whole-frame positives (aka film) directly to digital, whereas TV evolved from a scanning electron beam (line by line) to digital. This is evident in both broadcast, and the way that VHS operates too.
As a little extra, I've included the graphic below to expand on the "image changes shape" statement from above.
The graphic (from here) indicates the various "television safe" areas... note the rounded corners...
Importantly:
- 5 is the "television scanned area"
- 6 is the "television safe area for action"
- Faces and important plot information should not fall outside this area
- 7 is the "television safe area for titles"
- No text should be outside this area to ensure that subtitles / news tickers / etc... don't overhang the edge of an old display
2
480i30 would be a weirdly low frame rate; much more common is 480i60, 60 fields per second (can be combed into 30 frames per second). You can deinterlace it to 480p60 (doubling the vertical spatial resolution) or 480p30 (discarding half the temporal resolution). Agreed that interlacing is stupidly horrible now that progressive displays are nearly universal, and it always sucked to store digitally.
â Peter Cordes
Aug 18 at 19:20
@Peter thanks - I couldn't remember if the number was field rate or frame rate...
â Attie
Aug 18 at 20:10
1
The rate number is the temporal resolution. Number of times the display changes (or could change) per second. I think this sometimes gets fudged with progressive content stored as interlaced (evil, bad, and stupid), where both fields are actually from the same time, rather than evenly spaced times. (And of course there's 24p content hard-telecined to 60i, or worse telecined to 60i and then having VFX added at 60i or 30p, like the Star Trek DVDs before they remastered. Those DVDs switched from soft-TC 24p for most scenes to 60i for VFX shots, because the show was edited on video.)
â Peter Cordes
Aug 18 at 20:26
2
There are actually a few different resolutions that have been called "4K". The first usage however (according to Wikipedia) was a camera using 4096Ã2160. Someone thought 4096 was pretty close to 4000, and decided that saying their camera had a 4K resolution sounded cool. It's really just marketing.
â Yay295
Aug 19 at 3:47
1
It should also be noted that 480i or 480p use 720 horizontal pixels with a width to height aspect ratio of 8/9 (= .88888....), which creates an interpolation issue on digital monitors. For CRT displays, this isn't an issue because it just involves a change in the sweep rate (and effective thickness) of the beam used to light up the phosphors. CRT based HDTV's generally support 480i / 480p / 1080i as "native" resolutions, with a better image for 480i/480p than digital displays, but not as good as 1080p digital displays.
â rcgldr
Aug 19 at 21:17
 |Â
show 1 more comment
up vote
7
down vote
Attie has explained how vertical pixel counts were derived for traditional resolutions such as SD (480p), HD (720p) and Full HD (1080p). I would like to discuss K measurements which originate from the cinema industry. A K is about 1000 horizontal pixels and 2K and 4K were casual terms referring to resolutions of approximately 2000 and 4000 horizontal pixels.
Different K measurements were later standardised for digital cinema (2005) and consumer television (2007).
Digital Cinema Initiatives has specified the DCI 4K and 2K resoultion standards (4096âÂÂÃÂâÂÂ2160 and 2048 à1080). These are the full frame resolutions for digital cinema projectors, with many films displayed in the a crop format such as flat crop (3996âÂÂÃÂâÂÂ2160 , 1.85â¶1 aspect ratio) or CinemaScope crop (4096âÂÂÃÂâÂÂ1716, âÂÂ2.39â¶1 aspect ratio).
The 4K standard for consumer television (UHDTV1) can be more accurately labelled as Ultra HD or Quad HD. It is equivalent to 4 Full HD frames placed together. It is 3840âÂÂÃÂâÂÂ2160 with âÂÂ1.78â¶1 aspect ratio.
The Wikipedia article on 4K resolution provides a helpful visualisation which compares different frame sizes.
add a comment |Â
up vote
3
down vote
Your assessment that resolution is pixel pitch is kind of correct. In fact, for everything except digital displays, that's the correct definition of resolution (well, more accurately, it's the smallest detail that can be resolved accurately, which functionally translates to the size of the dots making up an image, and from that pixel pitch). If you ever talk to somebody in the printing industry, this is usually what they mean when they say 'resolution'.
The reason that the term has a different meaning for computer displays is largely historical, but is actually pretty easily explained without getting into much history.
On a raster display1, you have what's known as a 'native' resolution. This is the highest number of pixels you can have on each side of the screen, and is usually what is used to describe the screen. For example, a Full HD display has a native resolution of 1920 pixels horizontally, and 1080 vertically, so a Full HD display with a roughly 18 inch diagonal has a resolution (in the classical sense) of about 120 dots per inch (DPI) (which is actually pretty low compared to print media).
However, you can run most displays at lower than their native resolution, which is functionally equivalent to having a larger pixel pitch on the same display. Taking that same 18 inch Full HD display and running it at 1280x720 gives you the equivalent of 80 DPI on the same screen.
Now, the OS and application software (usually) don't care about the exact physical dimensions of the screen, because that information is not really all that useful to displaying information unless you need to display something 'real-size'. Given this, the OS just assumes the size of the display never changes, which means that the different pixel counts are functionally equivalent to the resolution.
- A display that uses rows and columns of individual points that can be turned on and off to produce an image. Almost all modern computer displays are raster displays, because they're fare easier to make. Compare to a vector display, which just draws lines directly (with one of the most recognizable examples being the displays used in the original Asteroids arcade cabinets).
1
"Resolution" doesn't have any bearing on dot size, pitch or aspect ratio as such - it refers to the smallest addressable "thing" - DAC resolution of 16-bits, screen resolution, etc... interestingly with RGB displays, you quite whole pixels, not sub pixels. With print it's quite meaningless to advertise a resolution across a page due to various sizes (A4 / A5 / etc...), margins, etc... so instead they specify the area (typically an Inch) and count the dots inside it (DPI)
â Attie
Aug 19 at 10:34
@Attie No, it does have bearing on dot size and/or pixel pitch (though not really aspect ratio). If you have a display of a given size, higher resolution means smaller pixel pitch, period (and thus higher detail and less aliasing). The reason resolution matters for actually displaying anything when dealing with anything higher level than drivers is pixel pitch and pixel count (which is a side effect of pixel pitch and display size), not addressing. Also, the use of dots-per-unit-area in print far predates computer displays.
â Austin Hemmelgarn
Aug 19 at 12:55
I completely agree that "resolution", "pixel pitch" and "physical dimensions" are all intertwined (how could they not be)... but "resolution" by itself has no bearing on the other two. The fact that software is typically aware of a display's physical dimensions or PPI means that it's possible to display things at approximately real size.
â Attie
Aug 19 at 13:54
As for DPI in print and pixel aspect ratio - until recently pixels were not square! IIRC facsimile had pixels with a 2:1 aspect ratio, and NTSC was 10:11... rendering the image with an incorrect PAR produces squashed/stretched looking images. Additionally, DPI in print usually needs to be so high because of the limited count of discernible colors - dithering coming to the rescue.
â Attie
Aug 19 at 14:00
add a comment |Â
3 Answers
3
active
oldest
votes
3 Answers
3
active
oldest
votes
active
oldest
votes
active
oldest
votes
up vote
44
down vote
In the "old days", TVs were analog devices - a CRT scans an electron beam across the front of the display, from left to right. In this sense, people have tried to argue that analog TV has "infinite" horizontal resolution, but they have an exact vertical resolution - the picture is formed of multiple horizontal lines.
Depending on where you are, this would have been either NTSC (525 lines) or PAL (625 lines).
Modern resolutions are still referred to by their "line count" - hence 1080 is the vertical resolution.
With such displays, the image was transmitted interlaced - i.e: the first field contains lines 1, 3, 5, etc... and the second field contains lines 2, 4, 6, etc...
With the advent of digital TV, the display technology changes - we now have discrete pixels instead of lines. A 2D image made from an array of pixels - the array having specific dimensions in the horizontal and vertical axes.
At this point, interlacing remains around only to reduce the bandwidth required for a video stream - in my opinion it's a horrible idea that is fundamentally incompatible with digital display systems.
As mentioned earlier, modern resolutions are still referred to by their "line count". But as shown above, we also have to identify whether we are referring to "interlaced" video (indicated by an i
), or "progressive" video (indicated by a p
).
The frame rate can also be specified, for example:
480i60
- 480 rows, interlaced, 60 Hz field rate (i.e: 30 Hz frame rate)1080p30
- 1080 rows, progressive, 30 Hz frame rate
"... okay, but where did 480 come from?"
The analog electronics involved in CRTs are imprecise, and a particular feature of early models was that as the set warmed up, or the capacitors and electronics aged, the image started to change shape. In addition to this, the electron beam has to be turned off and then redirected to the left of the screen for each line, and to the top for each new field/frame - this takes time, and is the reason for "blanking".
To account for this somewhat, only a handful of the scanned lines were intended / expected to be displayed. NTSC scans 525 lines, but only 483 if these are intended for display - each field shows ~241.5 lines to a viewer (with an additional 21 lines of "blanking"). ~240 lines per field (remember, this is interlaced) equates to 480 lines per frame (in the progressive world). Thus 480. Yay.
For the digital resolutions we follow a pattern... ish:
480 * 1.5 = 720
- "HD Ready"720 * 1.5 = 1080
- "Full HD"1080 * 2 = 2160
- "4k" or "Ultra HD"
So actually "4k" isn't following the * 1.5
pattern from before, and it's not really 4000 pixels in either dimension - it's 3840 ÃÂ 2160.
"4k" is actually "Four times the number of pixels as Full HD". 1920 * 2 = 3840
and 1080 * 2 = 2160
. Or lay out four 1080p displays in a 2 ÃÂ 2 grid - you get 3840 ÃÂ 2160.
Additionally, if we're using 1080p
as a resolution description, then really "4k" should be called 2160p
(which it is in the technical world).
In summary, in the consumer / broadcast space:
480
is because that's approximately the number of visible lines that NTSC would display720
is because that's 1.5ÃÂ480
1080
is because that's 1.5ÃÂ720
2160
is because that's 2ÃÂ1080
- "4k" is because it's a marketing thing - it isn't a technical specification, and I don't believe that there are any stipulations around the use of it...
Note: I've been taking about consumer / broadcast...
Within Cinema there is DCI 2K (capital K
, 2048 ÃÂ 1080) and DCI 4K (capital K
, 4096 ÃÂ 2160), where the 'K' presumably refers to kibi and the horizontal resolution. "DCI 4K" predates consumer "4k".
The aspect ratio of DCI 4K is not 16:9, but a slightly wider 256â¶135... to bring the resolution back in line with 16:9, you can increase the vertical resolution, or decrease the horizontal... But I'm not entirely convinced that the cinema and broadcast standards evolved this closely.
Cinema evolved from whole-frame positives (aka film) directly to digital, whereas TV evolved from a scanning electron beam (line by line) to digital. This is evident in both broadcast, and the way that VHS operates too.
As a little extra, I've included the graphic below to expand on the "image changes shape" statement from above.
The graphic (from here) indicates the various "television safe" areas... note the rounded corners...
Importantly:
- 5 is the "television scanned area"
- 6 is the "television safe area for action"
- Faces and important plot information should not fall outside this area
- 7 is the "television safe area for titles"
- No text should be outside this area to ensure that subtitles / news tickers / etc... don't overhang the edge of an old display
2
480i30 would be a weirdly low frame rate; much more common is 480i60, 60 fields per second (can be combed into 30 frames per second). You can deinterlace it to 480p60 (doubling the vertical spatial resolution) or 480p30 (discarding half the temporal resolution). Agreed that interlacing is stupidly horrible now that progressive displays are nearly universal, and it always sucked to store digitally.
â Peter Cordes
Aug 18 at 19:20
@Peter thanks - I couldn't remember if the number was field rate or frame rate...
â Attie
Aug 18 at 20:10
1
The rate number is the temporal resolution. Number of times the display changes (or could change) per second. I think this sometimes gets fudged with progressive content stored as interlaced (evil, bad, and stupid), where both fields are actually from the same time, rather than evenly spaced times. (And of course there's 24p content hard-telecined to 60i, or worse telecined to 60i and then having VFX added at 60i or 30p, like the Star Trek DVDs before they remastered. Those DVDs switched from soft-TC 24p for most scenes to 60i for VFX shots, because the show was edited on video.)
â Peter Cordes
Aug 18 at 20:26
2
There are actually a few different resolutions that have been called "4K". The first usage however (according to Wikipedia) was a camera using 4096Ã2160. Someone thought 4096 was pretty close to 4000, and decided that saying their camera had a 4K resolution sounded cool. It's really just marketing.
â Yay295
Aug 19 at 3:47
1
It should also be noted that 480i or 480p use 720 horizontal pixels with a width to height aspect ratio of 8/9 (= .88888....), which creates an interpolation issue on digital monitors. For CRT displays, this isn't an issue because it just involves a change in the sweep rate (and effective thickness) of the beam used to light up the phosphors. CRT based HDTV's generally support 480i / 480p / 1080i as "native" resolutions, with a better image for 480i/480p than digital displays, but not as good as 1080p digital displays.
â rcgldr
Aug 19 at 21:17
 |Â
show 1 more comment
up vote
44
down vote
In the "old days", TVs were analog devices - a CRT scans an electron beam across the front of the display, from left to right. In this sense, people have tried to argue that analog TV has "infinite" horizontal resolution, but they have an exact vertical resolution - the picture is formed of multiple horizontal lines.
Depending on where you are, this would have been either NTSC (525 lines) or PAL (625 lines).
Modern resolutions are still referred to by their "line count" - hence 1080 is the vertical resolution.
With such displays, the image was transmitted interlaced - i.e: the first field contains lines 1, 3, 5, etc... and the second field contains lines 2, 4, 6, etc...
With the advent of digital TV, the display technology changes - we now have discrete pixels instead of lines. A 2D image made from an array of pixels - the array having specific dimensions in the horizontal and vertical axes.
At this point, interlacing remains around only to reduce the bandwidth required for a video stream - in my opinion it's a horrible idea that is fundamentally incompatible with digital display systems.
As mentioned earlier, modern resolutions are still referred to by their "line count". But as shown above, we also have to identify whether we are referring to "interlaced" video (indicated by an i
), or "progressive" video (indicated by a p
).
The frame rate can also be specified, for example:
480i60
- 480 rows, interlaced, 60 Hz field rate (i.e: 30 Hz frame rate)1080p30
- 1080 rows, progressive, 30 Hz frame rate
"... okay, but where did 480 come from?"
The analog electronics involved in CRTs are imprecise, and a particular feature of early models was that as the set warmed up, or the capacitors and electronics aged, the image started to change shape. In addition to this, the electron beam has to be turned off and then redirected to the left of the screen for each line, and to the top for each new field/frame - this takes time, and is the reason for "blanking".
To account for this somewhat, only a handful of the scanned lines were intended / expected to be displayed. NTSC scans 525 lines, but only 483 if these are intended for display - each field shows ~241.5 lines to a viewer (with an additional 21 lines of "blanking"). ~240 lines per field (remember, this is interlaced) equates to 480 lines per frame (in the progressive world). Thus 480. Yay.
For the digital resolutions we follow a pattern... ish:
480 * 1.5 = 720
- "HD Ready"720 * 1.5 = 1080
- "Full HD"1080 * 2 = 2160
- "4k" or "Ultra HD"
So actually "4k" isn't following the * 1.5
pattern from before, and it's not really 4000 pixels in either dimension - it's 3840 ÃÂ 2160.
"4k" is actually "Four times the number of pixels as Full HD". 1920 * 2 = 3840
and 1080 * 2 = 2160
. Or lay out four 1080p displays in a 2 ÃÂ 2 grid - you get 3840 ÃÂ 2160.
Additionally, if we're using 1080p
as a resolution description, then really "4k" should be called 2160p
(which it is in the technical world).
In summary, in the consumer / broadcast space:
480
is because that's approximately the number of visible lines that NTSC would display720
is because that's 1.5ÃÂ480
1080
is because that's 1.5ÃÂ720
2160
is because that's 2ÃÂ1080
- "4k" is because it's a marketing thing - it isn't a technical specification, and I don't believe that there are any stipulations around the use of it...
Note: I've been taking about consumer / broadcast...
Within Cinema there is DCI 2K (capital K
, 2048 ÃÂ 1080) and DCI 4K (capital K
, 4096 ÃÂ 2160), where the 'K' presumably refers to kibi and the horizontal resolution. "DCI 4K" predates consumer "4k".
The aspect ratio of DCI 4K is not 16:9, but a slightly wider 256â¶135... to bring the resolution back in line with 16:9, you can increase the vertical resolution, or decrease the horizontal... But I'm not entirely convinced that the cinema and broadcast standards evolved this closely.
Cinema evolved from whole-frame positives (aka film) directly to digital, whereas TV evolved from a scanning electron beam (line by line) to digital. This is evident in both broadcast, and the way that VHS operates too.
As a little extra, I've included the graphic below to expand on the "image changes shape" statement from above.
The graphic (from here) indicates the various "television safe" areas... note the rounded corners...
Importantly:
- 5 is the "television scanned area"
- 6 is the "television safe area for action"
- Faces and important plot information should not fall outside this area
- 7 is the "television safe area for titles"
- No text should be outside this area to ensure that subtitles / news tickers / etc... don't overhang the edge of an old display
2
480i30 would be a weirdly low frame rate; much more common is 480i60, 60 fields per second (can be combed into 30 frames per second). You can deinterlace it to 480p60 (doubling the vertical spatial resolution) or 480p30 (discarding half the temporal resolution). Agreed that interlacing is stupidly horrible now that progressive displays are nearly universal, and it always sucked to store digitally.
â Peter Cordes
Aug 18 at 19:20
@Peter thanks - I couldn't remember if the number was field rate or frame rate...
â Attie
Aug 18 at 20:10
1
The rate number is the temporal resolution. Number of times the display changes (or could change) per second. I think this sometimes gets fudged with progressive content stored as interlaced (evil, bad, and stupid), where both fields are actually from the same time, rather than evenly spaced times. (And of course there's 24p content hard-telecined to 60i, or worse telecined to 60i and then having VFX added at 60i or 30p, like the Star Trek DVDs before they remastered. Those DVDs switched from soft-TC 24p for most scenes to 60i for VFX shots, because the show was edited on video.)
â Peter Cordes
Aug 18 at 20:26
2
There are actually a few different resolutions that have been called "4K". The first usage however (according to Wikipedia) was a camera using 4096Ã2160. Someone thought 4096 was pretty close to 4000, and decided that saying their camera had a 4K resolution sounded cool. It's really just marketing.
â Yay295
Aug 19 at 3:47
1
It should also be noted that 480i or 480p use 720 horizontal pixels with a width to height aspect ratio of 8/9 (= .88888....), which creates an interpolation issue on digital monitors. For CRT displays, this isn't an issue because it just involves a change in the sweep rate (and effective thickness) of the beam used to light up the phosphors. CRT based HDTV's generally support 480i / 480p / 1080i as "native" resolutions, with a better image for 480i/480p than digital displays, but not as good as 1080p digital displays.
â rcgldr
Aug 19 at 21:17
 |Â
show 1 more comment
up vote
44
down vote
up vote
44
down vote
In the "old days", TVs were analog devices - a CRT scans an electron beam across the front of the display, from left to right. In this sense, people have tried to argue that analog TV has "infinite" horizontal resolution, but they have an exact vertical resolution - the picture is formed of multiple horizontal lines.
Depending on where you are, this would have been either NTSC (525 lines) or PAL (625 lines).
Modern resolutions are still referred to by their "line count" - hence 1080 is the vertical resolution.
With such displays, the image was transmitted interlaced - i.e: the first field contains lines 1, 3, 5, etc... and the second field contains lines 2, 4, 6, etc...
With the advent of digital TV, the display technology changes - we now have discrete pixels instead of lines. A 2D image made from an array of pixels - the array having specific dimensions in the horizontal and vertical axes.
At this point, interlacing remains around only to reduce the bandwidth required for a video stream - in my opinion it's a horrible idea that is fundamentally incompatible with digital display systems.
As mentioned earlier, modern resolutions are still referred to by their "line count". But as shown above, we also have to identify whether we are referring to "interlaced" video (indicated by an i
), or "progressive" video (indicated by a p
).
The frame rate can also be specified, for example:
480i60
- 480 rows, interlaced, 60 Hz field rate (i.e: 30 Hz frame rate)1080p30
- 1080 rows, progressive, 30 Hz frame rate
"... okay, but where did 480 come from?"
The analog electronics involved in CRTs are imprecise, and a particular feature of early models was that as the set warmed up, or the capacitors and electronics aged, the image started to change shape. In addition to this, the electron beam has to be turned off and then redirected to the left of the screen for each line, and to the top for each new field/frame - this takes time, and is the reason for "blanking".
To account for this somewhat, only a handful of the scanned lines were intended / expected to be displayed. NTSC scans 525 lines, but only 483 if these are intended for display - each field shows ~241.5 lines to a viewer (with an additional 21 lines of "blanking"). ~240 lines per field (remember, this is interlaced) equates to 480 lines per frame (in the progressive world). Thus 480. Yay.
For the digital resolutions we follow a pattern... ish:
480 * 1.5 = 720
- "HD Ready"720 * 1.5 = 1080
- "Full HD"1080 * 2 = 2160
- "4k" or "Ultra HD"
So actually "4k" isn't following the * 1.5
pattern from before, and it's not really 4000 pixels in either dimension - it's 3840 ÃÂ 2160.
"4k" is actually "Four times the number of pixels as Full HD". 1920 * 2 = 3840
and 1080 * 2 = 2160
. Or lay out four 1080p displays in a 2 ÃÂ 2 grid - you get 3840 ÃÂ 2160.
Additionally, if we're using 1080p
as a resolution description, then really "4k" should be called 2160p
(which it is in the technical world).
In summary, in the consumer / broadcast space:
480
is because that's approximately the number of visible lines that NTSC would display720
is because that's 1.5ÃÂ480
1080
is because that's 1.5ÃÂ720
2160
is because that's 2ÃÂ1080
- "4k" is because it's a marketing thing - it isn't a technical specification, and I don't believe that there are any stipulations around the use of it...
Note: I've been taking about consumer / broadcast...
Within Cinema there is DCI 2K (capital K
, 2048 ÃÂ 1080) and DCI 4K (capital K
, 4096 ÃÂ 2160), where the 'K' presumably refers to kibi and the horizontal resolution. "DCI 4K" predates consumer "4k".
The aspect ratio of DCI 4K is not 16:9, but a slightly wider 256â¶135... to bring the resolution back in line with 16:9, you can increase the vertical resolution, or decrease the horizontal... But I'm not entirely convinced that the cinema and broadcast standards evolved this closely.
Cinema evolved from whole-frame positives (aka film) directly to digital, whereas TV evolved from a scanning electron beam (line by line) to digital. This is evident in both broadcast, and the way that VHS operates too.
As a little extra, I've included the graphic below to expand on the "image changes shape" statement from above.
The graphic (from here) indicates the various "television safe" areas... note the rounded corners...
Importantly:
- 5 is the "television scanned area"
- 6 is the "television safe area for action"
- Faces and important plot information should not fall outside this area
- 7 is the "television safe area for titles"
- No text should be outside this area to ensure that subtitles / news tickers / etc... don't overhang the edge of an old display
In the "old days", TVs were analog devices - a CRT scans an electron beam across the front of the display, from left to right. In this sense, people have tried to argue that analog TV has "infinite" horizontal resolution, but they have an exact vertical resolution - the picture is formed of multiple horizontal lines.
Depending on where you are, this would have been either NTSC (525 lines) or PAL (625 lines).
Modern resolutions are still referred to by their "line count" - hence 1080 is the vertical resolution.
With such displays, the image was transmitted interlaced - i.e: the first field contains lines 1, 3, 5, etc... and the second field contains lines 2, 4, 6, etc...
With the advent of digital TV, the display technology changes - we now have discrete pixels instead of lines. A 2D image made from an array of pixels - the array having specific dimensions in the horizontal and vertical axes.
At this point, interlacing remains around only to reduce the bandwidth required for a video stream - in my opinion it's a horrible idea that is fundamentally incompatible with digital display systems.
As mentioned earlier, modern resolutions are still referred to by their "line count". But as shown above, we also have to identify whether we are referring to "interlaced" video (indicated by an i
), or "progressive" video (indicated by a p
).
The frame rate can also be specified, for example:
480i60
- 480 rows, interlaced, 60 Hz field rate (i.e: 30 Hz frame rate)1080p30
- 1080 rows, progressive, 30 Hz frame rate
"... okay, but where did 480 come from?"
The analog electronics involved in CRTs are imprecise, and a particular feature of early models was that as the set warmed up, or the capacitors and electronics aged, the image started to change shape. In addition to this, the electron beam has to be turned off and then redirected to the left of the screen for each line, and to the top for each new field/frame - this takes time, and is the reason for "blanking".
To account for this somewhat, only a handful of the scanned lines were intended / expected to be displayed. NTSC scans 525 lines, but only 483 if these are intended for display - each field shows ~241.5 lines to a viewer (with an additional 21 lines of "blanking"). ~240 lines per field (remember, this is interlaced) equates to 480 lines per frame (in the progressive world). Thus 480. Yay.
For the digital resolutions we follow a pattern... ish:
480 * 1.5 = 720
- "HD Ready"720 * 1.5 = 1080
- "Full HD"1080 * 2 = 2160
- "4k" or "Ultra HD"
So actually "4k" isn't following the * 1.5
pattern from before, and it's not really 4000 pixels in either dimension - it's 3840 ÃÂ 2160.
"4k" is actually "Four times the number of pixels as Full HD". 1920 * 2 = 3840
and 1080 * 2 = 2160
. Or lay out four 1080p displays in a 2 ÃÂ 2 grid - you get 3840 ÃÂ 2160.
Additionally, if we're using 1080p
as a resolution description, then really "4k" should be called 2160p
(which it is in the technical world).
In summary, in the consumer / broadcast space:
480
is because that's approximately the number of visible lines that NTSC would display720
is because that's 1.5ÃÂ480
1080
is because that's 1.5ÃÂ720
2160
is because that's 2ÃÂ1080
- "4k" is because it's a marketing thing - it isn't a technical specification, and I don't believe that there are any stipulations around the use of it...
Note: I've been taking about consumer / broadcast...
Within Cinema there is DCI 2K (capital K
, 2048 ÃÂ 1080) and DCI 4K (capital K
, 4096 ÃÂ 2160), where the 'K' presumably refers to kibi and the horizontal resolution. "DCI 4K" predates consumer "4k".
The aspect ratio of DCI 4K is not 16:9, but a slightly wider 256â¶135... to bring the resolution back in line with 16:9, you can increase the vertical resolution, or decrease the horizontal... But I'm not entirely convinced that the cinema and broadcast standards evolved this closely.
Cinema evolved from whole-frame positives (aka film) directly to digital, whereas TV evolved from a scanning electron beam (line by line) to digital. This is evident in both broadcast, and the way that VHS operates too.
As a little extra, I've included the graphic below to expand on the "image changes shape" statement from above.
The graphic (from here) indicates the various "television safe" areas... note the rounded corners...
Importantly:
- 5 is the "television scanned area"
- 6 is the "television safe area for action"
- Faces and important plot information should not fall outside this area
- 7 is the "television safe area for titles"
- No text should be outside this area to ensure that subtitles / news tickers / etc... don't overhang the edge of an old display
edited Aug 20 at 10:03
answered Aug 18 at 17:20
Attie
8,50231934
8,50231934
2
480i30 would be a weirdly low frame rate; much more common is 480i60, 60 fields per second (can be combed into 30 frames per second). You can deinterlace it to 480p60 (doubling the vertical spatial resolution) or 480p30 (discarding half the temporal resolution). Agreed that interlacing is stupidly horrible now that progressive displays are nearly universal, and it always sucked to store digitally.
â Peter Cordes
Aug 18 at 19:20
@Peter thanks - I couldn't remember if the number was field rate or frame rate...
â Attie
Aug 18 at 20:10
1
The rate number is the temporal resolution. Number of times the display changes (or could change) per second. I think this sometimes gets fudged with progressive content stored as interlaced (evil, bad, and stupid), where both fields are actually from the same time, rather than evenly spaced times. (And of course there's 24p content hard-telecined to 60i, or worse telecined to 60i and then having VFX added at 60i or 30p, like the Star Trek DVDs before they remastered. Those DVDs switched from soft-TC 24p for most scenes to 60i for VFX shots, because the show was edited on video.)
â Peter Cordes
Aug 18 at 20:26
2
There are actually a few different resolutions that have been called "4K". The first usage however (according to Wikipedia) was a camera using 4096Ã2160. Someone thought 4096 was pretty close to 4000, and decided that saying their camera had a 4K resolution sounded cool. It's really just marketing.
â Yay295
Aug 19 at 3:47
1
It should also be noted that 480i or 480p use 720 horizontal pixels with a width to height aspect ratio of 8/9 (= .88888....), which creates an interpolation issue on digital monitors. For CRT displays, this isn't an issue because it just involves a change in the sweep rate (and effective thickness) of the beam used to light up the phosphors. CRT based HDTV's generally support 480i / 480p / 1080i as "native" resolutions, with a better image for 480i/480p than digital displays, but not as good as 1080p digital displays.
â rcgldr
Aug 19 at 21:17
 |Â
show 1 more comment
2
480i30 would be a weirdly low frame rate; much more common is 480i60, 60 fields per second (can be combed into 30 frames per second). You can deinterlace it to 480p60 (doubling the vertical spatial resolution) or 480p30 (discarding half the temporal resolution). Agreed that interlacing is stupidly horrible now that progressive displays are nearly universal, and it always sucked to store digitally.
â Peter Cordes
Aug 18 at 19:20
@Peter thanks - I couldn't remember if the number was field rate or frame rate...
â Attie
Aug 18 at 20:10
1
The rate number is the temporal resolution. Number of times the display changes (or could change) per second. I think this sometimes gets fudged with progressive content stored as interlaced (evil, bad, and stupid), where both fields are actually from the same time, rather than evenly spaced times. (And of course there's 24p content hard-telecined to 60i, or worse telecined to 60i and then having VFX added at 60i or 30p, like the Star Trek DVDs before they remastered. Those DVDs switched from soft-TC 24p for most scenes to 60i for VFX shots, because the show was edited on video.)
â Peter Cordes
Aug 18 at 20:26
2
There are actually a few different resolutions that have been called "4K". The first usage however (according to Wikipedia) was a camera using 4096Ã2160. Someone thought 4096 was pretty close to 4000, and decided that saying their camera had a 4K resolution sounded cool. It's really just marketing.
â Yay295
Aug 19 at 3:47
1
It should also be noted that 480i or 480p use 720 horizontal pixels with a width to height aspect ratio of 8/9 (= .88888....), which creates an interpolation issue on digital monitors. For CRT displays, this isn't an issue because it just involves a change in the sweep rate (and effective thickness) of the beam used to light up the phosphors. CRT based HDTV's generally support 480i / 480p / 1080i as "native" resolutions, with a better image for 480i/480p than digital displays, but not as good as 1080p digital displays.
â rcgldr
Aug 19 at 21:17
2
2
480i30 would be a weirdly low frame rate; much more common is 480i60, 60 fields per second (can be combed into 30 frames per second). You can deinterlace it to 480p60 (doubling the vertical spatial resolution) or 480p30 (discarding half the temporal resolution). Agreed that interlacing is stupidly horrible now that progressive displays are nearly universal, and it always sucked to store digitally.
â Peter Cordes
Aug 18 at 19:20
480i30 would be a weirdly low frame rate; much more common is 480i60, 60 fields per second (can be combed into 30 frames per second). You can deinterlace it to 480p60 (doubling the vertical spatial resolution) or 480p30 (discarding half the temporal resolution). Agreed that interlacing is stupidly horrible now that progressive displays are nearly universal, and it always sucked to store digitally.
â Peter Cordes
Aug 18 at 19:20
@Peter thanks - I couldn't remember if the number was field rate or frame rate...
â Attie
Aug 18 at 20:10
@Peter thanks - I couldn't remember if the number was field rate or frame rate...
â Attie
Aug 18 at 20:10
1
1
The rate number is the temporal resolution. Number of times the display changes (or could change) per second. I think this sometimes gets fudged with progressive content stored as interlaced (evil, bad, and stupid), where both fields are actually from the same time, rather than evenly spaced times. (And of course there's 24p content hard-telecined to 60i, or worse telecined to 60i and then having VFX added at 60i or 30p, like the Star Trek DVDs before they remastered. Those DVDs switched from soft-TC 24p for most scenes to 60i for VFX shots, because the show was edited on video.)
â Peter Cordes
Aug 18 at 20:26
The rate number is the temporal resolution. Number of times the display changes (or could change) per second. I think this sometimes gets fudged with progressive content stored as interlaced (evil, bad, and stupid), where both fields are actually from the same time, rather than evenly spaced times. (And of course there's 24p content hard-telecined to 60i, or worse telecined to 60i and then having VFX added at 60i or 30p, like the Star Trek DVDs before they remastered. Those DVDs switched from soft-TC 24p for most scenes to 60i for VFX shots, because the show was edited on video.)
â Peter Cordes
Aug 18 at 20:26
2
2
There are actually a few different resolutions that have been called "4K". The first usage however (according to Wikipedia) was a camera using 4096Ã2160. Someone thought 4096 was pretty close to 4000, and decided that saying their camera had a 4K resolution sounded cool. It's really just marketing.
â Yay295
Aug 19 at 3:47
There are actually a few different resolutions that have been called "4K". The first usage however (according to Wikipedia) was a camera using 4096Ã2160. Someone thought 4096 was pretty close to 4000, and decided that saying their camera had a 4K resolution sounded cool. It's really just marketing.
â Yay295
Aug 19 at 3:47
1
1
It should also be noted that 480i or 480p use 720 horizontal pixels with a width to height aspect ratio of 8/9 (= .88888....), which creates an interpolation issue on digital monitors. For CRT displays, this isn't an issue because it just involves a change in the sweep rate (and effective thickness) of the beam used to light up the phosphors. CRT based HDTV's generally support 480i / 480p / 1080i as "native" resolutions, with a better image for 480i/480p than digital displays, but not as good as 1080p digital displays.
â rcgldr
Aug 19 at 21:17
It should also be noted that 480i or 480p use 720 horizontal pixels with a width to height aspect ratio of 8/9 (= .88888....), which creates an interpolation issue on digital monitors. For CRT displays, this isn't an issue because it just involves a change in the sweep rate (and effective thickness) of the beam used to light up the phosphors. CRT based HDTV's generally support 480i / 480p / 1080i as "native" resolutions, with a better image for 480i/480p than digital displays, but not as good as 1080p digital displays.
â rcgldr
Aug 19 at 21:17
 |Â
show 1 more comment
up vote
7
down vote
Attie has explained how vertical pixel counts were derived for traditional resolutions such as SD (480p), HD (720p) and Full HD (1080p). I would like to discuss K measurements which originate from the cinema industry. A K is about 1000 horizontal pixels and 2K and 4K were casual terms referring to resolutions of approximately 2000 and 4000 horizontal pixels.
Different K measurements were later standardised for digital cinema (2005) and consumer television (2007).
Digital Cinema Initiatives has specified the DCI 4K and 2K resoultion standards (4096âÂÂÃÂâÂÂ2160 and 2048 à1080). These are the full frame resolutions for digital cinema projectors, with many films displayed in the a crop format such as flat crop (3996âÂÂÃÂâÂÂ2160 , 1.85â¶1 aspect ratio) or CinemaScope crop (4096âÂÂÃÂâÂÂ1716, âÂÂ2.39â¶1 aspect ratio).
The 4K standard for consumer television (UHDTV1) can be more accurately labelled as Ultra HD or Quad HD. It is equivalent to 4 Full HD frames placed together. It is 3840âÂÂÃÂâÂÂ2160 with âÂÂ1.78â¶1 aspect ratio.
The Wikipedia article on 4K resolution provides a helpful visualisation which compares different frame sizes.
add a comment |Â
up vote
7
down vote
Attie has explained how vertical pixel counts were derived for traditional resolutions such as SD (480p), HD (720p) and Full HD (1080p). I would like to discuss K measurements which originate from the cinema industry. A K is about 1000 horizontal pixels and 2K and 4K were casual terms referring to resolutions of approximately 2000 and 4000 horizontal pixels.
Different K measurements were later standardised for digital cinema (2005) and consumer television (2007).
Digital Cinema Initiatives has specified the DCI 4K and 2K resoultion standards (4096âÂÂÃÂâÂÂ2160 and 2048 à1080). These are the full frame resolutions for digital cinema projectors, with many films displayed in the a crop format such as flat crop (3996âÂÂÃÂâÂÂ2160 , 1.85â¶1 aspect ratio) or CinemaScope crop (4096âÂÂÃÂâÂÂ1716, âÂÂ2.39â¶1 aspect ratio).
The 4K standard for consumer television (UHDTV1) can be more accurately labelled as Ultra HD or Quad HD. It is equivalent to 4 Full HD frames placed together. It is 3840âÂÂÃÂâÂÂ2160 with âÂÂ1.78â¶1 aspect ratio.
The Wikipedia article on 4K resolution provides a helpful visualisation which compares different frame sizes.
add a comment |Â
up vote
7
down vote
up vote
7
down vote
Attie has explained how vertical pixel counts were derived for traditional resolutions such as SD (480p), HD (720p) and Full HD (1080p). I would like to discuss K measurements which originate from the cinema industry. A K is about 1000 horizontal pixels and 2K and 4K were casual terms referring to resolutions of approximately 2000 and 4000 horizontal pixels.
Different K measurements were later standardised for digital cinema (2005) and consumer television (2007).
Digital Cinema Initiatives has specified the DCI 4K and 2K resoultion standards (4096âÂÂÃÂâÂÂ2160 and 2048 à1080). These are the full frame resolutions for digital cinema projectors, with many films displayed in the a crop format such as flat crop (3996âÂÂÃÂâÂÂ2160 , 1.85â¶1 aspect ratio) or CinemaScope crop (4096âÂÂÃÂâÂÂ1716, âÂÂ2.39â¶1 aspect ratio).
The 4K standard for consumer television (UHDTV1) can be more accurately labelled as Ultra HD or Quad HD. It is equivalent to 4 Full HD frames placed together. It is 3840âÂÂÃÂâÂÂ2160 with âÂÂ1.78â¶1 aspect ratio.
The Wikipedia article on 4K resolution provides a helpful visualisation which compares different frame sizes.
Attie has explained how vertical pixel counts were derived for traditional resolutions such as SD (480p), HD (720p) and Full HD (1080p). I would like to discuss K measurements which originate from the cinema industry. A K is about 1000 horizontal pixels and 2K and 4K were casual terms referring to resolutions of approximately 2000 and 4000 horizontal pixels.
Different K measurements were later standardised for digital cinema (2005) and consumer television (2007).
Digital Cinema Initiatives has specified the DCI 4K and 2K resoultion standards (4096âÂÂÃÂâÂÂ2160 and 2048 à1080). These are the full frame resolutions for digital cinema projectors, with many films displayed in the a crop format such as flat crop (3996âÂÂÃÂâÂÂ2160 , 1.85â¶1 aspect ratio) or CinemaScope crop (4096âÂÂÃÂâÂÂ1716, âÂÂ2.39â¶1 aspect ratio).
The 4K standard for consumer television (UHDTV1) can be more accurately labelled as Ultra HD or Quad HD. It is equivalent to 4 Full HD frames placed together. It is 3840âÂÂÃÂâÂÂ2160 with âÂÂ1.78â¶1 aspect ratio.
The Wikipedia article on 4K resolution provides a helpful visualisation which compares different frame sizes.
answered Aug 20 at 2:58
Wes Toleman
2013
2013
add a comment |Â
add a comment |Â
up vote
3
down vote
Your assessment that resolution is pixel pitch is kind of correct. In fact, for everything except digital displays, that's the correct definition of resolution (well, more accurately, it's the smallest detail that can be resolved accurately, which functionally translates to the size of the dots making up an image, and from that pixel pitch). If you ever talk to somebody in the printing industry, this is usually what they mean when they say 'resolution'.
The reason that the term has a different meaning for computer displays is largely historical, but is actually pretty easily explained without getting into much history.
On a raster display1, you have what's known as a 'native' resolution. This is the highest number of pixels you can have on each side of the screen, and is usually what is used to describe the screen. For example, a Full HD display has a native resolution of 1920 pixels horizontally, and 1080 vertically, so a Full HD display with a roughly 18 inch diagonal has a resolution (in the classical sense) of about 120 dots per inch (DPI) (which is actually pretty low compared to print media).
However, you can run most displays at lower than their native resolution, which is functionally equivalent to having a larger pixel pitch on the same display. Taking that same 18 inch Full HD display and running it at 1280x720 gives you the equivalent of 80 DPI on the same screen.
Now, the OS and application software (usually) don't care about the exact physical dimensions of the screen, because that information is not really all that useful to displaying information unless you need to display something 'real-size'. Given this, the OS just assumes the size of the display never changes, which means that the different pixel counts are functionally equivalent to the resolution.
- A display that uses rows and columns of individual points that can be turned on and off to produce an image. Almost all modern computer displays are raster displays, because they're fare easier to make. Compare to a vector display, which just draws lines directly (with one of the most recognizable examples being the displays used in the original Asteroids arcade cabinets).
1
"Resolution" doesn't have any bearing on dot size, pitch or aspect ratio as such - it refers to the smallest addressable "thing" - DAC resolution of 16-bits, screen resolution, etc... interestingly with RGB displays, you quite whole pixels, not sub pixels. With print it's quite meaningless to advertise a resolution across a page due to various sizes (A4 / A5 / etc...), margins, etc... so instead they specify the area (typically an Inch) and count the dots inside it (DPI)
â Attie
Aug 19 at 10:34
@Attie No, it does have bearing on dot size and/or pixel pitch (though not really aspect ratio). If you have a display of a given size, higher resolution means smaller pixel pitch, period (and thus higher detail and less aliasing). The reason resolution matters for actually displaying anything when dealing with anything higher level than drivers is pixel pitch and pixel count (which is a side effect of pixel pitch and display size), not addressing. Also, the use of dots-per-unit-area in print far predates computer displays.
â Austin Hemmelgarn
Aug 19 at 12:55
I completely agree that "resolution", "pixel pitch" and "physical dimensions" are all intertwined (how could they not be)... but "resolution" by itself has no bearing on the other two. The fact that software is typically aware of a display's physical dimensions or PPI means that it's possible to display things at approximately real size.
â Attie
Aug 19 at 13:54
As for DPI in print and pixel aspect ratio - until recently pixels were not square! IIRC facsimile had pixels with a 2:1 aspect ratio, and NTSC was 10:11... rendering the image with an incorrect PAR produces squashed/stretched looking images. Additionally, DPI in print usually needs to be so high because of the limited count of discernible colors - dithering coming to the rescue.
â Attie
Aug 19 at 14:00
add a comment |Â
up vote
3
down vote
Your assessment that resolution is pixel pitch is kind of correct. In fact, for everything except digital displays, that's the correct definition of resolution (well, more accurately, it's the smallest detail that can be resolved accurately, which functionally translates to the size of the dots making up an image, and from that pixel pitch). If you ever talk to somebody in the printing industry, this is usually what they mean when they say 'resolution'.
The reason that the term has a different meaning for computer displays is largely historical, but is actually pretty easily explained without getting into much history.
On a raster display1, you have what's known as a 'native' resolution. This is the highest number of pixels you can have on each side of the screen, and is usually what is used to describe the screen. For example, a Full HD display has a native resolution of 1920 pixels horizontally, and 1080 vertically, so a Full HD display with a roughly 18 inch diagonal has a resolution (in the classical sense) of about 120 dots per inch (DPI) (which is actually pretty low compared to print media).
However, you can run most displays at lower than their native resolution, which is functionally equivalent to having a larger pixel pitch on the same display. Taking that same 18 inch Full HD display and running it at 1280x720 gives you the equivalent of 80 DPI on the same screen.
Now, the OS and application software (usually) don't care about the exact physical dimensions of the screen, because that information is not really all that useful to displaying information unless you need to display something 'real-size'. Given this, the OS just assumes the size of the display never changes, which means that the different pixel counts are functionally equivalent to the resolution.
- A display that uses rows and columns of individual points that can be turned on and off to produce an image. Almost all modern computer displays are raster displays, because they're fare easier to make. Compare to a vector display, which just draws lines directly (with one of the most recognizable examples being the displays used in the original Asteroids arcade cabinets).
1
"Resolution" doesn't have any bearing on dot size, pitch or aspect ratio as such - it refers to the smallest addressable "thing" - DAC resolution of 16-bits, screen resolution, etc... interestingly with RGB displays, you quite whole pixels, not sub pixels. With print it's quite meaningless to advertise a resolution across a page due to various sizes (A4 / A5 / etc...), margins, etc... so instead they specify the area (typically an Inch) and count the dots inside it (DPI)
â Attie
Aug 19 at 10:34
@Attie No, it does have bearing on dot size and/or pixel pitch (though not really aspect ratio). If you have a display of a given size, higher resolution means smaller pixel pitch, period (and thus higher detail and less aliasing). The reason resolution matters for actually displaying anything when dealing with anything higher level than drivers is pixel pitch and pixel count (which is a side effect of pixel pitch and display size), not addressing. Also, the use of dots-per-unit-area in print far predates computer displays.
â Austin Hemmelgarn
Aug 19 at 12:55
I completely agree that "resolution", "pixel pitch" and "physical dimensions" are all intertwined (how could they not be)... but "resolution" by itself has no bearing on the other two. The fact that software is typically aware of a display's physical dimensions or PPI means that it's possible to display things at approximately real size.
â Attie
Aug 19 at 13:54
As for DPI in print and pixel aspect ratio - until recently pixels were not square! IIRC facsimile had pixels with a 2:1 aspect ratio, and NTSC was 10:11... rendering the image with an incorrect PAR produces squashed/stretched looking images. Additionally, DPI in print usually needs to be so high because of the limited count of discernible colors - dithering coming to the rescue.
â Attie
Aug 19 at 14:00
add a comment |Â
up vote
3
down vote
up vote
3
down vote
Your assessment that resolution is pixel pitch is kind of correct. In fact, for everything except digital displays, that's the correct definition of resolution (well, more accurately, it's the smallest detail that can be resolved accurately, which functionally translates to the size of the dots making up an image, and from that pixel pitch). If you ever talk to somebody in the printing industry, this is usually what they mean when they say 'resolution'.
The reason that the term has a different meaning for computer displays is largely historical, but is actually pretty easily explained without getting into much history.
On a raster display1, you have what's known as a 'native' resolution. This is the highest number of pixels you can have on each side of the screen, and is usually what is used to describe the screen. For example, a Full HD display has a native resolution of 1920 pixels horizontally, and 1080 vertically, so a Full HD display with a roughly 18 inch diagonal has a resolution (in the classical sense) of about 120 dots per inch (DPI) (which is actually pretty low compared to print media).
However, you can run most displays at lower than their native resolution, which is functionally equivalent to having a larger pixel pitch on the same display. Taking that same 18 inch Full HD display and running it at 1280x720 gives you the equivalent of 80 DPI on the same screen.
Now, the OS and application software (usually) don't care about the exact physical dimensions of the screen, because that information is not really all that useful to displaying information unless you need to display something 'real-size'. Given this, the OS just assumes the size of the display never changes, which means that the different pixel counts are functionally equivalent to the resolution.
- A display that uses rows and columns of individual points that can be turned on and off to produce an image. Almost all modern computer displays are raster displays, because they're fare easier to make. Compare to a vector display, which just draws lines directly (with one of the most recognizable examples being the displays used in the original Asteroids arcade cabinets).
Your assessment that resolution is pixel pitch is kind of correct. In fact, for everything except digital displays, that's the correct definition of resolution (well, more accurately, it's the smallest detail that can be resolved accurately, which functionally translates to the size of the dots making up an image, and from that pixel pitch). If you ever talk to somebody in the printing industry, this is usually what they mean when they say 'resolution'.
The reason that the term has a different meaning for computer displays is largely historical, but is actually pretty easily explained without getting into much history.
On a raster display1, you have what's known as a 'native' resolution. This is the highest number of pixels you can have on each side of the screen, and is usually what is used to describe the screen. For example, a Full HD display has a native resolution of 1920 pixels horizontally, and 1080 vertically, so a Full HD display with a roughly 18 inch diagonal has a resolution (in the classical sense) of about 120 dots per inch (DPI) (which is actually pretty low compared to print media).
However, you can run most displays at lower than their native resolution, which is functionally equivalent to having a larger pixel pitch on the same display. Taking that same 18 inch Full HD display and running it at 1280x720 gives you the equivalent of 80 DPI on the same screen.
Now, the OS and application software (usually) don't care about the exact physical dimensions of the screen, because that information is not really all that useful to displaying information unless you need to display something 'real-size'. Given this, the OS just assumes the size of the display never changes, which means that the different pixel counts are functionally equivalent to the resolution.
- A display that uses rows and columns of individual points that can be turned on and off to produce an image. Almost all modern computer displays are raster displays, because they're fare easier to make. Compare to a vector display, which just draws lines directly (with one of the most recognizable examples being the displays used in the original Asteroids arcade cabinets).
answered Aug 18 at 21:12
Austin Hemmelgarn
1,89418
1,89418
1
"Resolution" doesn't have any bearing on dot size, pitch or aspect ratio as such - it refers to the smallest addressable "thing" - DAC resolution of 16-bits, screen resolution, etc... interestingly with RGB displays, you quite whole pixels, not sub pixels. With print it's quite meaningless to advertise a resolution across a page due to various sizes (A4 / A5 / etc...), margins, etc... so instead they specify the area (typically an Inch) and count the dots inside it (DPI)
â Attie
Aug 19 at 10:34
@Attie No, it does have bearing on dot size and/or pixel pitch (though not really aspect ratio). If you have a display of a given size, higher resolution means smaller pixel pitch, period (and thus higher detail and less aliasing). The reason resolution matters for actually displaying anything when dealing with anything higher level than drivers is pixel pitch and pixel count (which is a side effect of pixel pitch and display size), not addressing. Also, the use of dots-per-unit-area in print far predates computer displays.
â Austin Hemmelgarn
Aug 19 at 12:55
I completely agree that "resolution", "pixel pitch" and "physical dimensions" are all intertwined (how could they not be)... but "resolution" by itself has no bearing on the other two. The fact that software is typically aware of a display's physical dimensions or PPI means that it's possible to display things at approximately real size.
â Attie
Aug 19 at 13:54
As for DPI in print and pixel aspect ratio - until recently pixels were not square! IIRC facsimile had pixels with a 2:1 aspect ratio, and NTSC was 10:11... rendering the image with an incorrect PAR produces squashed/stretched looking images. Additionally, DPI in print usually needs to be so high because of the limited count of discernible colors - dithering coming to the rescue.
â Attie
Aug 19 at 14:00
add a comment |Â
1
"Resolution" doesn't have any bearing on dot size, pitch or aspect ratio as such - it refers to the smallest addressable "thing" - DAC resolution of 16-bits, screen resolution, etc... interestingly with RGB displays, you quite whole pixels, not sub pixels. With print it's quite meaningless to advertise a resolution across a page due to various sizes (A4 / A5 / etc...), margins, etc... so instead they specify the area (typically an Inch) and count the dots inside it (DPI)
â Attie
Aug 19 at 10:34
@Attie No, it does have bearing on dot size and/or pixel pitch (though not really aspect ratio). If you have a display of a given size, higher resolution means smaller pixel pitch, period (and thus higher detail and less aliasing). The reason resolution matters for actually displaying anything when dealing with anything higher level than drivers is pixel pitch and pixel count (which is a side effect of pixel pitch and display size), not addressing. Also, the use of dots-per-unit-area in print far predates computer displays.
â Austin Hemmelgarn
Aug 19 at 12:55
I completely agree that "resolution", "pixel pitch" and "physical dimensions" are all intertwined (how could they not be)... but "resolution" by itself has no bearing on the other two. The fact that software is typically aware of a display's physical dimensions or PPI means that it's possible to display things at approximately real size.
â Attie
Aug 19 at 13:54
As for DPI in print and pixel aspect ratio - until recently pixels were not square! IIRC facsimile had pixels with a 2:1 aspect ratio, and NTSC was 10:11... rendering the image with an incorrect PAR produces squashed/stretched looking images. Additionally, DPI in print usually needs to be so high because of the limited count of discernible colors - dithering coming to the rescue.
â Attie
Aug 19 at 14:00
1
1
"Resolution" doesn't have any bearing on dot size, pitch or aspect ratio as such - it refers to the smallest addressable "thing" - DAC resolution of 16-bits, screen resolution, etc... interestingly with RGB displays, you quite whole pixels, not sub pixels. With print it's quite meaningless to advertise a resolution across a page due to various sizes (A4 / A5 / etc...), margins, etc... so instead they specify the area (typically an Inch) and count the dots inside it (DPI)
â Attie
Aug 19 at 10:34
"Resolution" doesn't have any bearing on dot size, pitch or aspect ratio as such - it refers to the smallest addressable "thing" - DAC resolution of 16-bits, screen resolution, etc... interestingly with RGB displays, you quite whole pixels, not sub pixels. With print it's quite meaningless to advertise a resolution across a page due to various sizes (A4 / A5 / etc...), margins, etc... so instead they specify the area (typically an Inch) and count the dots inside it (DPI)
â Attie
Aug 19 at 10:34
@Attie No, it does have bearing on dot size and/or pixel pitch (though not really aspect ratio). If you have a display of a given size, higher resolution means smaller pixel pitch, period (and thus higher detail and less aliasing). The reason resolution matters for actually displaying anything when dealing with anything higher level than drivers is pixel pitch and pixel count (which is a side effect of pixel pitch and display size), not addressing. Also, the use of dots-per-unit-area in print far predates computer displays.
â Austin Hemmelgarn
Aug 19 at 12:55
@Attie No, it does have bearing on dot size and/or pixel pitch (though not really aspect ratio). If you have a display of a given size, higher resolution means smaller pixel pitch, period (and thus higher detail and less aliasing). The reason resolution matters for actually displaying anything when dealing with anything higher level than drivers is pixel pitch and pixel count (which is a side effect of pixel pitch and display size), not addressing. Also, the use of dots-per-unit-area in print far predates computer displays.
â Austin Hemmelgarn
Aug 19 at 12:55
I completely agree that "resolution", "pixel pitch" and "physical dimensions" are all intertwined (how could they not be)... but "resolution" by itself has no bearing on the other two. The fact that software is typically aware of a display's physical dimensions or PPI means that it's possible to display things at approximately real size.
â Attie
Aug 19 at 13:54
I completely agree that "resolution", "pixel pitch" and "physical dimensions" are all intertwined (how could they not be)... but "resolution" by itself has no bearing on the other two. The fact that software is typically aware of a display's physical dimensions or PPI means that it's possible to display things at approximately real size.
â Attie
Aug 19 at 13:54
As for DPI in print and pixel aspect ratio - until recently pixels were not square! IIRC facsimile had pixels with a 2:1 aspect ratio, and NTSC was 10:11... rendering the image with an incorrect PAR produces squashed/stretched looking images. Additionally, DPI in print usually needs to be so high because of the limited count of discernible colors - dithering coming to the rescue.
â Attie
Aug 19 at 14:00
As for DPI in print and pixel aspect ratio - until recently pixels were not square! IIRC facsimile had pixels with a 2:1 aspect ratio, and NTSC was 10:11... rendering the image with an incorrect PAR produces squashed/stretched looking images. Additionally, DPI in print usually needs to be so high because of the limited count of discernible colors - dithering coming to the rescue.
â Attie
Aug 19 at 14:00
add a comment |Â
11
You thought what is determined by horizontal width? Screen resolution is width by height. Multiply one by the other to get "K" [in the broadest terms]. 720p & 1080p are rendering standards, not resolution standards. The 'p' means progressive scan, to differentiate from the inferior 'i' for interleaved scan.
â Tetsujin
Aug 18 at 16:53
23
It's an idiotic change made by manufacturers so that the latest jump to "4k" sounds like a huge deal rather than the on going incremental change that it is. Up to 4k we were naming the resolution by vertical resolution and then suddenly with 4k they decided to use the horizontal width instead.
â Mokubaiâ¦
Aug 18 at 17:00
4
Look at this answer I posted here.
â JakeGould
Aug 18 at 17:32
4
@Mokubai - yeah the next step is for them to go all hard drive on us and we'll lose the last 96 pixels too.
â davidbak
Aug 18 at 22:45
5
I'm very confused because I don't see a difference between the two examples of 1920*1080 and 1280*720. The "p" number is the second number in both cases. Maybe the "but" should be an "and", because, as written, it reads very odd to me.
â Greg Schmit
Aug 19 at 23:39