Recuperando da história, esse negocio de resolução de 72 DPI ou 96 DPI foi um padrão que nos primórdios da computação (1980) fora definido como padrão de escala para monitores gráficos da Apple (72 DPI) e Microsoft (96 DPI). Mas como ja foi dito, acima, em fotografia atualmente o que importa é a resolução de saída, de acordo com a mídia escolhida.
veja mais em:
http://en.wikipedia.org/wiki/Dots_per_inch"
Computer monitor DPI standards
Since the 1980s, the Microsoft Windows operating system has set the default display "DPI" to 96 PPI, while Apple/Macintosh computers have used a default of 72 PPI.[3] These default specifications arose out of the problems rendering standard fonts in the early display systems of the 1980s, including the IBM-based CGA, EGA, VGA and 8514 displays as well as the Macintosh displays featured in the 128K computer and its successors. The choice of 72 PPI by Macintosh for their displays arose from the convenient fact that the official 72 points-per-inch mirrored the 72 pixels-per-inch that actually appeared on their display screens. (Points are a physical unit-of-measure in typography dating to the days of printing presses, where 1 point by the modern definition is 1/72 of the international inch (25.4 mm), which therefore makes 1 point approximately 0.0139 in or 352.8 µm). Thus, a 72 pixels-per-inch seen on the display was exactly the same physical dimensions as the 72 points-per-inch later seen on a printout, with 1 pt in printed text equal to 1 px on the display screen. As it is, the Macintosh 128K featured a screen measuring 512 pixels in width by 342 pixels in height, and this corresponded to the width of standard office paper (512 px ÷ 72 px/in ≈ 7.1 in, with a 0.7 in margin down each side when assuming 8.5 in × 11 in North American paper size).[citation needed]
A consequence of Apple's decision was that the widely used 10 point fonts from the typewriter era had to be allotted 10 display pixels in em height, and 5 display pixels in x-height. This is technically described as 10 pixels per em (PPEm). This made 10-point fonts render crudely and difficult to read on the display screen, particularly for lowercase characters. Furthermore, there was the consideration that computer screens are typically viewed (at a desk) at a distance 1/3 or 33% greater than printed materials, causing a mismatch between the perceived sizes seen on the computer screen versus those on the printouts.[citation needed]
Microsoft tried to solve both problems with a hack that has had long-term consequences for the understanding of what DPI and PPI mean.[4] Microsoft began writing its software to treat the screen as though it provided a PPI characteristic that is \tfrac{4}{3} of what the screen actually displayed. Because most screens at the time provided around 72 PPI, Microsoft essentially wrote its software to assume that every screen provides 96 PPI (because 72 * (1+\tfrac{1}{3}) = 96). The short-term gain of this trickery was twofold:
It would seem to the software that \tfrac{1}{3} more pixels were available for rendering an image, thereby allowing for bitmap fonts to be created with greater detail.
On every screen that actually provided 72 PPI, each graphical element (such as a character of text) would be rendered at a size \tfrac{1}{3} larger than it "should" be, thereby allowing a person to sit a comfortable distance from the screen. However, larger graphical elements meant less screen space was available for programs to draw; indeed, although the default 720-pixel wide mode of a Hercules mono graphics adaptor (the one-time gold standard for high resolution PC graphics) – or a "tweaked" VGA adaptor – provided an apparent 7.5-inch page width at this resolution, the more common and color-capable display adaptors of the time all provided a 640-pixel wide image in their high resolution modes, enough for a bare 6.67 inches at 100% zoom (and barely any greater visible page height – a maximum of 5 inches, versus 4.75). Consequently, the default margins in Microsoft Word were set, and still remain at 1 full inch on all sides of the page, keeping the "text width" for standard size printer paper within visible limits; despite most computer monitors now being both larger and finer-pitched, and printer paper transports having become more sophisticated, the Mac-standard half-inch borders remain listed in Word 2010's page layout presets as the "narrow" option (versus the 1-inch default).[citation needed]
Without using supplemental, software-provided zoom levels, the 1:1 relationship between display and print size was (deliberately) lost; the availability of different-sized, user-adjustable monitors and display adaptors with varying output resolutions exacerbated this, as it was not possible to rely on a properly-adjusted "standard" monitor and adaptor having a known PPI. For example, a 12" Hercules monitor and adaptor with a thick bezel and a little underscan may offer 90 "physical" PPI, with the displayed image appearing nearly identical to hardcopy (assuming the H-scan density was properly adjusted to give square pixels) but a thin-bezel 14" VGA monitor adjusted to give a borderless display may be closer to 60, with the same bitmap image thus appearing 50% larger; yet, someone with an 8514 ("XGA") adaptor and the same monitor could achieve 100 DPI using its 1024-pixel wide mode and adjusting the image to be underscanned. A user who wanted to directly compare on-screen elements against those on an existing printed page by holding it up against the monitor would therefore first need to determine the correct zoom level to use, largely by trial and error, and often not be able to obtain an exact match in programs that only allowed integer percent settings, or even fixed pre-programmed zoom levels. For the examples above, they may need to use respectively 94% (precisely, 93.75) – or 95/90, 63% (62.5) – or 60/66.7; and 104% (104.167) – or 105, with the more commonly accessible 110% actually being a less precise match.[citation needed]
Thus, for example, a 10-point font on a Macintosh (at 72 PPI) was represented with 10 pixels (i.e., 10 PPEm), whereas a 10-point font on a Windows platform (at 96 PPI) at the same zoom level is represented with 13 pixels (i.e., Microsoft rounded 13.3333 to 13 pixels, or 13 PPEm) – and, on a typical consumer grade monitor, would have physically appeared around 15/72 to 16/72 of an inch high instead of 10/72. Likewise, a 12-point font was represented with 12 pixels on a Macintosh, and 16 pixels (or a physical display height of maybe 19/72 of an inch) on a Windows platform at the same zoom, and so on.[5] The negative consequence of this standard is that with 96 PPI displays, there is no longer a 1-to-1 relationship between the font size in pixels and the printout size in points. This difference is accentuated on more recent displays that feature higher pixel densities. This has been less of a problem with the advent of vector graphics and fonts being used in place of bitmap graphics and fonts. Moreover, many Windows software programs have been written since the 1980s which assume that the screen provides 96 PPI. Accordingly, these programs do not display properly at common alternative resolutions such as 72 PPI or 120 PPI. The solution has been to introduce two concepts:[4]
logical PPI: The PPI that software claims a screen provides. This can be thought of as the PPI provided by a virtual screen created by the operating system.
physical PPI: The PPI that a physical screen actually provides.
Software programs render images to the virtual screen and then the operating system renders the virtual screen onto the physical screen. With a logical PPI of 96 PPI, older programs can still run properly regardless of the actual physical PPI of the display screen, although they may exhibit some visual distortion thanks to the effective 133.3% pixel zoom level (requiring either that every third pixel be doubled in width/height, or heavy-handed smoothing be employed)."