5

Acquisition

Any digital intermediate process begins with the acquisition of media (sometimes referred to as the “ingest” of media). In many cases, the digital intermediate process is the final stage of a production, following shooting and editing. In these cases, the type of media required can be worked out before the digital intermediate process begins. Typically, there is footage for the final cut and there can also be audio and textual elements, such as titles and dialog. The focus of the majority of this book will be on the visual component of the digital intermediate, because the audio is almost always mastered independently and added separately at the end.

5.1 Crunching Numbers

Computers don’t understand analog information (which includes printed images, reels of film, people talking, and so on). This means you can’t just show a film, edited or otherwise, to a computer system and get it to create a color-graded digital master (not yet anyway).

It is important to remember throughout the intermediate process that from the moment you begin, all you’re doing is crunching numbers. Even though a skilled colorist can turn a blue sweater red, or selectively add or remove colors throughout a scene; to the computer systems involved in the process all that is happening is that one long number is converted into another long number. The system works because of the innovative software involved and the experienced operators who can predict how the system is likely to react to changes and can understand the limitations and shortcuts.

Computers can do very complicated mathematics very, very quickly compared to humans. Computers are able to take a digital image with the same amount of detail as a piece of film and almost instantly produce a perfect copy. However, all they are doing is performing some repetitive calculations at very high speed. This strength can also be a weakness. Because the computer has no knowledge of what the image contains and can associate no meaning to it, it cannot tell which parts are important. For this reason, defects such as noise, dust, and scratches are treated with the same care as desirable parts of an image. For this reason, you must always assume the golden rule: what you get out of a digital system will only ever be as good as what you put into it.

Actually, that rule isn’t strictly true. Some methods for automated processes can perform feature analysis on images and sequences of images, can estimate which parts of an image or sequence of images are important and use that information to enhance the images—“estimate” being the keyword. Also, at every stage of the digital intermediate process are operators who are able to make subjective decisions about the processing and quality of the images, ensuring the highest possible standard throughout.

5.2 Digital Image Quality

The fundamental advantage that digital media has over analog materials is that digital information can be transferred across large distances with no loss of quality. A digital image looks exactly the same whether you’re accessing a digital camera directly, or from the other side of the world, or accessing the image via the Internet, bounced off of satellites, or through some other transfer method. As long as the data remains unaltered, it’s as good as the original. Sadly, the same cannot be said for video and film, which are both analog formats and therefore subject to many forms of degradation, the most significant being “generation loss.”

5.2.1 Generation Loss

Every time you copy an analog source, such as a video tape or a reel of film, it degrades quality. In fact, every time you even view certain analog sources (including both film and video), it suffers a loss of quality. This is because most analog-viewing devices are mechanical and have moving parts that can damage the source. Viewing (or even handling) film prints or negatives can cause scratches or dust and other damage to them.1 Video tapes risk decay or becoming demagnetized every time they’re moved across a video head.

Each copy of an analog source introduces additional, lasting damage to the copy. A copy of a copy retains all the errors so far and then adds more errors. Errors are caused by effects such as noise (in the case of video) and chemical stains (in the case of film), among others.2 Each successive copy is called a “generation.” A fifth-generation copy is therefore typically of lower quality than a second-generation copy. To maximize the quality, it’s preferable to work from original material whenever possible.

5.2.2 What is Quality?

This book includes discussions of various factors that influence the somewhat esoteric criteria of “quality.” As we know, quality, especially picture quality, can be a very subjective thing. For example, a well-lit scene recorded on DVCAM video might be considered of superior quality to a poorly lit scene filmed on 35mm negative. There are ways to measure image quality objectively though; if you assume all other things (such as lighting, set design, and so on) to be equal, then “quality” can be defined as “how closely an image matches the original content,” which is how the term is used throughout this book. There is another issue at stake though. What if you don’t want the final project to closely resemble the original? After all, part of the “magic” of filmmaking is in creating imagery that looks far removed from or even better than real life. However, even in those situations, you want to produce images that are of high quality—with as much detail as possible—so that you selectively control what stays and what doesn’t later on.

In the digital intermediate pipeline, the aim is to maintain the highest possible quality throughout, with the material prepared for supervised sessions with the filmmakers, who then make decisions about how to affect the images and control the subjective level of quality. For example, if you want a shot to look blurry in the final production, it can be argued that it doesn’t matter if the scene is transferred at low quality, making it visibly blurry. However, in this case, there is no way to “unblur” the image later on (even digital sharpening techniques won’t create as good an image as the original footage), and the level of blurriness can’t be adjusted interactively, in context with the scene. Chapter 14 contains information about using digital processes to help define the aesthetics and style of a production.

5.2.3 Resolving Power

Resolving power is one objective method for measuring the quality of an image. It is found by measuring how much detail is crammed into a given area. The simplest way to measure is to capture an image of a chart of lines that alternate between black and white and get increasingly thinner. When the lines are too thin for the recorder, they are recorded as gray. The point where this occurs is used to calculate the “resolution” of the recorder.

images

Figure 5–1   A chart such as this can be used to determine the resolving power of an imaging system. © Norman Koren, Available at www.normankoren.com

With film, the resolution depends upon the average grain size, and the resolution of video tape depends upon the properties of the type of signal received. Other factors can also play a part in influencing the resolving power, such as the optical system (i.e., lenses). In reality, the derived resolving power is actually a measurement of the entire imaging system.3

Ultimately though, to get the most relevant result, the resolving power, and thus the maximum resolution of any system, should be measured by looking at the final output.

5.2.4 Spread Function

Further evaluation of any optical system can be made by determining its spread function. The spread function is essentially a measurement of how different a single point of light (or a line or a hard edge) appears when recorded and demonstrates many different characteristics of degradation in a system.

5.2.5 Modulation Transfer Function

An even more useful result is the modulation transfer function (MTF). In simple terms, MTF is a measurement of the response (or accuracy) of any system across varying frequencies (which amounts to the level of detail). You would expect that the accuracy of reproduction for most systems would be directly proportional to the level of detail, so that, for example, an image of large squares would be more faithfully reproduced than an image of very thin lines. However in practice this often isn’t the case. MTF measurements and graphs serve to illustrate exactly how the systems respond to different frequencies, thus highlighting potential problems, as well as serving as a useful method for making objective comparisons of systems and components.

One of the reasons the MTF measurement is useful is because certain systems respond better than others under certain conditions. For example, some types of photographic film exhibit greater granularity in the mid-tone regions than in shadow or highlight areas. An MTF can provide an objective measurement of these types of effects.

images

Figure 5–2   A perfect hard edge is plotted with a vertical line

images

Figure 5–3   With a blurred (more realistic) edge, the transition is more gradual

images

Figure 5–4   An artificially sharpened edge may exaggerate the edge

images

Figure 5–5   Factors such as film grain may also play a part in the measurement of an edge

MTFs are measured by imaging a special chart (which displays a series of sine wave patterns) and performing various measurements and calculations on it.

5.2.6 Signal-to-Noise Ratios

Another useful measurement of the quality of a system is its signalto-noise ratio. This is calculated by dividing the signal strength of a system by the inherent noise. Every system, or signal (such as a video signal) has some degree of noise inherent, which comes from a variety of sources. Every time the signal is amplified, the noise is amplified along with it. A higher signal-to-noise ratio indicates a higher quality system than a lower ratio, because it means the effects of noise are less pronounced.

5.2.7 Detective Quantum Efficiency

Modulation transfer functions don’t take into account the effects of noise in a system, and signal-to-noise ratios don’t tell us anything about the accuracy of a system’s response to light (i.e., its contrast performance). A method that combines these two measurements is the detective quantum efficiency (DQE). A low signal-to-noise ratio, combined with high-contrast performance results in a high DQE (and a superior image). As with the other methods, a system’s DQE is typically measured across a range of frequencies input (such as by using a standard resolution chart), and the results allow direct comparison between different systems.

5.2.8 Compression Ratios

Certain digital and analog compression methods (such as video compression or JPEG digital compression) discard a set amount of information (and by extension, reduce quality by a proportional amount). While some of these methods are designed to discard information that is imperceptible to the human eye, the image is still considered degraded. Digital systems tend to measure compression as a percentage of the available information, so that a lower percentage is more degraded than a higher compression. The same is also true of video compression, which describes compression ratios in terms of YIQ or YUV levels. Video compression is discussed in Chapter 2 and digital compression is discussed in Chapter 4.

Lossless Compression

Interestingly, “lossless” digital compression isn’t a particularly popular method for storing digital images in the digital intermediate environment, despite the fact that it reduces each file’s size without any quality degradation whatsoever. There are several reasons to explain this:

  1. Corruption. If a digital image that uses lossless compression is damaged, even by a single pixel, the damage can be propagated over a much larger area, potentially destroying the entire frame.

  2. Budgeting. The size of files that use lossless compression vary depending on the information contained within them. For this reason, it can be impossible to predict how much disk space is required for a given number of digital images that use lossless compression, which, of course, makes it difficult to budget disk space for a project.

  3. Speed. Additional computation must be performed to compress and decompress digital images. The performance difference might be insignificant for a single frame, but when trying to display a sequence of images at a rate of 25 frames per second (or higher), the computations can add up and can be an unnecessary usage of computer resources. What is unclear is how this performance hit may be offset by the associated gain in transfer speed (i.e., the files are smaller and can therefore be transferred faster).

  4. Compatibility. Use of any compression method makes the file format more complicated. In general, it can be relatively simple to extract image data from simple file formats, particularly when they’re stored using the RGB color model, even if the specifics of the format are unknown. However, as soon as compression is used, the exact specification of the file format must be known before the image data can be accessed. The reason this is an issue is because all the software used in the pipeline must be able to process files according to the exact specification. Given that it’s possible that a file format (such as TIFF) might support multiple compression methods, it becomes likely that at least some of software won’t be able to access a given compressed file format. Furthermore, it means that long-term storage of compressed files may not be accessible in the future, particularly if the compression standards change (or are made redundant). The most efficient use of lossless compression probably lies in transferring digitalfiles over relatively long distances. For instance, if a shot needs to be transferred across an Internet link (such as by virtual private networking—VPN—or file transfer protocol—

FTP), then the system can be designed to take each frame, compress it via lossless compression, transfer the file, and then uncompress it at the other end (probably combined with some form of digital verification method, which is covered in Chapter 6). Also, lossless compression is ideal for short-term data backups because it can reduce the time both to store and to retrieve a backup, as well as the amount of storage space required.

5.2.9 The Eye of the Beholder

With all this discussion about image degradation and quality, it’s important to remember that the human eye itself isn’t a perfect optical system. Many of the shortcuts used by video-and digital-imaging devices are based upon inaccuracies in the human eye. For example, the human eye is much more sensitive to red and green light than blue.4 Also, due to the large number of rod (luminance-sensitive) cells compared to cone (chroma-sensitive) cells, humans can more readily detect changes in luminance than in color. For this reason, many devices (in particular, video systems) are optimized for these parameters. Because of this, it’s important to remember that even if some degradation occurs, it may be within the parameters of degradation experienced by the human eye anyway and is thus undetectable to end users.

There are two main reasons why you might want to keep information that is imperceptible to the human eye. The first is machine vision (which is less relevant to the entertainment industry) and the second is image manipulation, such as digital color grading (which is covered later in Chapter 8). Ultimately, it’s desirable that a pipeline not suffer any degradation at all, whether or not it’s perceptible. In fact, selective information loss can be more of a problem than overall degradation—first, because it means an image may not look as degraded as it is, and second, because it degrades the affected parts of the image more (which then means that those parts of the image don’t respond well to manipulation).

5.3 Media Acquisition

There are two distinct ways to acquire media for inclusion within a digital intermediate: by electronic transfer or by analog-to-digital conversion. These methods vary in efficiency, speed, and quality.

If the source material is already in a digital form, it simply needs to be transferred from the original storage device into the storage devices used for the digital intermediate pipeline. If the material exists in some other form, then it must first be digitized (or converted to a digital format).

5.4 Data Transfer

A data transfer involves taking a set of images that are already in digital form (or taking some other data such as audio tracks or timecode information) and copying or moving the associated files from one device to another. For example, if you’ve shot something on a DV (digital video) camera, you can just pop the tape in a DV tape deck connected to a digital-editing system via a firewire cable, and all the data can be copied digitally from the tape, ready for editing.5

This process becomes complicated because there are many different ways to store, or encode digital information. In the same way that different languages can be used to describe the same things, different images can be recorded in a variety of different ways. It’s often necessary to perform additional data management, such as file format or color-space conversion during or after transfer. The transfer method might use a transcoding process, which effectively reinterprets the images from scratch, and thereby subjects them to further degradation.

A Perfect Copy Every Time?

Just because a transfer is made digitally, doesn’t mean it isn’t subject to quality loss. In many cases, a bit-for-bit copy is made in the same way that files can be moved about on a hard drive. However, this may not be the case in several cases, such as copying files between different operating systems (although even in this instance the data should have the same content) or when transferring data between different software packages or devices. In some of these cases, the information may be “transcoded,” potentially resulting in a loss of quality. Always be sure to check that data transfers are bit-for-bit whenever possible.

Data conversion notwithstanding, digital transmission allows data, such as images or video streams, to be copied or moved over long distances, with no loss in quality at all. For this reason, data transfer is preferable to every other acquisition method. However, everything is ultimately sourced from analog media, so the ideal scenario is to keep everything digital beginning with the step of analog-to-digital conversion and continuing onward. Whether the analog-to-digital conversion happens in the camera, as in the case of digital cameras, or after shooting, as when scanning film negatives, everything should be digital from that point onward.

An interesting factor affecting data transfer is speed. Video is always copied in real time (i.e., it takes one minute to copy one minute of footage, and so on). Video also has an advantage because it can be copied to many tapes at the same time (to as many recorders as available) through a process of signal amplification. Film can be duplicated at a rate of hundreds of feet per minute.

Data, however, is transferred at varying speeds. The transfer speed completely depends on a number of factors, such as bandwidth, network traffic, type of cabling, speed of the host computer, and so on. In some cases, it’s possible that a digital video sequence can be copied or moved faster than real time, and in others, it may be much slower than real time. It’s also possible that a transfer may begin rapidly and then slow to a crawl. The important thing is that each copy is identical to the original. Further, a single piece of data can be accessed by multiple devices at the same time, which is a (theoretically) more efficient way of working. Given a reel of film or a video tape, only one person can modify the contents at a time. But the digital equivalent allows each frame to be treated and modified independently, and in some highly specialized applications, even a single frame can be modified by different people at the same time. There are some pitfalls though. Certain applications “lock” files so that they cannot be accessed by others, to avoid reducing the speed of throughput. Locking files is especially common on softwarebased grading systems, which prevent frames that are in use from being copied or modified from outside the grading system. Although file locking is sensible, it can cause problems (e.g., when data transfers overrun their allotted time, as when the process of copying a number of frames overnight hasn’t been completed by the time the colorist starts work the next day), or when the system contains software errors, requiring it to be rebooted to “unlock” certain files.

Some digital media formats (e.g., DV or Internet streaming) are specified to work at a preset speed. In the case of DV video, for example, transfer from a DV tape to a capture system always occurs in real time. If the transfer speed drops during the capture process, the process might halt, normally reporting “dropped frames” or some other error, and have to be restarted. Internet streaming is designed to work at a given transfer speed (normally significantly lower than the maximum transfer speed) and will continue to transfer video at speeds above that rate—although the end user may experience viewing problems (e.g., slow downs or dropped frames) if the average transfer speed falls below the target rate.

A computer accessing its own files uses data-transfer operations. For example, displaying a digital image on a screen typically involves several processes: the data is transferred from the source storage device to the computer system through a number of cables and junctions, into the computer’s RAM. From there, it’s transferred to the computer’s graphics subsystem, which in turn transmits the data to the monitor through another cable. This description somewhat oversimplifies what exactly happens; in practice, even more factors play a part—for instance, displaying an image from a network location rather than a local disk drive necessitates even more transfers, through more cables and numerous network routers.

Data-transfer operations are normally used only for computergenerated or digital video productions, because most footage is generated using analog means, such as shooting on film, which is unfortunate because it’s the most convenient method for acquiring images. As digital filmmaking matures and gains popularity, the data-transfer method of acquisition will become more common. Until that time, the majority of footage will be fed into the digital intermediate pipeline using digitization methods.

5.5 Digitization

Digitization is the first step in converting an analog source to a digital format. It is currently the most common method for image acquisition in filmmaking. Almost everything is sourced from an analog format at some stage, with the notable exception of computer-generated imagery.6

To produce a digital version of anything analog, an analog-to-digital conversion must be made. Without question, this event ultimately has the most impact on quality in any digital pipeline (coupled, of course, with whatever was shot in the first place). From this moment on, you can only reduce quality—purposefully in many cases, but irreversibly nonetheless. How you go from an analog source to a digital source has a bearing on every subsequent stage, and therefore this process must be planned carefully in advance based upon the requirements of the film.

5.5.1 Sampling

The process of digitization is a statistical reduction of information. Analog sources effectively have an unlimited level of detail. The closer you examine any analog source, the more detail you find. Some of that is “useful” detail, providing more of the image content, and some of the detail is just a factor of defects, such as noise. The first step in digitization is to decide how much of this detail you actually need. It isn’t possible to capture absolutely everything, but in most cases, it isn’t necessary. In the same way that the attitudes of an entire population can be determined by interviewing a small, carefully selected “sample” of individuals, an analog source can be sampled to build a complete picture. An entire branch of mathematics, sampling theory, is devoted to methods for doing this, but the basic idea is that if you make enough samples of an image at regular intervals, you’ll be able to re-create the original faithfully.7 This means that you start with an analog source, which has an infinite level of detail, and you split it into parts.

Imagine, for instance, that you’re painting a scene that includes a tree. Let’s say you use a large paintbrush, one that can cover the entire canvas in a few strokes, and a pot of black paint. After maybe four strokes, you might have a vague impression of the tree. Now let’s say you use a smaller brush (or a bigger canvas). It takes a lot more strokes to paint a picture of the tree, but now you can paint more of its subtle details. But it’s still not an exact image of the tree. Let’s now assume that you start the painting again, but this time with a few different colors of paint. You can even mix these colors to create new colors. This provides an even more accurate image of the tree.

What you’ve done, in a very loose sense, is digitized the tree onto the canvas. Even with a very thin brush, you can’t replicate all the detail of the scene. The more carefully you inspect the scene, the more details become revealed. But at some point, you decide that the level of detail is close enough to make no visible difference. For example, make the canvas as big as a cinema screen, and the level of detail high enough that your eyes can’t actually perceive all the detail from a few feet away. Then say that you don’t want to be able to perceive specific points where one color of paint begins and another ends. Maybe now you have something that approximates a projected photograph.

Anything analog has to be digitized somehow so it can be replicated in another form. Take photographic film, for instance. The detail level of a scene imaged on a photograph is determined by the grain size (among other things) of the film. You can think of taking a photograph as an analog-to-analog process, which digitizes a scene at the level of the grain size.8

As another example, consider the way that the moving picture actually works—by projecting a series of stills at a rate of around 24 frames a second. What has happened is that time itself has been digitized by this process, breaking down a continuous (analog) motion into discrete (digital) parts. But when a moving picture is played back, we perceive the motion as smooth, because it occurs at a slightly higher rate than our eyes can detect.

You can use many different ways to digitize an image, but the most common method is to break down an image into regularly arranged, repeating squares. There are many reasons for doing it this way, but the main reason is that it allows the data collected to be identical to how digital images are stored and displayed, resulting in no necessary further degradation of the image.

In the simplest case, a single point of a source is sampled, which then corresponds to a single pixel of the digital image. This process is repeated at regularly spaced intervals, horizontally and vertically, until the entire image is digitized.

Are Pixels Best?

Surprisingly, the use of a grid of square-shaped pixels is a bad way to create a representation of an image. One study has shown that digital images have a higher MTF when rotated 45 degrees during digitization (but the downside to this method is that rotating the image back for viewing purposes degrades it, so destroying the quality gain of rotating it in the first place). However, a so-called “honeycomb” structure of digital images, where the pixels are diamonds rather than squares, produces visibly sharper results (both objectively and subjectively) than a regular pixel arrangement.

The main reason that pixels are square is because it simplifies the mathematics involved in displaying and manipulating digital images. And now, everything, from software to monitors, is designed to work with square-pixel images, so it’s unlikely to change in the near future. The notable exception is that it’s possible to have nonsquare (meaning rectangular) pixels in some cases. Images with nonsquare pixels work on the basis that you can just add or remove a certain proportion of square pixels to enable them to be viewed at the correct ratio on square-pixel devices. (See Chapter 4 for more information about nonsquare pixels.) One of the reasons that grain-for-pixel, film still produces a superior image is because photographic grains are randomly shaped and randomly distributed, making them less vulnerable to effects such as aliasing.

images

Figure 5–6   Digital images may be created by sampling an analog source and then encoding it into a digital format

5.5.2 Interpolation

Digital images can change resolution as required, through a process called “resampling.”9 For example, a 200 × 200 image can be resampled to a 100 × 100 image or to a 400 × 400 image. Increasing the number of pixels is referred to as “upsampling” (or “uprez’ing”), while decreasing the pixel count is “downsampling” (or “downrez’ing”). Resampling is used for a variety of reasons, often so that different footage is processed at the same time (so that video footage can be recorded onto film, it must first be upsampled to match the film image resolution). However, increasing the pixel resolution in this way won’t increase the image’s level of quality (though decreasing the resolution will cause a reduction of quality). This is because you can never get additional quality from a digital image that wasn’t captured in the first place.

To resample an image, a calculation must be made to work out what the new pixels look like. This process is known as “interpolation.” There are different interpolation methods (or algorithms) for different situations, which are explored in Chapter 10. Interpolators merely estimate how additional detail would look (in the case of upsampling) based upon existing pixels. When downsampling, estimation is less an issue, because a reduction in quality doesn’t have to estimate detail that doesn’t exist. However, upsampling should be avoided whenever possible.

Digital Zoom

Certain digital-imaging devices have a digital zoom function. This function is just a method of upsampling parts of an image, to make those parts appear larger. However, there is no quality benefit to using such features, especially because they can be replicated later in the digital intermediate pipeline, using higher quality interpolation algorithms.

images

Figure 5–7   Upsampling artificially increases the resolution of an image, while downsampling decreases it. © 2005 Andrew Francis

Quality and Wine

Digital image quality is analogous to a bottle of wine. Imagine that your original scene is a vat of wine, and you fill a bottle with a sample. The size of that bottle is equivalent to the quality of the imaging format, so that 35mm film is equivalent to a large bottle, whereas VHS video is a much smaller bottle. Between the time the wine is in the vat, and the time you get to pour yourself a glass, the wine might have been transferred to many different bottles. Anytime it’s transferred to a smaller bottle (or put through a quality-reducing process), some of the wine will be permanently discarded. Similarly, transferring the wine to a larger bottle doesn’t give you any more wine, but you don’t necessarily lose any either. Interpolation is like watering down the wine: you can use it to fill a larger bottle, but it won’t give you any more wine.

5.5.3 Color Space

So far, we’ve mainly looked at the RGB model of digital images, because the most common types of digital image are of this model and are ideally suited to being viewed on an RGB display device, such as a computer monitor. However, there are many different ways of using digital images, other than displaying them on a monitor. Digital images can be printed on paper, put on video, recorded by lasers onto photographic film, or even just analyzed by machines without ever being looked at.

Each of these different methods understands color in a different way. For example, many colors that can be seen on projected film can’t be displayed on video (and vice versa). The range of colors that a system is able to reproduce is known as its “color space” (or “gamut”). Most of the time, the color spaces of different systems overlap, and the systems can show the same color. So, for example, an image displayed on a (properly calibrated) monitor will look exactly like the color print. However, where areas of color don’t match, the colors are said to be “out of gamut” and may look wrong.

Even the human eye has its own gamut. For example, some video cameras can detect infrared or ultraviolet, which is outside of the gamut of the human eye. However, the color space of the human eye is much larger than most color-reproduction systems, which is why we can easily detect differences between colors viewed in different mediums.

Digital images are often optimized for a specific color space. This means that the images may be interpreted a certain way, depending on their application. For example, the common three-channel RGB color image discussed earlier, aims to be suitable for viewing on most computer monitors. Printing houses may adopt the four-channel CYMK color space, which is interpreted in a similar way to the cyan, yellow, magenta, and black inks used for printing on paper. Images for photographic film typically use a three-channel RGB color space but by using a logarithmic scale. This is done for efficiency because film responds to light differently at the extremes of color than in the mid-tones. This topic is covered in Chapter 8.

5.5.4 Nonlinear Color Space

Certain materials, including photographic film, don’t respond to light in a linear fashion. That is, adding twice as much light to a photographic material won’t necessarily result in an image that is twice as bright. For this reason, digital images can be stored in a nonlinear fashion. With linear-color-space images, pixels are encoded with brightness values on each channel. Nonlinear color spaces work this way too, but the difference is that the values are mapped onto another scale, so that values of 0, 1, 2, 3, etc. might be mapped to brightness values of 0, 1, 4, 9, respectively (in this case, an exponential color space is used) instead. This method makes it possible for very high values and very low values to be stored together in the same image, resulting in more efficient use of disk space.

Where problems occur with nonlinear-color-space images is in trying to convert data between different spaces, because it can be difficult to correlate colors from one color space to another. Color-space conversion is covered further in Chapter 8.

5.5.5 Gamma

Although RGB provides a suitable model for displaying images on a number of devices, such as televisions and computer monitors, there is no direct correlation between the brightness of a pixel of an RGB digital image and the corresponding brightness of a point on a monitor. Part of the problem is that monitors have a luminance range that is nonlinear, but another problem is that there is so much variance between monitors.

The first problem is resolved by applying gamma correction to the image. In gamma-correcting an image, the luma signal sent to the monitor for display is adjusted so that it’s represented on a nonlinear scale, compensating for the response of the monitor (so that the image looks as it should, at least in theory). Gamma is a specific scale that combines both contrast and brightness. The amount of gamma to be applied is variable, with a value of 1.0 making no difference to the input and output of the pixel values, and higher and lower values increasing or decreasing the gamma correction, respectively. This is complicated somewhat by the use of different standards for different applications. For example, older Macintosh computers use a gamma of 1.8, whereas Windows computers typically use a value of 2.2.

Because of the variance between different monitors (and lighting conditions), the gamma correction must be adjusted for each monitor for accurate results. The most common way of doing this is using a gamma-calibration system. Using such a system, a series of readings of the response of the viewing device to different images and luma values is made, and the optimum gamma correction is determined. Many digital images may also have additional gamma information encoded into them, so that they can be displayed more accurately on other calibrated systems (although such embedded information is useful only if the originator of the material was working with a calibrated system). During acquisition, gamma correction may be applied automatically to certain types of media, such as video or digital images from elsewhere, and can affect the images’ colors. Many digital intermediate pipelines try to avoid altering color content in any way prior to the color-grading stage because such alterations may degrade the images. Color-grading and calibration are covered further in Chapter 8.

5.5.6 Dynamic Range

The intensity range of an image is the ratio of its brightest point to the darkest point. The density range of a photographic image is the ratio of the highest density on the film (the D-max) to the lowest (the D-min). Either of these terms can be taken as the dynamic range of an image. In terms of digital image formats, the dynamic range of the image can be derived from the bit depth: an 8-bit image, having 256 values per pixel, has a dynamic range of 256:1. A (linear) 10-bit image has a dynamic range of 1024:1. The dynamic range of the human eye is roughly 10,000:1, whereas most CRT displays are approximately 100:1. Televisions are around 30:1 or 40:1. Where this becomes important is in trying to capture the available intensity range from another source (either analog or digital). If the dynamic range of the capture device (such as a film scanner or digital camera) is lower than the dynamic range of the scene being imaged, some tonal information will be lost. If the dynamic range of the image format being used is lower than the information being captured, some information will be lost. Similarly, using a file format with a high dynamic range doesn’t necessarily mean that the image itself will have a high dynamic range.

There’s a subtle difference between dynamic range and color precision. Color precision is a purely mathematical concept, determined by the bit depth of an image. A 10-bit file has greater precision (and is therefore less prone to color artifacts) than an 8-bit file. On the other hand, a nonlinear 8-bit image might have a greater dynamic range than a linear 10-bit image, meaning it could reconstruct a greater range of luminance but at a lower degree of accuracy.

Color Versus Resolution

If you look at the specifications for most digital-imaging devices, the pixel resolution is always the most predominant factor. However, it isn’t necessarily the most important. As discussed previously, resolution is theoretically equivalent to the system’s resolving power, although this isn’t necessarily the case. In practice, the question of image quality relies on two main factors—the (true) resolution of the image and its color range. It’s not clear which of these two values is more important, although there seems to be an upper limit on the “useful” color range. Essentially, if the color range is high enough to allow color grading without introducing banding or other artifacts, then increasing the color range further has no benefit. Current tests determine this limit to be about 16 bits per channel for RGB channels, giving approximately 280 trillion possible colors for each image. All things being equal, this means that if the imagecapture device already has a sufficiently high color range, the determining factor is going to be resolution; images can always benefit from extra sharpness. However, too often, a high resolution is quoted alongside a substantially low color range, especially in the case of digital cameras.

Another caveat to this argument is that the color range is inherently limited by the viewing device. Even the best monitors are configured to display a maximum of 10 bits per channel, which means that anything higher can’t even be displayed properly. Similarly, other output devices such as film recorders may have a maximum color range on a similar scale. Therefore, even an image with 16 bits per channel may have to be reduced to a lower color range before printing (but the higher range will still help to reduce color artifacts). On the other hand, there isn’t really an upper limit on useful pixel resolution. Even if an image has more pixels than can be displayed at once, it still has the benefit of being able to be viewed at different scales. For instance, you can look at a large image, viewing it at the full level of quality (one pixel of image occupies a single pixel on the display) so that you only see a quarter of it at any time, or you have the option to zoom out the image, reducing the number of pixels displayed but enabling you to see the entire image. In terms of color range, there is no real equivalent. You could opt to view a section of the color range at once, such as having a virtual exposure control to expand the highlight or shadow areas as needed, but such controls aren’t particularly common.

This limitation is a bit of a setback for the digital intermediate pipeline, because if devices were able to register or print a greater color range, it would contribute to a higher level of detail in an image with a reduced number of pixels. If you consider the edge of an object being imaged, such as the leaves on a tree, a capture device would better reproduce the veins on the leaves if the sample pixel count was increased. However, it’s also feasible to assume that a more accurate color sample would also improve the rendition of detail, because each pixel would be afforded a slightly more accurate “average” of the color of the pixel. Clearly this isn’t a perfect solution, because you would also benefit from a reduction of aliasing artifacts by having a higher pixel count. However, it’s worth noting that in terms of file size at least, doubling the number of colors (e.g., going from 256 colors per pixel to 512 colors) requires only a 12.5% increase in file size (the increase of 8 bits per pixel to 9 bits per pixel is 12.5%), whereas doubling the number of pixels doubles the file size.

For moving images, a third factor is also relevant to quality: time, or the number of images per second. Currently, this factor doesn’t really undergo significant change during the intermediate process (with the exception of format conversion or motion effects), because there isn’t as much stretching and squashing of time as with colors or pixels. There may be a point in the future where filmmakers experiment more with the process of manipulating the speed of a shot during the digital intermediate process, which may require scenes to be filmed at a higher frame rate to allow for this possibility to work well. For now though, such shots are usually planned well in advance, or the filmmaker uses interpolation methods.

What tends to happen is that the digital intermediate process uses the same frame rate as used by the majority of the source material (such as 24fps for film productions, or 29.97 for NTSC video productions), and everything else is changed to match it. Ultimately, the use of this method will normally be decided by the lead editor on the project.

5.5.7 Acquisition for Digital Intermediates

The aim of acquisition for the digital intermediate pipeline is to obtain, whenever possible, digital images whose quality isn’t compromised in any way. Current technology dictates a trade-off between speed and quality, particularly in analog-to-digital conversions. However, most digital intermediate facilities believe that quality must take precedence over speed at this stage, especially when dealing with the high-quality demands of cinema or (to a slightly lesser extent) high-definition video formats. The quality level set at the stage of image acquisition will have a profound effect on the quality through the pipeline, regardless of the target quality of the final product. This is true of many mediums—for instance, a VHS video originated from a HD video source will look superior to a VHS video originated from a VHS source, even though the content might be the same. There are other considerations—for example, the quality level shouldn’t be so high that it impacts the performance of the playback system. Most facilities require that all screenings for the production team, or any interactive sessions (such as color grading) be done at the maximum level of quality in real time.

5.6 Acquisition from Video Sources

Acquisition of digital images from video sources is actually a fairly straightforward process. All that is required is a video playback device, an analog-to-digital converter, and a data storage device. The quality of the conversion depends entirely upon the analog-to-digital converter (ADC).

images

Figure 5–8   Digital media may be acquired easily using a data-transfer method, while analog sources must undergo a digital-to-analog conversion

Video signals are formatted in a way that translates to digital images very easily. Depending upon the video format, the necessary resolution can be derived that corresponds exactly to the number of lines in the image. In this case, no additional benefit is derived from digitizing the video to a higher resolution (because there will be no additional information to record, unlike for photographic sources). Both YIQ- and YUV-encoded colors can be converted to RGBencoded ones, although some RGB color spaces have different gamuts than the video ones.10 A higher bit depth may provide a quality advantage in terms of color precision, because a wider variety of colors can be sampled, although this will inevitably depend on the precision of the source video. However, many video formats inherently use color compression to reduce the transmission overhead, and so the benefit may be slight. In most cases, it’s unlikely that any benefit will be gained from using more than 10 bits per channel RGB images.

In any case, provided the file format is appropriate, any reduction of quality will largely be due to noise added by the ADC and its subsystems. With most ADCs, the noise introduced will be negligible; otherwise, a noise-reduction technique, such as image averaging, can be used (see Chapter 9 for more on this technique). A table of digital resolutions for various video formats is included in the Appendix.

A simple, effective system for digitizing video sources is a dedicated computer workstation with a video capture card. The source deck (or video matrix, in facilities with a large number of video decks) connects to the capture card, and software within the computer saves the incoming video as a series of still images onto a local disk or network. Video is normally transferred in real time, meaning it takes one hour to digitize an hour’s worth of video.

5.6.1 Digital Video

With digital video formats (such as DV, DVCAM, and HD video), the video is already stored in a digital format. Therefore, acquiring video of this type requires no analog-to-digital conversion and is usually facilitated by directly transferring the data from the tape deck to the host computer (such as with a firewire connection). When transferring digital video this way; there will be a much lower degree of quality loss compared to transferring video using analog means.

Digital Tape Damage

In the event that the tape has been damaged too much, rendering some of the data inaccessible, portions of the footage will also be damaged, or corrupted. This problem is commonly seen when large blocks of the picture are missing (called “picture dropout”), or when playback isn’t smooth (in this case, frames are said to have been “dropped”). Digital formats are inherently more robust than analog ones. However, there is a point where errors can cause digital formats to suddenly break. This is entirely dependent upon the ability of the system to access the data correctly.

images

Figure 5–9   When using digital tapes, a single tape error can impact a sequence of frames

Certain digital formats, especially those with a high level of compression, have a further weakness in this area, where a single data error could destroy an entire image or movie sequence. This problem occurs because some formats work by using interframe encoding, storing the data with references to past or future data. In this case, if some data is damaged, some or all of the data referencing the damaged data may become damaged as a result.

As an analogy, imagine that you stored your reels of film in a number of locked vaults. If one of the keys to the vaults became damaged, you wouldn’t be able to access any of the film within that vault. It’s important to remember that just because something is stored digitally, it isn’t impervious to physical damage.

The process isn’t exactly the same as copying a set of digital images from one disk to another, however. Digital video is always transferred in real time, and thus, any read errors can’t be corrected by attempting to re-read the data from the tape, which is why some errors may be introduced during capture. However, this disadvantage is the only real weakness of digital video capture, and in general, very highquality transfers can be achieved. In the event that the errors are significant, problems may be corrected by redigitizing the tape (assuming the tape itself is not damaged).

5.6.2 Timecode

Video timecode is often used to determine the parts of a video tape that have to be captured. For example, a video tape may contain 90 minutes of footage, but only 2 minutes are actually needed. In this case, the capture system might allow for the capture of video starting at a timecode of 01:00:00:00 (the “in point”) and ending at 01:02:00:00 (the “out point”).

To enable video footage to be referenced correctly later on, it’s often desirable to capture this timecode information along with the image content for every frame. Capturing this information is usually done by using a simple cable between the capture system and the VCR, but the information somehow must be tied with the image data. Certain digital file formats, such as MXF, allow timecode to be embedded within the file, but for less-sophisticated formats, a common method is to convert the timecode to a frame number, as covered in Chapter 6. Even for systems that have no capability for capturing timecode, the frame-numbering method can be used to retrospectively attach timecode to captured footage by renumbering the data.

In some cases, it may be necessary to include additional information, such as the tape reel number or the date of shooting.

5.6.3 Frames Versus Fields

One stumbling block in video acquisition is that many standard definition video formats are interlaced, which means that each frame of video actually consists of two interlaced fields. Typically, each field is recorded with a slight delay in between. This means that a frame of a fast-moving object might be at a slightly different position on each field of a frame, and that visually, the object will appear to be shifted horizontally on alternate lines (which can look similar to aliasing artifacts). If the final output is a video format anyway, this doesn’t necessarily matter (because it would be split back into fields anyway). But often, especially when mastering to film or other formats, it’s desirable to remove the interlacing motion artifacts on each frame. The methods for doing this are covered in Chapter 9.

Progressive Scan

Many modern video cameras (especially HD cameras) have the capability to capture video in a “progressive scan” mode. Rather than record two separate fields for each frame, a single whole frame is recorded. This is a preferable capture method for digital imaging as it eliminates the need to later de-interlace them. Progressive scan video formats are often denoted with a “p” following the frame rate, such as “24p” (interlaced formats are followed with an “i”, such as “50i”. It is also worth noting that interlaced formats don’t necessarily have interlacing—for example, a set of (progressive) digital stills can be output to a 50i video format, simply dividing each frame into two fields. Effectively, each frame is identical to a 25p equivalent; it’s just stored in a different way.

5.7 Acquisition from Photographic Sources

At present, an undeveloped piece of film can’t be digitized. For this reason, the initial step in the digital intermediate chain is to develop the photographic material. This is an inherently chemical process and should be done as soon after exposure as possible, to prevent the risk of “fogging” or otherwise spoiling the film.

Once the film has been processed, the bulk of image acquisition involves using specialized film scanners to digitize reels of film, creating a series of stills that correspond to each frame on a reel. Depending upon the particular digital intermediate pipeline you’re using, you may wish to scan everything that has been shot as soon as the negative is developed, or (if the film has already been completed in a chemical lab environment) to scan a theatrical release print (e.g., for transfer to video or digital master). A common method at the moment is to scan “master negative reels,” which are reels that comprise only the selected takes that are to be used in the final edit.

5.7.1 Film Scanners

Regardless of the paradigm, the process of digitizing film is the same. The basic methodology of a scanning process is that a reel of film is loaded onto the film scanner (in a way similar to using a projector or an ultrasonic cleaner).11 The film is scanned by illuminating the surface, usually with a very intense light source (such as a metal halide lamp). The light is then measured in terms of transmittance (i.e., how much light passes through to the other side), typically a line at a time by a trilinear CCD array (a single line of photosensitive CCD elements for each channel of red, green, and blue light) or by some other method, depending upon the scanner configuration. Each point measured, or sampled, along the line corresponds to a single pixel in the image, with separate measurements for the red, green, and blue light. Motors in the scanner advance the film to the next line, and the process is repeated. Finally, when the entire frame has been scanned, the image is saved, and the scanner advances to the next frame. It can take anywhere between 4 and 320 seconds to scan a foot (i.e., 16 frames) of film, depending upon the resolution and type of scanner.12 It is interesting to note that the maximum color precision (or bit depth) of the scans doesn’t affect the scanning time at all but is instead dependent upon the sensitivity of the scanner.

images

Figure 5–10   Film material must be developed and then scanned before it can be used in a digital intermediate process

Most high-end film scanners are made to exceptionally high optical standards and don’t introduce any significant artifacts, such as noise. In addition, scanners don’t capture motion; they merely record a sequence of stills. The distinction is important because it means that they don’t introduce (or remove) any additional temporal artifacts (such as motion blur) that weren’t originally recorded. However, spatial and chromatic artifacts may still be introduced as a result of the analog-to-digital conversion. If the sensitivity of the scanner to color or if the scanned pixel resolution is too low for the type of film, artifacts such as banding or aliasing may occur.

With current systems, the color information contained on a frame of film is higher than can be captured by most scanners. Most film scanners capture 10-bit (linear or nonlinear) Cineon or DPX files (see the sidebar, “Cineon/DPX Images,” later in this chapter). Ideally, nonlinear log (or logarithmic) files are preferable because they provide a better representation of film color space. If the color-correction, conforming, and finishing systems aren’t designed to handle log files, it may be better to scan to linear color space files, avoiding problems later on. Even with log files, it isn’t possible to capture the entire dynamic range of film with most image formats.

Either way, the scanner process should be set up to maximize the contrast range of the images captured, so that the brightest point on the film (i.e., the white point) is set to a value just under the maximum brightness value of the file, and the darkest point on the image (i.e., the black point) is just above a value of pure black. Note that values that are outside the picture area (and therefore irrelevant) can safely be ignored. Maximizing the images’ color ranges ensures that no clipping or crushing of colors will occur and affords the maximum possible flexibility and quality levels later on in the process. Most scanners have this capability, and settings can usually be made on a shot-by-shot basis.

The usual work flow for the scanner operator, having loaded the film onto the scanner, is to create black-and-white point settings for each shot on the film, while playing through the reel at high speed. (These point settings are typically stored as a small “grading list” that can then be saved if needed later—e.g., for rescans. Most scanners are able to display the film on a monitor at a faster rate than the time required to digitize the film to data files.) While it’s possible to set these levels visually (in a properly calibrated environment and with a highly experienced operator, at least), it’s preferable to use more objective means to determine these settings. The most common method is to use a vectorscope, which can measure the intensity of the video signal generated by the scanner and match the peak white level to the white point of the output image (and do the same for the black point). Some scanners provide this information automatically, by taking density readings across an image.

images

Figure 5–11   Filmlight’s Northlight scanner is able to calculate the scanning parameters by analyzing the film

Once the levels have been set, the scanner returns to the start of the reel and digitizes each frame, applying the point settings, typically generating a series of digital images to a local disk drive. With certain scanners, such as Filmlight’s Northlight scanner (www.filmlight.ltd.uk), shots don’t have to be set up this way. Instead, such scanners employ an “unattended scanning” paradigm, which means that once the film is put onto the scanner, the scanner identifies the base density of the film (either by matching stock presets from the key numbers or by taking a reading in between frames), calibrates itself, and digitizes the film without having to set up grading parameters for each shot. This capability is ideal because it guarantees that the film is captured quickly, with the maximum possible color range for the film stock. This system may fail, however, when scanning a reel made up of different film stocks spliced together.13

5.7.2 Scan Dimensions

Unlike video, there is no perfect translation between the dimensions of photographic film and digital images. Photographic grains are randomly shaped and randomly arranged, unlike the precisely structured video image. It is therefore reasonable to suggest thatscanned images can be any resolution. It can also be suggested that the ideal image resolution should be as high as possible, while remaining practical to work with (i.e., not taking too long to process, copy, or display). Other considerations include the possibility that the film scanner color-correction and conforming system may be limited to working with images of a specific or optimum size.

Cineon/DPX Images

Cineon and DPX files are currently the most popular formats in the digital intermediate environment. The two formats are almost identical (i.e., the information they contain is exactly the same, they’re just structured slightly differently); the Cineon format is a slight modification of the Digital Moving Picture Exchange (DPX), SMPTE-specified format.14

Cineon and DPX files are typically 8 bits or 10 bits per channel (the latter is more common) of three-channel (RGB) data, which can be encoded as either linear or nonlinear RGB and are uncompressed. The linear color space has applications for storing video, while the nonlinear color space is more suitable for representing film. This duality makes the two formats very useful to a digital intermediate environment, which frequently switches between film and video sources. The nonlinear color space is designed to match the characteristics of photographic film by modeling transmission density as the brightness component of each channel. Photographic film doesn’t respond in a linear way to light; it responds logarithmically, so that more of a brightness (or contrast) range is at the extremes (blacks or whites) of a negative. Images saved in this logarithmic (or log) color space are a much more accurate representation of the film image they were digitized from (i.e., having a higher dynamic range); however, they won’t display properly on a linear device, such as a computer monitor. To compensate for this, applications working with log Cineon or DPX files must employ special lookup tables (or LUTs) to alter the way the images are displayed on a monitor (but without actually affecting the data).

HDR Images

High dynamic range (HDR) images can overcome many of the problems of moving between different color spaces. HDR images encompass a very large and very precise color space (some implementations are large enough to include every useful color space that exists) and utilize tone-mapping functions to correctly translate between different display devices. At the present time, the digital intermediate pipeline is concerned only with the color spaces of photographic film, of the Cineon/DPX image format, and of the RGB model (for computer monitors), but as this expands to include more applications, HDR images and tone mapping will play a more important role. HDR images may also preclude the need to perform a “one-light” grading pass on scanned film, because the format has sufficient precision to make this stage redundant, which will reduce the scanning time somewhat. (HDR images are discussed in Chapter 13.)

Of the several accepted standards, the most common is 2k and 4k resolution for 35mm negatives. This value refers to the number of pixels along the horizontal edge. A typical, full frame of a 35mm negative scans at 2048 pixels (the width) by 1556 pixels (the height), making it a 2k image (i.e., having approximately 2000 pixels along the width). A 4k image is 4096 by 3112 pixels in size (resulting in a file size that is four times larger for each frame). Different specifications apply to different formats (e.g., when you want to scan only the Academy area). These specifications are listed in the Appendix.

5.7.3 Nyquist Frequency

The Nyquist frequency (also called the “Shannon sampling frequency”) is a simple mathematical equation used to determine the amount of information required to successfully reproduce a signal (i.e., an analog source). It is used to determine the point where all the relevant information is captured (and so recording any more will have no benefit).

This equation—Nyquist frequency = 1 / (2 × sampling interval)—can be used when digitizing an image to determine the maximum useful resolution, using a measurement of the resolving power of the image. For example, for a typical 35mm negative, the resolving power is around 150 lines per millimeter.15 The Nyquist frequency, in this case, is equivalent to the resolving power (because no more information can be successfully recorded beyond this amount). Putting this value into the equation gives us a sampling interval of 0.0033, meaning that a pixel must be sampled every 0.0033mm (or that there must be 300 pixels for each millimeter of film).16 In practical terms, given that the height of a full frame of 35mm negative is 18.67mm, a completely accurate representation of a frame of 35mm film requires approximately 5600 lines of resolution. Assuming the resolving power of film is the same horizontally as it is vertically, our desired image size is approximately 5600 × 7500 pixels per frame, which is required to completely capture all of the detail in a frame of film. At 10 bits per channel, it would make each frame about 150MB.

Megapixels

A trend in digital camera manufacture is to talk about “megapixels,” which is a somewhat arbitrary measurement and refers to the total number (in millions) of pixels in an image. To calculate image size in megapixels, multiply the width of the image by its height, and then divide by one million. A 2k image is thus equivalent to a 3.2 megapixel image, and a 4k image is 12.7 megapixels.

Unfortunately, current technology simply cannot store or process files that large in a practical manner. However, this limit is theoretical; in reality the difference in quality between an image that large and a 4k image, for example, may be negligible. It is questionable whether that much detail really can be perceived, given factors such as film grain; and in any case, the ultimate question is whether an audience in a cinema is able to perceive any difference. The bottom line, at the moment at least, is to aim to try to capture as high a resolution as possible.

5.7.4 Punch Holes

Many scanner systems require that each reel of film includes a “punch hole” (or a marker frame at the beginning of each reel). The easiest way to make a punch hole is to literally punch out a hole in the negative at the beginning of the reel (before any useful images!). The reason for doing this is to provide an arbitrary sync point that will allow each frame of data to be matched back to a specific frame on the reel. In simple terms, each frame scanned is given a frame number, so that the punch hole is frame 0, and every frame after goes up by one. For example, the thirty-first frame after the punch hole is simply frame 31.

Enhanced 2K

In many cases, film scanners can scan images that are much larger than is practical to work with. For instance, many scanners can scan 4k images, but the constraints of the rest of the system might require images no larger than 2k. In this situation, it may seem more efficient to scan at 2k and keep everything at that resolution. However, tests have shown that the quality improves when the film is instead scanned at 4k and resampled down to 2k. Provided that the appropriate resampling method is used, the so-called “enhanced 2k” footage appears almost as sharp as 4k footage and is a vast improvement over a regular 2k scan.

Scan Edls

Lengthy reels of film, especially those with many shots, can require a long time to set up, even before anything is scanned. One of the reasons for the prolonged setup time is that the scanner operator has to go through the entire reel first, marking cut points to divide each shot. One way to speed up this process is through the use of a scan edit decision list (EDL), which essentially is a list of cut points for every shot on the reel. Using scan EDLs isn’t usually an option when the film has just been developed, but in productions where each reel of film has been telecined already, scan EDLs may be available and can reduce the length of time required to scan each reel. EDLs come in many different flavors and are discussed in Chapter 7.

5.7.5 Splices

When scanning film, you may encounter a problem if the reel contains splices. When a film splice runs through certain scanners, it can cause the frame to “bounce,” meaning that the image shifts vertically. This problem is less likely to happen with pin-registered scanners. Even when using scanners that are not pin-registered, the problem of bounce can be corrected digitally, but doing so is usually an operator-controlled procedure (rather than an automatic one) and as such can add to the cost and duration of the project. Perception of bounce is much greater on monitors than on projected images; therefore, bouncing shots may not be noticeable on cinema releases.

Additionally, the materials used to make a splice may intrude into the image area, damaging the image and requiring its restoration. For all of these reasons, it is preferable for productions to minimize the number of splices made to film material intended for scanning.

5.7.6 Datacines

In addition to film scanners, telecines are used to convert film reels to other media (typically to video). Telecines, particularly older machines, may perform an analog-to-analog conversion rather than an analog-to-digital one and work in real time. Film scanners traditionally differ from telecines in a number of ways. An important advantage of using scanners is that they have the capability to output to a digital format. On the other hand, telecines traditionally enable color grading and the viewing of film material in real time. However, in recent years, the differences between the two have become more subtle. Some telecines, such as Thomson’s Spirit 4K (www.thomsongrassvalley.com), can directly output high-quality data and are referred to as “datacines,” whereas some scanners provide rudimentary color-grading and display options. At the present time, the main difference between scanners and telecines seems to be that telecines are operated by experienced colorists, while scanners operate almost automatically.

Color grading performed at the acquisition stage may have less of a destructive effect on the quality of the images because most high-endscanners perform color grading prior to actually capturing the image, effectively recalibrating the scan parameters to maximize the captured dynamic range within the new color parameters. In fact, this paradigm is a perfectly viable one for the digital intermediate pipeline. Even so, most pipelines prefer to leave color correction to a later stage because doing so allows more flexibility and makes it possible to view changes in a more useful context, with all the material available at the same time.

images

Figure 5–12   Film frames are generally counted beginning with the first frame that follows a punch hole frame

Telecines are routinely used by the film industry to allow the viewing of rushes (or dailies) of film that has just been developed, although this practice eventually may be replaced by the advent of digital dailies.

images

Figure 5–13   Thomson’s Spirit 4K can scan film as well as telecine it

5.7.7 Digital Dailies

One of the advantages of digital media is that it can be viewed simultaneously from a number of places over a great distance. When shooting film, filmmakers typically send the day’s filmed material to be developed and then printed or telecined, or both (depending on their preference and the budget). The developed negative is stored, and the filmmakers watch the telecine or print to assess their work. However, as more filmmakers turn to the digital intermediate pipeline, the idea of digital dailies may become more attractive. As soon as the film negative is developed, it can be sent to the digital intermediate facility to be scanned. The scanned data can then be made available for viewing and copying. The negative can be stored safely (and theoretically is never needed again), and the data can be archived for post-production. The copies can be manipulated (e.g., compressed or encrypted) in many ways that don’t affect the original data, and they can be transmitted across the Internet, providing them to filmmakers, editors, and any number of people who may require an instant copy.

The Death of Film?

Ever since the boom of the digital image industry in the early 1990s, people have been wildly proclaiming an end to celluloid. However, these claims seem to be unfounded, at least for the time being. While it’s true that sales of consumer photographic film have fallen in recent years (reflecting the widespread use of digital imaging devices), sales of film for the motion picture industry haven’t fallen.

It is feasible that technology eventually will reach a point where digital imaging becomes technically superior to film in every way, but that point is still a long way off. In the immediate future, it’s likely that celluloid will still be used on the set as a general purpose recording device, although digital capture devices may be preferable for certain types of shots, such as effects shots. At the present time, photographic film produces subjectively better imagery than its digital counterparts. The current thrust of technology is to try to maintain the quality recorded on film from the point of digitization onward. In the motion picture industry at least, nothing suggests that film, as a shooting format, is due to be retired any time soon. Furthermore, photographic film is more reliable—experienced filmmakers know what is and what isn’t possible to do with film and are highly aware of the necessary precautions that must be taken when working with film. In the time-pressured environment of the set, producers cannot afford to take unnecessary risks, such as having a digital-imaging device fail for any number of reasons. In fact, when digital counterparts fail within the digital intermediate pipeline, producers are often relieved that they can fall back on the camera negative or source video tapes. In these early stages of digital film, a reel of film certainly has a greater lifespan than the latest digital storage device. A properly stored negative will last longer than a hard disk, because film is highly standardized as compared to computer components. Some computer hardware that is only a few years old already is incompatible with new systems. The same cannot be said for film, which relies on the correct positioning of the images, a constant frame rate, the size and shape of the perforations, and little else.

Digital photography is much more appropriate with many applications than chemical photography. A great cost and speed advantage is associated with digital cameras as opposed to film cameras, particularly if the images are ultimately used for websites, emails, or small prints. In these cases, there is absolutely no reason to go through the time-consuming (and relatively expensive) processes of chemical development, printing, and the scanning of the negative, when you ultimately end up with a picture that is marginally different from a digital equivalent. But the motion picture industry is highly specialized, and it requires a massive volume of imagery be recorded over a short, intense time frame. The motion picture process benefits greatly from having the luxury of a post-production phase, so it’s here that the digital advantages come into effect. Finally, as digital cameras get better and cheaper, so will film scanners; thus it may become feasible to shoot on larger formats than 35mm, resulting in a corresponding increase in overall quality. So if you shoot on film, you’ll still be able to take advantage of the improvements in technology for many years to come.

In any case, the digital film process eventually will probably stop trying to emulate the look of photographic film and instead explore an exciting new range of possibilities now available, in exactly the same way that some digital photographers are doing with still photography. When that happens, lots of people will just prefer the methodology, techniques, and output that photographic film provides. And more than this, lots of people will use both.

Digital dailies aren’t presently a popular option, mainly because the speed of the technology is much slower than a traditional telecine process, and the space requirements for storing a high volume of scanned film (especially when most of it will never be used) are too large to be practical. (There is certainly a positive element to limiting the number of people who have access to production material as well). However, as the technology becomes faster, the need for digital dailies will increase. Digital dailies may even be generated using more automation to produce footage that is displayed by a single piece of equipment, and may also include the means to apply some basic color grading to the footage.

5.7.8 Scanner Gauges

Although 35mm film is by far the most common gauge of film that is scanned, some film scanners accept a variety of other gauges, such as 16mm or 70mm. In addition, many scanners can capture specific regions of each frame, such as the “Academy aperture” region (which has a 1.37 aspect ratio). Capturing film of different gauges normally involves replacing the scanner gate to allow the different size of film to run smoothly through the machinery. It may also be necessary to change scanning parameters, such as the scanned image dimensions, to accommodate differences between formats.

Original Film Negatives

A scanner normally works just as well using a film positive (or print) as a negative. However, the process of making a print from a negative, and then a negative from the print, and so on, results in generation loss for each successive copy. Therefore, it’s always advisable to scan the original negative that was shot on the set. Even when a copy might look better (e.g., when it has been processed and color-timed in a lab), it’s still going to be of a lower quality than the original. Finally, because the original negative can be damaged just as easily during the scanning process as during the duplication procedure, it might be advisable to duplicate the negative after it has been scanned. Ultimately though, the possibility of damaging the negative depends on whether the scanning facility takes appropriate precautions to protect original materials (and fortunately, most of them do).

5.7.9 Anamorphic Film

Anamorphic images have a ratio of 2.35 to 1, which is much wider than the 1.33 to 1 ratio of full-aperture 35mm film. Rather than waste film area, anamorphic images use a special lens to “squeeze” the picture horizontally to make it fit much more comfortably on the film area. When projected, the film is “unsqueezed” to give it the proper dimensions.

images

Figure 5–14   The digital dailies paradigm allows film to be scanned for viewing and simultaneously acquired into the digital intermediate pipeline for later use

In the digital intermediate pipeline, anamorphic images can either be unsqueezed at the time of scanning and then saved as images with a wide aspect ratio, or be saved as narrower, squeezed images (with a wide pixel aspect ratio). The latter method has the advantage of saving disk space (and removing the need to resqueeze the images if out-puttingback to film) but requires that each device used to view the images (e.g., the conforming and color-correction systems) be able to automatically unsqueeze the images for display purposes. However, for images that have been squeezed in this way, the effective horizontal resolution (and by extension, the quality level) is halved compared to the vertical resolution. Superior quality images may be obtained by unsqueezing the images optically at the scanning stage (or by scanning twice the number of pixels horizontally than normal) and saving the images with square pixels.

Handling Film More than Once

One much-touted advantage of the digital intermediate process is that it requires a reel of original negative be handled only once. After development, the negative only has to be loaded onto the film scanner, scanned, and then stored away—it need not be handled again and thus isn’t subject to potential damage. That’s the theory. While this is the case for the majority of film that is scanned, there are a variety of situations that require the film be removed from storage and go through the scanning process again. Data corruption, scanner faults, missing frames, and a myriad of other problems can occur and necessitate the film being rescanned. Although there is not usually an impact on the image quality if a reel is rescanned (particularly with pin-registered scanners, and those that allow the calibration settings to be saved), it is worth recognizing that even after a reel has been scanned, it may be needed again.

5.7.10 Pin Registration

Pin registration ensures accurate positioning of each frame of film. Essentially, precisely positioned pins or “teeth” position film using the perforations along its edge, in the same way that a film is loaded into a camera. Using pin registration is necessary for two reasons. First, it reduces the possibility of the images “weaving” (i.e., moving up and down or from side to side in the frame) when played back in a sequence.17 Second, if any frames have to be rescanned at a later date, they’re more likely to match the originals exactly in terms of positioning and can therefore transparently replace the original scans. These effects become more pronounced at higher image resolutions.

The disadvantage of using pin registration is that it somewhat increases the scanning time because pin-registered scanners have to pause between each frame to ensure the frame’s correct positioning, whereas those scanners that aren’t pin-registered can continuously run the film over the scanning laser without stopping.

For nonpin-registered scanners (which includes most telecines), it’s possible to compensate for the lack of pin registration by using a digital process known as “motion stabilization” (or image stabilization), which is covered in Chapter 9. However, having to run this process on every scanned frame may be considered an extravagant waste of computer resources (and probably justifies the additional cost of a pin-registered scanner).

Forward Planning

Because the scanning process can be time consuming, several things can be done concurrently to save time later. For example, many color-correction systems use proxy images (or “proxies”) that are simply downsized copies of the scanned frames. Because it’s (usually) quicker to resample a file than to scan it as each frame is being scanned, the previous scan can be copied and resized. That way, by the time the scan is completed, a set of proxies is also saved alongside it. Beware, this process can actually impair the performance of faster scanners, because destination disks are being accessed more often, which can cause the scanner to stall.

Also, many conforming systems require that the data be assigned specific frame numbers. Normally, the film scanner saves the frames (starting with the punch hole) from 0 and then just increments the count for each frame. After that scanning, all the frames can be renumbered appropriately. To save time (and possibly prevent confusion later), some film scanners enable you to specify the numbering system to be used, which requires no extra time and has no impact on performance.

Finally, when presented with a high volume of film to be scanned all at once, it’s worth working out the order in which the reels will be needed. For instance, complex sequences may have to be conformed sooner if the filmmakers prefer working with particular sections of the film first, to define or experiment with specific “looks” during the color-correction process.

5.7.11 Key Numbers and Timecodes

A useful component of motion picture film is the key numbers. Every few frames of film negative manufactured is pre-exposed with a unique serial number, so that every frame can be independently referenced. These numbers were originally introduced to allow the use of cut lists, which specifically refer to individual or ranges of frames. The cut lists were useful because they enabled film editors to cut together sequences from prints of negatives (making it possible to avoid handling the original negative and subjecting it to possible damage) and then easily match each frame in the cut back to the corresponding negative frame.

More recently, a machine-readable code (or “bar code”) is included alongside each number. Some film scanners incorporate a key number reader that reads the bar code as the film is scanned and then embeds the corresponding number into the image file. This procedure is necessary for productions where the data will be conformed digitally, using a key-numbered cut list (see Chapter 7 for a discussion of conforming projects using key numbers).

Similar to key numbers is the idea of timecodes. Most video sources contain a timecode track that makes it possible to reference a particular frame. In some cases, it’s possible to record timecode information onto film for the same purpose. While timecodes aren’t unique like key numbers (i.e., two reels of film might contain the same timecodes but never the same key numbers), they’re useful for synchronizing film with other devices. For instance, when filming a music promotion, the timecodes might be linked to the soundtrack, so that the audio can be quickly matched to the picture. Sometimes this linking is performed with a “timecode slate” or a clapperboard containing timecodes, which is filmed at the beginning of each take, or alternatively, the timecode information can be embedded on the outside edge of the film in a machine-readable format. When the latter method is used, film scanners equipped with a timecode reader embed the timecode information into the scanned image.

Digital Cameras and Scanners

Digital cameras, camcorders, and image scanners are all analogto-digital devices. They each take an analog source (i.e., with cameras, the source is reflected light from a scene) and convert it to a digital format. However, many of these devices try to “improve” picture quality by applying a set of post-processing filters on the digitized images. In many cases, these filters can be beneficial—for instance, when compensating for CCD defects within the device. All too often however, the device automatically applies digital sharpening and color-correction filters, which degrade the image. When using such a device, be sure to disable color grading or sharpening that can be done later. Some newer digital cameras provide access to the “raw” data of the camera, providing the image information in the form it was received from the CCD array and allowing you to apply the filters. This also has the effect of acquiring images that are at the maximum possible quality level, provided all the software can correctly interpret the file.

However, in some cases (especially on high-end cameras and scanners), devices utilize optical filters that don’t degrade the image quality at all but just alter the way the information is captured prior to digitization (which is kind of like positioning actors where you want them in the first place, rather than positioning them later by editing the images). For example, applying a “one-light” grade to film on certain film scanners provides a quality gain in the captured data, with no loss of information.

IT Doesn’t Look Right!

Acquiring media for use in the digital intermediate pipeline has something of a pitfall, which is that you intuitively expect a properly captured, accurate representation of the source media to look as good as it should. That is, you naturally assume that a scanned reel of film, when viewed on a monitor, should look equally as good as (if not better than) a projected print of the same film. However, this assumption isn’t always correct, especially with scanned film. The aim of the acquisition stage isn’t necessarily to get the best-looking scanned data of a particular reel of film; it’s to ensure that the maximum level of quality is available to be used later down the line. Correctly scanned film, even when perfectly exposed, might look flat and colorless. This is because the colors may have been optimized to accommodate as wide a contrast ratio or dynamic range as possible, so that there are a greater number of possibilities for color grading and the other phases, without having to compromise image quality. It may also be that a lookup table must first be applied to the data so it can be correctly displayed on a monitor. In some pipelines, until this color-correction phase begins, you may not even be able to see the potential output of the images in terms of color. Some facilities, on the other hand, such as those offering digital dailies services, may be able to generate two types of images: one for viewing purposes and one for processing purposes (although both images would come from the same source data).

5.8 Summary

Before any material can be used in the digital intermediate environment, it must first be ingested. With digital source material, this ingestion may be as simple as transferring the source data by copying files between devices.

With analog material, such as film and video, analog-to-digital conversion, a largely mathematical process, must be used. Film material may be digitized using a digital film scanner, or a suitably equipped telecine, while video sources require a video capture device. In general, analog-to-digital conversion should be performed at the highest quality possible, working from original material whenever feasible, capturing as much useful information as practicable, to benefit later stages in the pipeline.

In theory, data transfer can occur at much higher speeds than video or film, but in practice, data transfer can be a very slow process. For this reason, a balance must be achieved between image quality and file-size considerations, which in turn impact the pipeline’s performance and relative cost.

The image quality of acquired footage depends on a number of factors, such as the color space, the precision, and the resolution, and there are methods for measuring each factor throughout the pipeline. The digital acquisition process typically generates a large volume of data, and the organization of this data is the subject of the next chapter

1 Celluloid is particularly susceptible to static electricity, which effectively attracts tiny particles to stick to the film.

2 Video tapes may also lose other information, such as timecodes, which are encoded directly on the tape. However, many video tape machines have mechanisms to reduce or prevent the likeliness of such loss occurring.

3 In the case of digital cameras, the resolving power can be calculated from the output resolution only if the optical components (such as the lens and CCD array) themselves have at least the same resolving power as the recorded resolution. Unfortunately, in many cases the resolving power of the optical components isn’t the same as the recorded resolution, and the resolution is “cheated” by “up-sampling” the captured image.

4 This is due to a concentration of red- and green-sensitive cells at the center of the retina, known as the macula lutea, or “yellow spot.”

5 Although some sources are listed as being digital, they may commonly be transferred by analog means. For instance, high-definition video is an inherently digital format, and yet it’s not unusual for HD tape decks to be wired into a video-switching matrix (a device that allows video to be routed to many different places at the same time) via analog means. Many facilities are actually set up to acquire HD video via analog means, rather than transferring the raw digital data.

6 Although even CG imagery may contain some elements from analog sources.

7 Even some analog devices use sampling techniques, such as telephones and faxes, which sample an analog source and then re-encode it as an analog electrical signal.

8 Actually it’s more like an analog-to-digital process, if you think of it in terms of the huge (but finite) number of photon and grain interactions on a piece of film, but thinking about it like that can be unnecessarily confusing and misleading.

9 This term is a little misleading because the process does not actually go back to the original source and sample it again. It merely uses the information already available.

10 The Adobe 98 RGB color space is similar to the NTSC color space, while the sRGB color space is similar to the PAL color space.

11 In most cases, it’s advisable to previously clean the film with a dedicated film cleaner, therefore ensuring that the reel is as free from defects as possible.

12 Not including the time to load the film and set up the scanner’s parameters.

13 Another interesting feature of the Northlight scanner is that the film is scanned horizontally rather than vertically. Combined with a slightly downward air pressure, this feature supposedly reduces the amount of dust on the film during the scanning process.

14 SMPTE is the abbreviation for the Society of Motion Picture and Television Engineers.

15 This result will vary, depending on the type of film stock. Also, different film stocks have different resolving powers, depending upon factors such as exposure.

16 A quicker way to get this result is to simply double the resolving power.

17 Unless, of course, the film was weaving during shooting.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset