Good and bad practices (II)
1st step : normalization
We explained in a few words in another page dealing with CCD how to preprocess raw digital images to substract dark frames and other bias to get frames ready for the post processing. Now, before stacking these images they need to be normalized so that they cover an uniform range of brightness.
The tools are usually automatic algorithms that ensure next functions :
- Display a histogram of the image light curve showing the "brightness" level of each pixel
- Process to stretching in order to extend the histogram over the full range of data.
In this way the dynamic range of the image will be corrected. The darkest level set by default to 0 will be set to the minimum available in the image (e.g. 8.7686 in the below image), and the brightnest level by default set to 255 will be assigned to the maximum available (e.g. 248.44). This process can also be performed by automated procedures like Auto contrast or Auto color in PhotoShop and in adjusting brightness or color curves.
A second issue is the color balance of the image. In theory, the brightest stars must be pure white and the background sky pure black, without red, green or blue dominance. Here also the correction can be set automatically in most imaging programs. In case of algorithm failure, the previous steps must be applied manually, shifting the brightness of each channel. But most of the time the aesthetic effect is improved when the background sky is deep blue or even black. Now our images are prepared and ready for registration and stacking.
2d step : registration
The next step is to register all pre-processed images in order to combine their data pixel by pixel without the slightest shift thanks to an alignment tool using two or several points in each image.
The pictures alignment is sometimes a hard-to-match task but a mandatory one if you want to get rigorous results. You will agree to recognize that it will be silly to mix two inverted images of Saturn or images of a galaxy where stars are misaligned...
The principle consists for each image of removing misalignments such as telescope drift, rotation and higher orders differences (including scanner nonlinearity and film unstability using analog peripherals) by positioning peculiar alignment points, typically stars.
This method usually aligns a serie of points in each images with a minimum of 2 points as far apart as possible for a better alignment accuracy. For the ease of the process images are treated two by two and the result image combined together, etc.
To read: Align and Stack Multiple Moon Pictures with Ninox and Registax, Wikihow
MaxIm DL Video Tutorials, Cyanogen Imaging
Download Registax, by Cor Berrevoets
Once the points are in place in each image the program can automatically refine the pixels position. Then using a bilinear interpolation, it creates automatically the composite by shifting the second image on the original. In some cases this alignment algorithm is able to coincide non-square pixels, different camera orientations or different optics with an accuracy down to a fraction of a pixel (the tricks is to zoom the normal view to enlarge the pixel on screen).
Like rotating a single image, an inconvenient of this registration is to soften your resulting image. This side effect can be suppressed by applying a sharpen function with a minimum of strength to restore the original resolution.
3d step : stacking
The basic concept consists of stacking or combining images together in respect to the exposure time of each image to simulate a longer exposition. So if you add 4 images, each of them will receive 25% of the weight of the total exposure time.
This step is simple and only the processing tool depends on the nature of your images, monochrom or RGB. Indeed, this last often uses a specific function to insure the stacking of the three monochrom images taken through red, green and blue filters, knowing the blue one requested the longer exposure.
For B/W images the method consists of using a non alignment points and a blending, adding images according to their time exposure.
Another way to combine images is to use a short exposure image to restore highlights in long exposures. The principle consists of creating a mask that isolates the highlights of the long exposure. This is the famous unsharp mask technique explained in next page, particularly impressing when picturing bright nebulae or the Sun corona during an total eclipse.
If a B/W stacking is quite easy to make in stacking many black-and-white frames, an LRGB composite is more subtil as explained in the next link dedicated to CCD's.
To read: The image processing
To increase the contrast of a RGB image, we can take profit of a slight different technique, combining a monochrom picture with our RGB's to produce what we call an LRGB image.
When amateurs speak of LRGB images, theoretically they speak of a processing more sophisticated than simply stacking four monochrom frames, 1 monochrom + 1R + 1G + 1B. In short and theoretically, to increase the signal-to-noise ratio, reduce the turbulence and others artifacts, RGB images should be the stacking of several dozen to hundreds of monochrom images as well.
The number is not very important, and in some cases even 1R+1G+1B added to the luminance are enough when seeing is excellent. But usually, as the seeing is far to be exceptional, most amateurs prefer to stack several LRGB to reduce the electronic noise and effects of the turbulence. So it is common to recording tens or hundreds and sometimes thousands of individual R, G, B frames.
Then these R, G and B images are stacked together with the luminance image. This last gives the contrast to the RGB picture, whithout which the resulting image looks fine of course but lack of depth, of contrast; it is not "crisper of details".
The luminance frame is the combination of one to some dozen to hundreds individual monochrom frames. This is particulary important when picturing highly featured surfaces like planets (Mars, Jupiter, Saturn and in a lesser extent the crescent of Venus or the one of Mercury).
The final image results then in the combination of all these individual frames. As you can see browsing the image gallery, results can be amazing.
At last you can merge several images to circle a vast area of the sky, larger that the field of your scope and create a mosaic. This method requests to extend the input image to enlarge its limits and create a linear gradient mask that runs from black and white across the overlap between the two images. Then you can create a blending mask for the overlay and composite all together, the extended input image, the overlay and the mask.
The basics being defined, in next pages we will see several techniques to improve the aesthetic of astronomical images and to correct their defaults. Unsharp mask, LRGB or mosaic are all terms that apply to image processing techniques that combine multiples photographs to enhance its features. But to get results as exceptional as the ones displayed above or in the gallery, some strict rules must be applied.
To read : Composites by Examples
Page 1 - 2 -