Jump to content

User:Sonny Day

From Wikipedia, the free encyclopedia

User:Sonny Day



Fibre Optics
The ancient romans knew how to heat and draw-out glass into fibers of such small diameter that they became flexible. They also observed that light falling on one end was transmitted to the the other. ( We now know this is due to multiple reflections at the internal surface of the fiber.) These multiple reflections in a sense mix the light beams together thereby preventing an image being transmitted by a single fiber - (more accurately, the different path-lengths experienced by individual light-rays alter their relative phases so rendering the beam incoherent and thus unable to reconstitute the image.) The end result is that the light emerging from a single fiber will be some kind of average of the intensity and colour of the light falling on the 'front' end.
Coherent Fibre Optics
If a bundle of fibers could be arranged such that the ends of the fibers were in matching locations at either end, then focusing an image on one end of the bundle would produce a pixelated version at the further end which could be viewed via an eyepiece or captured by a camera. However such a bundle would need to contain tens of thousands of fibers to produce a useful image. In the early 1950s, Hopkins realized a way to accomplish this. He proposed winding a single continuous length of fiber around a drum. Then, when sufficient turns had been added, the coil could be clamped, cut through and straightened out to produce the required coherent bundle. It only remained to calculate and fit the necessary lenses at either end and then to enclose the whole in a protective flexible jacket for the 'fibroscope' (now more commonly called a fiberscope) to be born.
Fibroscopes
Fibroscopes have proved extremely useful both medically and industrially. Subsequent research and development, not least by some of Hopkins' own research students led to further improvements in image quality. Other innovations included using additional fibres to channel light to the objective end from a powerful external source - thereby achieving the high level of full spectrum illumination that was needed for detailed viewing and high quality colour photography. At the same time this allowed the fibroscope to remain cool, which was especially important in medical applications. (The previous use of a small filament lamp on the tip of the endoscope had left the choice of either viewing in a very dim red light or increasing the light output at the risk of burning the inside of the patient.) Alongside the advances to the optical side, came the ability to 'steer' the tip via controls in the endoscopists hands and innovations in remotely operated surgical instruments contained within the body of the endoscope itself. It was the beginning of key-hole surgery as we know it today.

Rod-lens Endoscopes
There are, however, physical limits to the image quality of a fibroscope. In modern terminology, a bundle of say 50,000 fibres gives effectively only a 50,000 pixel image - in addition to which, the continued flexing in use, breaks fibres and so progressively loses pixels. Eventually so many are lost that the whole bundle must be replaced (at considerable expense). Hopkins realised that any further optical improvement would require a different approach. Previous rigid endoscopes suffered from very low light transmittance and extremely poor image quality. The surgical requirement of passing surgical tools as well as the illumination system actually within the endoscope's tube - which itself is limited in dimensions by the human body - left very little room for the imaging optics. The tiny lenses of a conventional system required supporting rings that would obscure the bulk of the lens' area; they were incredibly hard to manufacture and assemble and optically nearly useless. The elegant solution that Hopkins produced (in the 1960s) was to use glass rods to fill the air-spaces between the 'little lenses', which could then be dispensed with altogether. These rods fitted exactly the endoscope's tube - making them self-aligning and requiring of no other support. They were much easier to handle and utilized the maximum possible diameter available. With the appropriate curvature and coatings to the rod ends and optimal choices of glass-types, all calculated and specified by Hopkins, the image quality was transformed. Light levels were increased by as much as eighty-fold and diameters of a just few millimeters were possible. As with the fibroscopes a bundle of glass-fibers would relay the illumination from a powerful external source (typically a xenon arc lamp). With a high quality 'telescope' of such small diameter, the tools and illumination system could be comfortably housed within an outer tube. Once again, it was Karl Storz who in 1967 manufactured the first of these new endoscopes as part of a long and productive partnership between the two men. Whilst there are regions of the body that will forever require flexible endoscopes (principally the gastrointestinal tract), the rigid rod-lens endoscopes have such exceptional performance that they are to this day the instrument of choice and in reality have been the enabling factor in modern key-hole surgery.




1. In theory - as a more or less analogue system - the human eye can distinguish  an infinite number of shades within the overall range of perceivable colours. This 'overall range of perceivable colours' is effectively the 'colour-space' of the human visual system.

In very simplified terms the visible spectrum of light comes as a continuously changing range of 'colours' ( cf. rainbows/prism spectrums). The eye however only has colour receptors/sensors for the three primary colours - red, green and blue (RGB). The eye encodes each colour as a unique combination of output levels from these sensors which it sends to the brain, which in turn perceives that colour. Mixing all three in 'equal proportions' gives white light. The eye/brain cannot distinguish between the 'colour' produced by simply mixing the appropriate proportions of just the three primaries and that due to the whole range of natural light. This is central to the long-established RGB colour model . It explains how TV screens/ computer displays etc. only need to generate three different colours and cameras equally only need to record three different colour. As the intensity of each primary colour can be varied continuously, in theory at least an unlimited number of shades can be produced. See colour vision.


2. A digital recording system records an image as a finite array of pixels each of which has a specified colour. The colour is specified by recording the intensity levels - detected, for instance, by a camera sensor - of each of the three primary colours (RGB). However in a digital system, only a finite number of levels can be recorded (without using infinite data) which means only a finite number of different colours can be recorded.

A typical recording might employ '8-bit colour depth'. Only numbers between 0 and 255 can be written using 8-bit, so only 256 levels can be specified/recorded for each of the three primaries. So RGB [0,0,0] = black. RGB [255,255,255] = white . RGB [255,0,0] = saturated red etc.. 256 levels for each of three primaries gives 256 X 256 X 256 = 16,777,000 combinations/different colours. (often rounded down as 16 million colours). When light falls on a sensor its colour is recorded not as its true colour but as the closest to it of the 16 million possible 8-bit colours. Whilst each primary is represented by an 8-bit number, each pixel in RGB requires three of these numbers and therefore is also confusingly called 24-bit

3. If you distribute these 16 million colours evenly across the whole range of human vision, the difference between each discrete shade would be quite noticeable.

It is essential to have only small differences between shades in order to reproduce subtle details and provide a smooth transition between shades. This would require far more colours, requiring a greater bit depth - in other words much more data. (In fact 16-bit is frequently used in high quality work. This has 65,536 levels per primary giving billions of shades and confusingly with three 16-bit numbers to specify the colour of each pixel is often called 48-bit.)

4. Many of the 16 million colours specified in 8-bit depth are very dark or extreme shades and are more or less absent in the majority of images. Thus if we try and cover the whole range, many of the RGB values go unused whilst we do not have enough for the main colour region. It turns out, however that we can still have acceptable colour rendition for many/most purposes if these shades are simply ignored.

This is fortunate as in reality an RGB system cannot actually reproduce all the visible colours - in fact only those which lie within the range de-lineated by the three chosen primaries.

By distributing the 16 million colours (of 8-bit) across less than the full range of human vision - specifically excluding the very dark and extreme shades referred to above - it is possible to make the difference between adjacent colours smaller - thereby achieving sufficient shade differentiation where it matters and providing acceptable colour rendition without having to increase the amount of data required.

5. As an approximate guide sRGB covers 30% of the visual range whereas aRGB covers about 50%. (Depending somewhat on luminance levels and also whether one allows for certain perceptual factors.)

As both have the same number of colours to distribute,  the difference between individual colours in the sRGB colour-space is less and therefore it gives better shade differentiation within its range. The aRGB colour-space however covers some extra colours - more specifically, extended green range at low luminance levels and intense cyan/green, magenta, orange and yellows at high luminance. The loss of shade differentiation using aRGB is potentially significant at 8-bit but is not at 16-bit with its much greater number of levels and therefore colours to distribute.


6. These ranges are different because the two colour-spaces use different primary colours - those of sRGB being closer together than those of aRGB. An important consequence of this is that the actual colour corresponding to (for example) {100,150,255} in sRGB is not the same colour as {100,150,255} is in aRGB.


7. At its simplest, the colour space of a device refers to the range of colours that it can either record or reproduce.

To a greater or lesser extent, all recording devices, displays and printers have their own color-spaces. That is any given camera can only record and any given screen or printer can only reproduce a certain range of colours. Also, especially for digital systems, they will have their own levels of shade differentiation (dependent on the bit-depth employed)..

8.The colour-space of your recorded image, your monitor and your printer may all be different. In other words a given RGB value will correspond to a different colour in each device!


9. In order to display the original colour, the RGB value sent to the display must be changed to the one corresponding to that colour in the display.

If told which colour-space was employed to create a given recording/file, a display will hopefully have been programmed with a set of rules for converting the RGB values of the recording colour-space to that of the display. If not, then the RGB values must be changed before it is sent to the display - usually expressed "as converting the file to the display's colour-space". Likewise, one will usually have to "convert the file to the printer's color-space".

10. Good quality printers have a colour-space that exceeds aRGB and therefore will print the extra colour range if it exists.


11. Lesser quality printers  and some commercial  printers have a colour-space that is smaller than aRGB and so will not be able to reproduce these extra colours.


12. The majority  of computer screens  have a colour-space more or less the same as sRGB and cannot reproduce the extra possible colours of aRGB.


13. UNLESS an image contains deep dark greens and intense yellow / orange / magenta which are important to the final image   AND  the image is to be printed on a high quality machine which can utilise aRGB then there seems little point in using aRGB and many reasons to prefer sRGB. In any case if ones shoots in Raw and retains the file, then the data required for aRGB will not have been lost.


14. In order to display on a sRGB monitor with any accuracy, an aRGB file must be converted to sRGB (essentially this means re-allocating  the colour of each pixel  in aRGB  colorspace to its nearest equivalent colour in sRGB).


15. Having converted a file from aRGB to sRGB, the wider gamut colour information is lost and cannot be retreived by reconverting to aRGB. Equally the greater shade differentiation of sRGB will be lost on converting to aRGB.


16. Assigning a profile is not the same as converting. 

Assigning merely tells a printer or display to treat the data of one colour-space as though it were from another.  Most commonly, this occurs as assigning an sRGB profile to an aRGB colour-space image, with the result that the colours are incorrectly reproduced (typically washed-out).


17. Commercial photolabs very commonly work in sRGB.  It is not unheard of for commercial printers to assign sRGB to aRGB images rather than converting them - with  resultant colour errors.


18. If your browser is not colour managed (at present only Safari and FF3 are) it will assume that any image file is in sRGB colourspace. Since most computer screens are very close to sRGB in colourspace, then an sRGB colourspace recording (file) will  reproduce reasonably accurately. (Assuming of course that the screen has at least got the  colour, brightness and contrast  controls adjusted correctly, if not a proper calibration.) 


19. If you shoot in aRGB (or in Raw and then convert to aRGB),  your monitor will still show the sRGB version unless you have a wide gamut screen,  so you will not be able to see any difference until you print.  Generally it will an unexpected boost to the image, not damaging.


20. Unless you have a definite reason, you should always use sRGB on the web. Shooting/saving in Raw covers a wider colour-space with better shade differentiation than either sRGB or aRGB. You have to have a good reason not to be using Raw.This is a function of the colour management profile.