WP's SloMo


Deutsch

Image memory of digital high-speed camera systems

Where to save all the image data?

VGA, PAL/NTSC and HDTV formats
VGA, traditional and new TV formats

Emerging data: already at VGA resolution with 640 x 480 pixels and a frame rate of humble 100 frames/sec more than 30 Megabbyte data are sampled per second. But that's just to handle with standard interfaces and mass storage media, with USB 3.0 even 800 frames/sec. For megapixel resolutions and about 300 frames/sec therefore the frame grabber / control unit approach with the CoaXPress interface. A concept that is never outdated, because the problem getting rid of the huge data amounts is always present.

 

Δ

Image recording and data transmission

Imaging techniques

Existing digital high-speed cameras are not limited by the resolution of their sensors only, but also by the read-out rate of their sensors and the data transmission rate to their storage media and by the capacity of those frame buffers. The maximum read-out and data transmission rates restrict, according to the type of design, the recording frequency and/or the usable area of the image sensor whereas the storage capacity of the buffer memory is responsible for the comparatively short duration of the sequences. The buffer memory can be located inside the camera head or on a plug-in board of the control unit.

Name

Transfer rate nominal max. cable length
Fast Ethernet
(100 Base T)
100 Mbit/s 12.5 Mbyte/s 100 m
Gigabit Ethernet
(1000 Base T)
1000 Mbit/s 125 Mbyte/s 100 m
FireWire 400
(IEEE 1394a)
~400 Mbit/s 40 Mbyte/s 4.5m (14 m)
FireWire 800
(IEEE 1394b)
~800 Mbit/s 88 Mbyte/s ... (72 m)
USB (1.1
Full Speed)
12 Mbit/s 1.5 Mbyte/s 3 m
USB 2.0
High Speed
480 Mbit/s 60 Mbyte/s 5 m
USB 3.0
Super Speed
5000 Mbit/s 625 MBbyte/s 3 m
USB 3.1 10 Gbit/s 1212 Mbyte/s 1 m
USB 3.2 20 Gbit/s 2.5 GBytr/s 0.5 m
USB 4.0 / Thunderbolt 4 20 / 40 Gbit/s 2.5 / 5 Gbyte/s 0.5 m

Comparison of some PC interfaces (1 bit = 1/8 byte)

Of course, there are high-speed camera systems, whose resolution and recording frequency just permit continuous operation like a video recorder over a longer span of time. They store their data on magnetic tape (older models) inside the control unit or on the disk of a PC directly - streaming.

The integration of the camera(s) is a multilayered task. Primary it deals with the data transfer - image data back and control data to (and possibly back). Just expressed in sober numbers: a megapixel sensor, even with a comparatively humble 8 bit color depth per channel, generates at 1 000 frames/sec 1 Gigabyte after all, so 1 000 Megabyte, data per second in RAW format. First this amount of data has to be transferred and then has to be stored somewhere. Even the Gigabit Ethernet (= 1 000 Megabit/sec), the favorite at the moment, offers a nominal transfer rate of 125 Megabyte/sec (1 bit = 1/8 byte) only. In real about 100 Megabyte/sec remain due to a certain administrative overhead. Similar relations are valid for USB 3.0 with about 500 Megabyte/sec usable transfer rate.Intel Thunderbolt 3 or 4, USB 3.2 or 4, resp., with 20 Gigabit/sec at least would be fine, if there were no cable length limitations. Although one can quit real time requirements, but that is not possible with control data. What generates demands for an optional operating system.
Furthermore the cameras require (a good amount of) power, what is to notice for cable length, especially for one-cable solutions.
Finally there are security reasons to add. Often one works in research and development environment and not every company wants to know its »test mule« transferred thru radio or WLAN.
That makes storage concepts inside the camera head interesting further on and affordable solid state disks without moving parts complement conventional harddisk drives, if one will not rely on fast but expensive DRAMs.

 

Binning and format adaption

Engineering tricks to increase recording frequency despite of limited read-out rate are depending of the design of the sensor. CCD sensors rather offer a line or column reduction, e.g. in form of binning, which means two neighboring pixel rows are combined and are read out as one single row. Concerning CMOS sensors, which are designed similar to DRAMs with more or less pixel by pixel access, it is rather the reduction of read-out format simply by reducing the read-out area of the sensor, i.e. a smaller pixel count.

line and column binning
Line/column binning

Comparison of SpeedCam formats
Comparison of SpeedCam read-out areas

Thus for instance SpeedCam +500/+2000/lite (Dalsa CA-D1-0128A, CA-D1-0256A CCD) use line binning, SpeedCam 512/PRO (EG&G Reticon HS0512JAQ CCD) use column binning and SpeedCam Visario (FhG/CSEM Cam 2000 Visario CMOS) uses format adjustment. For comparison of (maximum) resolutions, see the figure on the right.

Whereas binning keeps the field of view constant, and the saving of pixels to be read out is done by reduction of resolution - lines or columns are combined - with format adaption one saves pixels by creating a pinhole effect. Binning makes images look more blurred, format adaption makes them smaller, but with constant quality. In the end it comes to the same thing: if one wants the same field of view, one will have to edge closer or spread the image - the image turns pixeled. Binning does this without changing the location or focus.

One cannot generally say what is better. During adjustment jobs one will appreciate to avoid changes in lens and camera settings (especially concerning exposure), when increasing the frame rate. On the other hand, however, there will be advantages, when all the images show the same resolution. In real format adaption is preferred: quality counts. Of course, this process is also supported by the increasing use of CMOS sensors, which are predestinated for this by design.

 

Read out methods (interlaced, non-interlaced, progressive scan)

Existing read out procedures for sensor are somehow similar to displaying on screens. In the easiest case half frames like in the interlaced mode of the CRT TV are used. The human eye is to slow to recognize this trick. Not until the still frame replay one perceives this deceit due to the common comb style artifacts, see also [SloMo Freq.].

  • 2:1 interlaced operation in frame integration mode: the first half frame consists of all odd lines the second half frame of all even ones. Both half frames are read out one after the other, but are displayed simultaneously. Disadvantage: one frame in real consists of two images with a time lap between them, which are mixed line by line. Lines 1, 3, 5 ... e.g. are from the present moment, whereas lines 2, 4, 6 ... yet show the image before. In the next step the even lines receive a new image, whereas the odd ones yet show the old one, and so on. Therefore this typical so called comb artifacts in animated scenes, the full frame resolution, however, is preserved.
    (In case of deinterlacing for video replay comparable with »weave«.)

  • 2:1 interlaced operation in field integration mode: two lines are already summed up (binning) during the read out and are displayed, resp. So in the first half frame line 1 and line 2 give line 1, lines 3 and 4 line 2 and so on. In the second half frame line 2 and line 3 give line 2, line 4 and line 5 give line 4 and so on. Therefore the resolution is reduced, possibly there are block or stairs artifacts at edges. But the resolution in time is better, and so any motion blur is reduced.
    (In case of deinterlacing for video replay rough comparable with »field averaging«.)

  • Non-interlaced: half frames are made, which if necessary, are complemented to images with full resolution by doubling the lines. Often semi-professional (VHS/SVHS) video recorders can (could ;-) operate in this mode. Resolution may be lost, but the relation in time is clear.
    (In case of deinterlacing for video replay rough comparable with »bob« (interpolation of missing lines in each half frame) or »skip field« (every second half frame is displayed only, its missing lines are interpolated), resp.)

  • Progressive scan: the sensor is read out completely at (as far as possible) one certain time. Highest resolution in image and time. Technical sophisticated, but most productive method for high-speed imaging and motion tracking.
    (There are no full images, so-called »frames« in traditional tv/video technique. Only half images, so-called »fields« are used.)

For high-speed cameras progressive scan is the most suitable approach. In the standard image processing sector (»machine vision«), however, the interlaced methods can provide reduced data rates and due to the double exposure and maybe a optimized fill-factor of the sensor they can increase light sensitivity.

Δ

Image memory

 

Buffer memory

Either the image data are gathered analogue or digital due to the sensor design. Usually analogue image data are digitized before storing, but there are exceptions. Normally the storage takes place in some kind of buffer. Its location varies with the state of the art and the demands for the camera and for the camera system, resp.
Even the - of course, very limited - buffering of image data on the sensor itself is possible. Especially in very fast cameras.
If the data rate is small enough the image data can be directly streamed to a storage medium (e.g. the harddisk of a notebook). Therefore concerning real high-speed cameras worth of their name this is not an option.

High-speed camera concepts in comparison of size
Digital high-speed camera concepts

Legend to the figure on the left: RAM = image memory; µC = microcontroller or processor; A/D = analogue to digital conversion (often already integrated in the sensor)
The green card should show a PC card, the red line connection possibilities (image and control data).
The standard connection can be e.g. (Gigabit) Ethernet or FireWire.

And so a crash-proofed digital high-speed camera interior may look, see [SloMo  HYCAM].

Cameras without analogue/digital converter on the sensor chip or inside the camera head transmit the analogue image data to the control host and only now they are converted into digital values and stored. That permits to keep the real camera head small and its power consumption (waste heat!) is modest. The demands for the cables, however, are comparatively high due to the analogue data transmission. On the other hand digital cameras already convert the data on the sensor chip or inside the camera head and store them there preferably.
Because of the limited capacity the buffer memory is all the time overwritten in some kind of endless loop. The trigger impulse controls this process and the images are done - cut!

This buffer memory, which usually holds rather raw image data without color recovery algorithms (therefore the expression RAW format), is mostly built with DRAM (dynamic random access memory) integrated circuits similar to those known from the memory banks of PC main memory. With the typical feature of DRAMs to loose their data during a power failure.
Memory ICs, which keep their data even without power supply, so called NVRAMs, SRAMs, Flash (non volatile; Static RAM), are hardly used for buffer memory due to different disadvantages like slower speed, higher costs, higher power consumption, shorter durability, ...
One makes do with a safety (rechargeable) battery, which supplies the buffer memory if necessary, or even with a full service (rechargeable) battery holding the complete camera in working order. Sometimes with an UPS for the complete camera system, especially for systems, which buffer their data at first in the control unit.

SpeedCam Visario image memory
Image memory, resolution levels, frequency and recording time
of SpeedCam Visario high-speed cameras

Explanation of the diagram on the right side:
The left axis defines the maximal possible recording time at given frame rate. The right axis shows the maximal possible x pixel × y pixel resolution in relation to the frame rate. These reduction levels are drawn for better comprehension.
The jumps in recording time derive from the reduction steps of the resolution with increasing frame rate. If one voluntarily accepts the reduction level through lower frame rates, one will be able to drastically increase the recording time in parts. Then one moves along the stroke-dotted lines. Perhaps in this case the recording time may be limited by the minimum frame rate of the system. It is about 50 frames/sec with SpeedCam +500 and SpeedCam PRO, and reaches about 10 frames/sec with SpeedCam Visario.
(The markers in the curves are given to identify the curves in a black-and-white print only. In real the frequencies can be selected without steps.)

 

Permanent memory

Equality is between the buffer memory and a role of film of traditional movie cameras. Its limited capacity forces to shift the data to a mass storage medium. Here one often uses the hard disk of the control unit and its optical disk or removable drive and the network and the cloud, of course.

Some high-speed cameras own a harddisk or a flash card inside her head. Concerning applications with high mechanic loads (e.g. usage in a crash vehicle), however, at least the harddisk, even if it is automatically parked during the trial, causes a rest of risk. Even if several models are specified for the loads in a crash test - when parked mind you.

These mass storage media in the camera head, however, can accelerate the work in a considerable manner. One takes one sequence after another in a short period, shifts the data to the mass storage medium and during a break or over night one downloads the data or just exchanges the storage medium.

Δ

Display and storage

Every manufacturer has its own philosophy presenting the image data more or less revised. For instance with automatic contrast or edge enhancement. Like in photography field professionals prefer the access on RAW images. They are not falsified and very efficient. For instance uncompressed AVI files are by factor 3.5 to 4 bigger than RAW files.

At first one sees the (potential) images through various preview or view finder channels, which may be processed by DSPs (digital signal processor) in real time or be compressed by them. Often simple sharpen, edge enhancement and color saturation filters are used. Not to mention the defective pixel correction, i.e. the interpolation of defect pixels by their neighbors.

For massive image processing in real time software is hardly to use. Usually it operates with the data saved on the mass storage medium and converts them to common file formats. Due to the calculation time this may be even over night.

In spot checks of serial production of safety relevant devices the image data are stored on durable media and archived. Concerning air-bags, e.g. for ten years under the scope of product liability law and additional three years to cover the juridical objection period. In all thirteen years. A lot of material will gather then.