Beruflich Dokumente
Kultur Dokumente
www.basler-vc.com
GigE Vision is becoming a more powerful and reliable interface The Basler performance driver is a hardware specific GigE Vision
in the machine vision market. In addition to IEEE 1394a/b network driver and is compatible with network interface cards
(FireWire) and Camera Link, GigE Vision is now found in many that use specific Intel chipsets.
applications, for example, in traffic control, PCB inspection, and
medical care. The GigE Vision standard makes a persuasive Filter and Performance Driver CPU Load
argument especially regarding bandwidth, transmission mechanism,
reliability, and cable length. GigE Vision is based on the topology Comparison
of well-established Gigabit Ethernet technology. Gigabit Ethernet
has the advantage of wide penetration in both the industrial Due to the different architectures of the drivers and to different
and consumer markets and serves as the basis for an easy-to- hardware platforms, the CPU load of the filter and the
use, low-cost interface for a wide variety of applications. performance drivers varies. There are other parameters that
also contribute to this effect including: network vendors, network
topology, packet size of the data transfer stream, bandwidth,
GigE Vision CPU Load and the network class (avoids packet resends).
In the machine vision market, there are still some open questions As described above, the filter driver and the performance driver
regarding GigE Vision. Whenever Gigabit Ethernet is discussed use the CPU with different intensities. The graphs in Figures 1
in connection with machine vision, CPU usage and latency times and 2 reflect typical results for a medium performance PC and
in the PC are among the most interesting and controversial a bandwidth situation comparable to IEEE 1394b. The PC is
topics. Different players in the machine vision industry still must equipped with a Pentium dual core 2.8 GHz processor, an Intel
be convinced that the CPU load generated by the use of a GigE Pro 1000 GT network card, and has the Basler pylon SDK
transmission mechanism is within acceptable limits. The main installed. The measurements were performed with a
difference between GigE and FireWire or Camera Link is that Basler piA640-210gm. Network workloads were between 15
with GigE there is no CPU-independent incoming data and 61 Mbytes/s and the packet size was either 500 or 4000
management. This means that every incoming packet must be bytes.
handled when arriving in the PC and copied afterwards. This
process always requires CPU involvement. To keep the CPU 14
load as low as possible, Basler offers two different GigE Vision Performance Driver Filter Driver
12
drivers: a filter driver and a performance driver.
10
CPU-Load [%]
6
The Basler filter driver quickly separates incoming packets and
4
transfers them directly to the application. The filter driver can
be used with all network interface cards available on the market. 2
The CPU load associated with the filter driver is generally
0
attributable to the fact that packets must normally be copied 15 30 45 61
two times. Bandwith [MB/s]
Figure 1: CPU load with a packet size of 500 Byte
The Basler Performance Driver 14
Performance Driver Filter Driver
The Basler performance driver also separates incoming packets 12
and transfers them directly to the application. The CPU load 10
CPU-Load [%]
driver. 2
0
15 30 45 61
Bandwith [MB/s]
Figure 2: CPU load with a packet size of 4000 Byte
AF000065 02 GigE Vision - CPU Load and Latency
As you can see, the performance driver has a significantly lower For the workload evaluation process, a sample program from
CPU load than the filter driver. In addition, you can see that a the pylon SDK called 'AcquireContinuous.cpp' is used. This
larger packet size has a positive impact on the CPU load. A data application continuously acquires images and stores them in the
stream based on small packets includes more overhead/protocol application's buffer. This program does not perform any additional
data than a data stream with large packets. A typical CPU load user interface activities such as displaying the image or analyzing
imposed by a GigE based application is 5 % the image content.
performance driver and an Intel network card that supports // Grab c_ImagesToGrab times
for ( int n = 0; n < c_ImagesToGrab; n++)
jumbo frames (i.e., a 6 Kbyte packet size). {
// Wait for the grabbed image with a timeout of 3 seconds
if ( StreamGrabber.GetWaitObject().Wait( 3000 ))
{
// Get the grab result from the grabber's result queue
To make the evaluation of measurement data GrabResult Result;
StreamGrabber.RetrieveResult( Result );
more transparent, Basler is providing a Java
if ( Grabbed == Result.Status() )
script that can be used to determine the current {
// Grabbing was successful, process image
CPU load on a machine vision system's host cout << "Image #" << n << " acquired!" << endl;
cout << "Size: " << Result.GetSizeX() << " x "
PC. This script uses the same Windows operating << Result.GetSizeY() << endl;
system functions as the Windows Task Manager, // Get the pointer to the image buffer
const uint8_t *pImageBuffer = (uint8_t *) Result.Buffer();
and the results are comparable. cout << "Gray value of first pixel:
" << (uint32_t) pImageBuffer[0]
<< endl << endl;
// CPU Load. // Reuse the buffer for grabbing the next image
// if ( n < c_ImagesToGrab - c_nBuffers )
// This script samples the systemwide CPU load at regular intervals. StreamGrabber.QueueBuffer( Result.Handle(), NULL );
// }
// Run by using "cscript /nologo cpuload.js" within a console. else if ( Failed == Result.Status() )
// Stop by pressing Ctrl-C. {
// // Error Handling
// Note: Because the Win32_PerfFormattedData_PerfOS_Processor formatted cerr << "No image acquired!" << endl;
// data class is not available on earlier Windows versions, cerr << "Error code : 0x" << hex
// this script only runs on Windows XP and above. << Result.GetErrorCode() << endl;
cerr <<"Error description : "
var wbemFlagReturnImmediately = 0x10; << Result.GetErrorDescription() << endl;
var wbemFlagForwardOnly = 0x20;
// Reuse the buffer for grabbing the next image
var strComputer = "."; if ( n < c_ImagesToGrab - c_nBuffers )
var objWMIService = GetObject("winmgmts:\\\\" + strComputer + "\\root StreamGrabber.QueueBuffer( Result.Handle(), NULL );
\\CIMV2"); }
}
var total = objWMIService.Get("Win32_PerfFormattedData_PerfOS_ else {
Processor.Name='_Total'") // Timeout
cerr << "Timeout occurred!" << endl;
for (;;)
{ // Get the pending buffer back (It is not allowed to deregister
total.Refresh_(); // buffers when they are still queued)
WScript.Echo("\rProcessor Time: StreamGrabber.CancelGrab();
" + total.PercentProcessorTime + "% ");
WScript.Sleep( 1000 ); // Get all buffers back
} for ( GrabResult r; StreamGrabber.RetrieveResult( r ););
// Cancel loop
Java script to determine CPU load }
break;
// Stop acquisition
Camera.AcquisitionStop.Execute();
[]
The receipt of the data stream from the camera by the host Figures 4 and 5 show timelines for the trigger signals and their
PC is identical in both scenarios. The camera transmits the jitter. The measurements were made for GigE and IEEE 1394 to
streaming data across the network layer to the GigE vision driver be able to compare both technologies. The measurements were
(performance or filter driver) via the Pylon API to the image performed with Basler scA640-70fm/gm cameras.
buffer of the application.
As Figure 4 shows, the timeline of the software trigger is very
Figure 3 shows the logical path of the software trigger (green) short in relation to the total timeline, but the variation is
from the application to the camera and the data stream path significant. This variation is caused by the IP stack of the operating
from the camera to the application (blue). system. There is no significant influence from the GigE Vision
driver. The latency time of GigE and FireWire are very similar
and are both in the range of 30%.
To determine the latency times of
scenario one and two, different time
stamps are defined: SW-Trigger:
GigE Vision
Software trigger start
Trigger Start Start of exposure
Exposure start command received by
670 - 950
the camera 0 300 600 900
[s]
Hardware trigger start IEEE1394
Start of exposure
Trigger Start Start of exposure
End of exposure
340 - 480
Images received by the camera 0 300 600 900
[s]
Figure 4: Latency time of software trigger for GigE vs. IEEE 1394
AF000065 02 GigE Vision - CPU Load and Latency
The timeline from the hardware trigger to the start of exposure The jitter of the timeline is very small and is approximately 5%
depends on the camera model and is in the area of 50s (please of the transmission time of the data stream. This is true because
refer to the user's manual for your specific camera model). the architecture of the GigE driver is very efficient.
Evaluations have also shown that the process priority has no
HW-Trigger: significant impact on latency time. The largest fraction of the
processing time is used in the IP stack or in the GigE Vision
driver, which are independent of the process priority.
Trigger Start Start of exposure
31,72
In addition to the latency time of the camera and the operating
0 20 40 60
[s] system, there are other components that can add additional
delays and jitter times. A network switch, for example, can cause
this effect. The extent of the delay depends on the hardware
Figure 5: Exposure start delay for the scA 640-70 vendor. The hardware vendor's data sheet can provide more
detailed information.
The timeline for the data stream from the camera to the image
buffer of the application depends on several different parameters To create a machine vision application with near to realtime
including: performance of the host PC, image size, packet size, conditions, we suggest using hardware triggered image acquisition
bandwidth of the network, etc. and a network topology that does not include a network switch.
IEEE1394
Start of End of
transmission transmission
0 5 10 13,5 - 13,7
[ms]
Figure 6: Latency time of image transmission for GigE vs. IEEE 1394