Sie sind auf Seite 1von 4

Technical Insights - Embedded

Remote HD and 3D image processing challenges in Embedded Systems


The Modern Applications has increased the complexity and demands of the video processing features and subsequent image data transfer. The Image Processing mission will typically comprise three key elements: data capture, data processing, and data transmission. In addition, applications such as Recognition and data fusion have a need for time sensitivity, spatial awareness, and mutual awareness to correctly understand and utilize the data. Low-latency processing and transmission are key performance metrics, particularly where there is a human operator and key decision maker situated in a location remote from the point of data gathering. An examination of key considerations sensor processing location trends, video fusion, and video compression and bandwidth, in addition to Size, Weight, and Power. Technological Challenges: In contrast to high power-consuming processors for PCs and workstations, embedded processors must operate on low power. However, conventional embedded processors without efficient parallel processing could not execute massive image processing and high speed data transfer with low power consumption. Another drawback was that in order to achieve high processor performance with low power consumption. Trends: Stay close to the sensor: As high-definition sensors become commonplace on modern Devices, their increased bandwidth demands place significant processing overhead on traditional video processing systems. This trend leads to adoption of video architectures that can process the higher pixel densities, frame rates, and multiple video feeds with minimal latency. Higher definition in turn generates an exponential increase in the volume of data with associated impact on the bandwidth requirements of the transmission Medium and the need to manage this through effective capture, conversion, and compression of video streams.

FossilShale Embedded Technologies Private Limited - www.fossilshale.com

Page 1 of 4

Technical Insights - Embedded

Processing the video locally at the sensor can be beneficial in that it is possible to extract pertinent information, such as target metrics, from the original high-fidelity imagery prior to any downscaling or compression losses, and this can be done with low latency (typically less than 1 frame) for driving subsequent decision making processes. However, this comes at the potential cost of greater power consumption in the Embedded Device, which can impact Battery driven device operation time range and endurance, particularly on smaller platforms with extremely tight power constraints.

Image fusion from multiple camera Inputs and 3D Image processing: The number of sensors on Embedded Device is increasing rapidly, leading to a requirement for intelligent ways to present information to the End user without information overload, while reducing the power consumption, weight, and size of systems. Embedded imaging systems can include sensors sensitive to multiple wavebands including color visible, intensified visible, near infrared, thermal infrared, and based on the application requirements. Typically embedded systems have a single display that is only capable of showing data from one or two camera at a time, so the End user must choose which image to concentrate on, or must cycle through the different sensor outputs. Image fusion is a technique that allows combining complementary information from each sensor into a single, superior image that can be displayed in case of 3D imaging or Industrial Image processing application. Approaches to fusion include the simple additive image fusion approach, which applies a weighting to each of the input images and then linearly combines them. This has the benefits of low latency and moderate processing power, but with variable quality output, and it cannot guarantee retaining the full image contrast. In most cases, a linearly weighted fusion algorithm will produce a perfectly acceptable image that is clearly a combination of the input images and is preferable to viewing the two camera outputs side-by-side. However, in some cases the weighted average technique will result in the loss of key scene features and the fused image might not offer an enhanced view of the scene.
FossilShale Embedded Technologies Private Limited - www.fossilshale.com Page 2 of 4

Technical Insights - Embedded

3D Image processing: More advanced techniques must be employed if a higher-quality image fusion system is required. The most reliable and successful approach to fusion of two sensors uses a multi resolution technique to analyze the input images to maximize scene detail and contrast in the fused image. The added complexity of the multi resolution approach introduces an additional processing load over the linear combination technique but offers much greater scope for tailoring the algorithm to requirements and a higher-quality, more reliable fused image. A key component of successful image fusion is input alignment to ensure that pixels in the different source images correspond to the same time and location in the real world Pixel Synchronization. If this were not the case, a feature in the real world could be represented twice in the fused image, creating a confusing representation of the scene. An ideal image fusion system would contain synchronized sensors and a common optical path, but this is often not possible unless the image sensor has the flexibility and option to synchronize two cameras. Some Aptina Sensors such as MT9V024 image sensor has this synchronization feature. Temporal alignment can be provided by buffering one image stream, which can compensate for unsynchronized imagers or sensors with different latencies. The process of matching one image to the other is achieved by creating a warped image. The pixel intensities of the warped image are evaluated by first calculating where in the original image the output pixel comes from and then interpolating its intensity using the values of the surrounding pixels. This can compensate for any relative rotation of the sensors, misalignment of the optical axes, or differences in the scale of the images. Built-in warp engines can provide rotation, scale, and translation for each video source to compensate for image distortion and misalignment between the imagers, reducing the need for accurate matching of imagers with a resulting reduction in overall system cost. The systems only require a single monitor, further reducing SWaP requirements. With the goal of fusion being to increase dynamic range and offer increased depth of field, sensor data and synthetic video (for example, Industrial Imaging) are now being fused to provide for enhanced local situational awareness for challenging environments. Video compression: Communications bandwidth is always at a premium between a remotely located Embedded Systems and the Remote Monitoring System (Figure 4). Transmitting raw captured video, for example, is at best likely to result in unacceptable delays. As such, the onboard system will typically be required to undertake significant local processing in order to identify valuable information, and discard that which has no value prior to transmission. As remotely located Embedded Systems become increasingly autonomous no longer Wired by Remote Monitoring System, and capable of capturing the real-time data such IP Camera Monitoring will assume even greater significance.

FossilShale Embedded Technologies Private Limited - www.fossilshale.com

Page 3 of 4

Technical Insights - Embedded

Beyond this, to reduce bandwidth consumption even further, the requirement exists to compress transmitted video, using codecs that minimize the data stream while maximizing image fidelity. The codec of choice today is H.264, also known as H.264/MPEG4, which uses around half the bit rate (or less) of previous video codecs. H.264 is popular not only because of its efficiency, but also because its applications are widespread including in broadcast television. The implication of this is that there is a substantial infrastructure of support and expertise that can make implementing an H.264-based system both quicker and less costly for the low budget embedded systems as well, when compared to alternatives such as JPEG2000. Effective video compression becomes even more critical considering that new systems are adding more video sources, and increasing the image resolution from lower quality to high definition at increased frame rates. The result is up to 12x more raw data per video stream, which requires significant data compression to enable the remote monitoring system to view even one video source at Monitoring center. The SWaP issue: Increasingly complex pixel processing chains (tracking and stabilization compression, for example) combined with a rise in the number of sensors used on embedded device has led to a tenfold increase in the number of pixels being processed. Meeting skyrocketing video processing demands for embedded systems while satisfying continually declining SWaP expectations is a daunting task. Combining processes such as Human Identification, Motion detection, image stabilization, image processing, and compression on a single board not only saves space but also results in a more tightly integrated system with lower overall power levels. The rationalization of disparate processing tasks into a single unified processing platform is one approach that is useful in tackling this problem, and SWaP-optimized based image processing modules are already available for MIL embedded applications that are designed to be placed on small platforms can be expected shortly on all common embedded platforms Source - Chris Jobling
FossilShale Embedded Technologies Private Limited - www.fossilshale.com Page 4 of 4

Das könnte Ihnen auch gefallen