Israel’s Mobileye N.V., a global designer and developer of camera-based Advanced Driver Assistance Systems, introduced its 4th generation system-on-chip, the EyeQ4 which the company boasts will bring the world closer to fully automated cars.
The EyeQ4 consists of 14 computing cores out of which 10 are specialized vector accelerators with extremely high utilization for visual processing and understanding. The first design win for EyeQ4 in series production has been secured for a global premium European car manufacturer for production to start in early 2018.
The EyeQ4 would be part of a scalable camera system starting from monocular processing for collision avoidance applications, in compliance with EU NCAP, US NHSTA and other regulatory requirements, up to trifocal camera configuration supporting high-end customer functions including semi-autonomous driving.
“Supporting a camera centric approach for autonomous driving is essential as the camera provides the richest source of information at the lowest cost package. To reach affordable high-end functionality for autonomous driving requires a computing infrastructure capable of processing many cameras simultaneously while extracting from each camera high-level meaning such as location of multiple types of objects, lanes and drivable path information, ” said Prof. Amnon Shashua, cofounder, CTO and Chairman of Mobileye.
“The EyeQ4 continues a legacy that began in 2004 with EyeQ1 where we leveraged deep understanding of computer vision processing to come up with highly optimized architectures to support extremely intensive computations at automotive compliant power consumption of 2-3 Watts.”
In addition to the EyeQ4 “high” capability version, at an ASP of approximately three times that of an ADAS functionality chip, Mobileye also plans the launch of the EyeQ4 “medium” variant within the same timeframe. The EyeQ4M will include a subset of EyeQ4’s computational cores, enabling a select group of functions.
Engineering samples of EyeQ4 are expected to be available by the fourth quarter of 2015. First test hardware with the full suite of applications including active safety suite of customer functions, environmental modeling (for each of the 8 cameras), path planning for hands-free driving and fusion with sensors, is expected to be available in the second quarter of 2016. Series production is supported for early 2018 start of production.