Unmanned Systems Technology 008 | Alti Transition UAS | Ground control systems | Xponential 2016 report | Insitu Orbital N20 | UAVs | Solar power | Oceanology International 2016 report
8 Platform one June/July 2016 | Unmanned Systems Technology Driverless cars Road ahead for data fusion Chip suppliers are positioning themselves for the transition from advanced driver assist systems (ADAS) to fully autonomous driverless cars as the architecture of the vision systems changes (writes Nick Flaherty). “We are certainly in the early days of autonomous driving already, and we believe it’s a full progression from ADAS to full autonomous operation, not a separate category,” said Brooke Williams, business manager of the ADAS system-on-chip (SoC) business unit at Texas Instruments. TI has shipped more than 15 million SoC devices for front camera, surround view, radar systems and sensor fusion to 15 tier one suppliers that are selling to 25 car makers for 100 separate lines. “The building blocks are in there now, but to enable more advanced driving, fusion of that data needs to be done in a very intelligent way,” Williams said. “As things progress to autonomous operation there will need to be a higher level of data integration, and what we are already seeing is the centralisation of that edge signal processing, so we will see dumb sensors feeding raw image data into a central location. “When we start to see raw data movement, that’s when we need a lot more signal processing or a board with multiple SoCs. The beauty is that the architecture for image processing and control is heavily optimised for algorithms that run on the edge or in the middle. “For centralised control you need to have a very efficient architecture, so when we talk about front camera algorithms they are the most challenging. “You have to run six to ten algorithms on a single device, and that has to be in a very low power envelope of 2-4 W, and very few architectures can handle that level of processing at that power,” he said. While TI would not comment on the next generation devices for this, ST Microelectronics has extended its relationship with Israeli chip designer Mobileye for exactly this application. The two are working on the fifth- generation EyeQ5 SoC as the central computer for performing sensor fusion for fully autonomous driving vehicles. The chip will be designed in advanced 10 nm or below FinFET silicon technology with eight multi-threaded CPU cores coupled with 18 cores of Mobileye’s next- generation vision processors. This will provide eight times the performance of current devices, processing 12 Tflops with a power consumption of less than 5 W. Engineering samples of EyeQ5 are expected to be available by the first half of 2018, with development boards in the second half for running traditional computer vision but also neural network algorithms. “EyeQ5 is designed to serve as the central processor for future fully autonomous driving for both the sheer computing density – it can handle around 20 high-resolution sensors – and for increased functional safety,” said Prof Amnon Shashua, co- founder, CTO and chairman of Mobileye. “This continues the development Mobileye began in 2004 with EyeQ1, with optimised architectures to support intensive computations at power levels below 5 W to allow passive cooling in an automotive environment.” Mobileye is also adding a hardware security module to the chip so that system integrators can support over- the-air software updates and secure in-vehicle communications. Creating the root of trust is based on a secure boot from an encrypted storage device. The data interface is also vital, and EyeQ5 will support at least 40 Gbit/s through two PCIe Gen4 ports for inter- processor comms with other chips. The EyeQ5 SoC is designed to be the central processor for fully autonomous driving
Made with FlippingBook
RkJQdWJsaXNoZXIy MjI2Mzk4