Snapdragon Ridge hails autonomous vehicles with customisable SoC

Scalable and open autonomous driving solutions from Qualcomm Technologies, a subsidiary of Qualcomm, consist of the family of Snapdragon Ride Safety system on chips (SoCs), Snapdragon Ride Safety Accelerator and Snapdragon Ride autonomous stack.

Snapdragon Ride aims to address the complexity of autonomous driving and advanced driver assistance systems (ADAS) by leveraging power-efficient hardware, artificial intelligence (AI) technologies and what Qualcomm describes as a pioneering autonomous driving stack. The combination of Snapdragon Ride SoCs, accelerator and autonomous stack offers vehicle manufacturers a scalable solution designed to support three industry segments, namely: L1/L2 active safety ADAS for vehicles that include automatic emergency braking, traffic sign recognition and lane keeping assist functions, L2+ convenience ADAS for vehicles featuring automated highway driving, self-parking and urban driving in stop-and-go traffic and L4/L5 fully autonomous driving for urban driving, robo-taxis and robo-logistics.

The Snapdragon Ride platform is based on the Snapdragon family of automotive SoCs and accelerator. It is built on scalable and modular heterogenous high-performance multi-core CPUs, energy efficient AI and computer vision (CV) engines, and a graphics processing unit (GPU). It offers 30 Tera operations per second (TOPS) for L1/L2 applications to over 700 TOPS at 130W for L4/L5 driving. This enables it to be used in designs that can be passively or air-cooled to reduce cost and increase reliability by avoiding the use of expensive liquid cooled systems. It also allows for simpler vehicle designs and extends the driving range for electric vehicles (EVs), says Qualcomm. The Snapdragon Ride SoCs and accelerator are designed for functional safety ASIL-D systems.

The Snapdragon Ride autonomous stack is modular and scalable for automotive manufacturers to use optimised software and applications for complex use cases, such as self-navigating human-like highway driving as well as modular options like perception, localisation, sensor fusion and behaviour planning. This software infrastructure for Snapdragon Ride supports customer-specific stack components to be co-hosted with the Snapdragon Ride autonomous stack components.

The Snapdragon Ride integrated safety board support package has safe OS and hypervisors and operates within safety frameworks from automotive industry leaders, including Adaptive AutoSAR. It has optimised foundational function libraries for computer vision, sensor signal processing, and standard arithmetic libraries

AI tools for improving model efficiencies, as well as optimising runtime on heterogeneous compute units

Snapdragon Ride is expected to be available for pre-development to automakers and tier-1 suppliers in the first half of 2020 and the company anticipates Snapdragon Ride-enabled vehicles to be in production in 2023.

http://www.qualcomm.com

> Read More

Siemens and Arm partner to develop and validate automotive safety systems

To help automotive designers develop and validate electronic systems, Siemens has paired its PAVE360 software with Arm’s automotive IP in a partnership that will bring IP, methodologies, processes and tools together to help automakers, integrators and suppliers collaborate, design and bring to market their next-generation platforms much faster, says Siemens.

Bringing together IC methodologies, processes and tools will help the automotive supply chain solve design and verification challenges, by validating differentiated safety enabled systems, ICs and software solutions in the context of the entire vehicle the company added.

Siemens’ PAVE360 is part of Siemens Digital Industries Software’s Xcelerator portfolio. This partnership was formed to address the increasingly complexity in developing platforms to realise active-safety, advanced driver assistance, in-vehicle infotainment, digital cockpits, vehicle-to-vehicle/vehicle-to-infrastructure and self-driving vehicles. Advances in computing and sensor technology are enabling companies to redefine mobility beginning with the ICs and software within automotive electronics systems.

Siemens’ PAVE360 digital twin environment, featuring Arm IP, applies high-fidelity modelling techniques from sensors and ICs to vehicle dynamics and the environment within which a vehicle operates. Using Arm IP, including Arm Automotive Enhanced (AE) products with functional safety support, digital twin models can run entire software stacks providing early metrics of power and performance while operating in the context of a high-fidelity model of the vehicle and its environment.

Using PAVE360 with Arm automotive IP allows automakers and suppliers to simulate and verify sub-system and system on chip (SoC) designs and understand how they perform within a vehicle design from the silicon level up, before the vehicle is built. By rethinking IC design, manufacturers can consolidate electronic control units (ECUs), saving thousands of dollars per vehicle through reduced the number of circuit boards and lengths of wire within the vehicle design, says Siemens. These savings also reduce vehicle weight which can promote longer range electric vehicles (EVs).

Siemens‘ PAVE360 platform will be demonstrated in the Siemens Mobility booth at CES in Las Vegas (7 to 10 January 2020).

http://www.siemens.com

> Read More

Harting snaps into Rinspeed concept car for power, data and signals

Swiss automotive manufacturer, Rinspeed is presenting its metroSnap concept vehicle at CES this week, and Harting has designed an interface that supplies the vehicle with power, data and signals.

The module helps the driver use the vehicle for a wide range of tasks. It is divided into the Skateboard, which serves as an optimised electric vehicle (EV), and the Pod, which can be switched to accommodate different tasks. Communication, signal transmission and the power supply must all work smoothly between these two elements. Connectivity is automatically established as soon as the Pod and the Skateboard are securely connected.

The two companies have partnered before, with a fast charger from Harting Automotive, Harting’s subsidiary which reported a sharp increase in demand for e-mobility solutions. The company leverages its decades of experience in the field of connection and transmission technology to develop and produce charging equipment for electric and plug-in hybrid vehicles (PHEVs).

In 2016, Harting’s Mica was integrated into Rinspeed’s Etos vehicle for autonomous emission and condition monitoring. In 2017, Harting its miniMICA to support Rinspeed’s Oasis car. The following year, Harting provided the fast-charging technology for the Snap and in 2019, the microSnap was charged using the fast-charging push.

http://www.harting.com

> Read More

NIR sensor saves power for mobile 3D optical sensing systems

3D optical sensing applications such as face recognition, payment authentication can operate at much lower power than alternative implementations, using the CGSS130, CMOS global shutter sensor (CGSS) near infra red (NIR) image sensor by ams.

The CGSS130 enables battery-powered devices to run for longer between charges while supporting sophisticated sensor functions.

According to ams, the CGSS130 sensor is four times more sensitive to NIR wavelengths than other image sensors on the market today, and detects reflections from very low power IR emitters in 3D sensing systems. It is the IR emitter that consumes most of the power in face recognition and other 3D sensing applications, says ams, which means using the CGSS130 sensor will enable manufacturers to extend battery runtime in mobile devices.

The 1.3Mpixel sensor also creates the opportunity to implement face recognition in wearable devices and in other products which are powered by a very small battery, or to enable a new range of applications beyond face recognition as the increased sensitivity extends the measurement range for the same power budget.

Following ams’ partnership with CMOS image sensor supplier, SmartSens Technology, the first 3D active stereo vision (ASV) reference design based on the CGSS130 was produced. The 1.3Mpixel stacked BSI sensor offers the highest quantum efficiency at 940nm, claims ams and, by supplying all main parts of the 3D system (illumination, receiver, software) it enables superior system performance with lower costs and a faster time to market.

The stacked BSI process used to fabricate the CGSS global shutter image sensors, results in a small footprint of 3.8 x 4.2mm. The sensor produces monochrome images with an effective pixel array of 1080 x 1280 at a maximum frame rate of 120 frames per second. The high frame rate and global shutter operation produce clean images free of blur or other motion artefacts, says ams.

The sensor also offers a high dynamic range (HDR) mode in which it achieves dynamic range of more than 100dB. It also implements advanced functions such as external triggering, windowing, and horizontal or vertical mirroring.

The CGSS130 is available for sampling.

ams is demonstrating the CGSS130 at CES, in the Venetian Tower, Suite 236 / 30th floor.

https://ams.com

> Read More

About Smart Cities

This news story is brought to you by smartcitieselectronics.com, the specialist site dedicated to delivering information about what’s new in the Smart City Electronics industry, with daily news updates, new products and industry news. To stay up-to-date, register to receive our weekly newsletters and keep yourself informed on the latest technology news and new products from around the globe. Simply click this link to register here: Smart Cities Registration