Skip to Main Content

      

ANSYS BLOG

January 06, 2020

The Autonomous Vehicle System iDAR Mimics Human Eyes

For autonomous driving to be safe, engineers need to design systems and sensors that can detect, interpret and react to hazards on the road.

That is why AEye is using simulation to create an Intelligent Detection and Ranging (iDARTM) platform that mimics how human eyes focus on the road.

The company will be using Ansys SPEOS to model the optics of its sensor platform and Ansys VRXPERIENCE to test and validate it within a realistic virtual environment.


Visit ANSYS at CES


The iDAR Platform Helps Autonomous Systems Assess the Road

Traditionally, engineers would have to test and validate their sensor systems using physical prototypes that consume a lot of time and budgets.

However, with the iDAR platform, a sensor system can be tested virtually over millions of miles in a few days.

Thanks to this autonomy on-demand, AEye and its OEM and Tier 1 customers will be able to address use cases systematically and gain more intelligence from the edge.

This systematic approach is important because you can’t put an autonomous system on the road and expect to test it through every potential scenario it could experience. The road is an unpredictable place so it’s impossible to tell when an edge case could pop up. Setting up these scenarios using physical prototypes could be impossible, expensive and dangerous.

As customer adopt iDAR, Ansys pervasive simulation software will be there to fully validate its performance. This will help to reduce development time and optimize autonomous implementation.

To see how this autonomous driving technology works, visit Ansys at CES.

To learn more about iDAR, read: AEye and Ansys Accelerate Autonomous Driving Safety.

불가능을 가능하게 만들 준비가 되셨나요?

문의하기

* = 필수 항목

문의해 주셔서 감사합니다!

We’re here to answer your questions and look forward to speaking with you. A member of our Ansys sales team will contact you shortly.

Racecars on a track