January 17, 2023
Engineering excellence doesn’t exist without empathy for the end user. You can design a structurally robust steering wheel, but if it’s in the wrong position, it doesn’t serve its main purpose — to be steered. Likewise, when designing for human vision, a display doesn’t need the highest possible luminosity or energy output. An optical engineer might be looking for these qualities, but that expectation overlooks the purpose of the design. Optimizing a vehicle display is about centering the end user and the way they perceive the world around them.
When it comes to functionality, there are a number of design areas that offer opportunities for improvement:
Early on, engineers work to design a display on the nanoscale by testing how light interacts with different geometries like LED crystals or OLED layers. As the design scales up from photonic component modeling to optical component modeling, the focus shifts to elements like polarizing layers and surface coatings. As most automakers rely on outside vendors for display technology, the final stages of design are focused on integrating the display into the vehicle.
Sun studies identify critical sun positions and account for reflections from other light sources. In some cases, post-processing algorithms allow users to see screen effects like glare. Ansys’ algorithm accounts for the biology of the human eye with features that emulate adaptation time or even color blindness. If the driver is in the dark for five minutes and their eyes suddenly hit the display, they’re going to perceive it differently than they would after five minutes in bright light. The eye is constantly adapting to light settings, and the right simulation tools can render these kinds of realities.
When simulating human vision using physics-based rendering, predicting the quality and performance of a future product is more valuable than creating a realistic render of an existing prototype. One of the key requirements for physics-based renderings is high-performance computing (HPC) on central processing units (CPUs) or graphics processing units (GPUs). Ansys is quickly adding GPU support because they operate closer to real time. These results can be analyzed and experienced, using the human eye parameters mentioned above, in the Ansys Human Vision Lab.
At the end of the day, specifications are important, but sometimes they come from uninformed sources or are simply stricter than the application demands. Simulation software can get results fast, but there’s something to be said for experiencing the end product. When a driver gets into a vehicle, they won’t be measuring how much light comes out of the display, but they will be thinking about the way that vehicle makes them feel. The ability to anticipate those feelings and perceptions — and make design decisions without over-engineering — makes virtual prototyping an incredibly valuable tool.
Don’t just look at the numbers — experience your results. For more information on luminance matching and human vision, watch our on-demand webinar.