Complexity Engineering

How Credible are FE Car Crash Simulation Models?

Car crash simulation is performed by means of explicit Finite Element models, containing many millions of elements. These models are computationally intensive and parallel computing is used to perform the task. Evidently, the reason car crash is simulated in computers is to reduce and/or avoid expensive physical testing. Abundant literature is available on the subject.

The present blog will illustrate an example of how QCM technology can help evaluate and measure the credibility of finite element crash models. Clearly, in order to replace a physical test with a computer simulation, it is not only necessary for the simulation to reflect correctly the underlying physical phenomena, one must also know to what degree this is accomplished. The issue is addressed here.

A physical crash test is performed, producing sensor outputs in 65 locations on the car. QCM produces the following Complexity Map, which shows the nearly 2000 interdependencies between the said 65 data channels. The map has a density of 98%, pointing to two things: 1) the phenomenon is complex and 2) probably some sensors are redundant.

In parallel, a crash test simulation is performed in the same configuration as the physical test and information is collected in the same 65 locations and with the same sampling frequency. The following Complexity Map is produced.

This time, map density is 91%, still very high but 7% less than the physical test.

There is an 11% difference in complexity, 18% difference in Total Entropy, and 7% difference in map density between the test and simulation cases.  Based on complexities, the model credibility can be quantified in 88%. In other words, the model misses approximately 12% of the total information contained in the test data. If the complexity between the two cases is “small”, the weak condition of model credibility is satisfied.

The strong condition of high model credibility is that the topologies of the test and simulation Complexity Maps be equal.  This can be verified by dividing the norm of the difference of the two maps (treated as matrices) by the norm of the test map.

Test:   norm (CT) = 273.65

Simulation: norm (CS) = 242.61

Difference: norm |(CT –  CS)| = 82.48

Model Credibility Index = 1 –  norm |(CT –  CS)| / norm (CT)  = 1 – 82.48/273.65 = 0.698

Based on the above approach, it may be concluded that the simulation embraces approximately 70% of the test. In other words, the model neglects approximately 30% of what took place in the physical test.

In conclusions, based on the weak condition, the model correctly reproduces approximately 88% of the information contained in the test data.

Based on the strong condition – requiring topological similarity of the two complexity maps –  it may be concluded that the simulation embraces approximately 70% of the test. In other words, the model neglects approximately 30% of what took place in the physical test.

In the past we have used QCM to analyse numerous finite element crash models for various car manufacturers. The conclusion is that so far we have been unable to find a model which misses less than 20% of the physics of a real car crash.

The question this begs is this: is it really necessary to build computer models with disputable levels of detail, which require immense compute power, and still miss a large amount of physics? While this is done in the name of so-called “precision”, it is often forgotten that models must be realistic, not precise. There is no need to play extravagant video games.

PS. No Artificial Intelligence, no Machine Learning is needed to do this. All it takes is a single execution of OntoTest™.

PPS. You can apply this technique to any test-computer simulation case.

Unknown's avatar

Established originally in 2005 in the USA, Ontonix is a technology company headquartered in Como, Italy. The unusual technology and solutions developed by Ontonix focus on countering what most threatens safety, advanced products, critical infrastructures, or IT network security - the rapid growth of complexity. In 2007 the company received recognition by being selected as Gartner's Cool Vendor. What makes Ontonix different from all those companies and research centers who claim to manage complexity is that we have a complexity metric. This means that we MEASURE complexity. We detect anomalies in complex defense systems without using Machine Learning for one very good reason: our clients don’t have the luxury of multiple examples of failures necessary to teach software to recognize them. We identify anomalies without having seen them before. Sometimes, you must get it right the first and only time!

2 comments on “How Credible are FE Car Crash Simulation Models?

  1. Mike Sheh's avatar

    Interesting!

    So have you compared CFD models, NVH models, etc. And 1-D models like GT-Power or Adams?

    Mike

    Liked by 1 person

  2. ontonixqcm's avatar

    We have compared models of all sorts over the years. The ones that perform best – this is NOT a general statement for all such models – are linear models. Whenever there are elements of nonlinearity, or uncertainty, capturing the physics of a real phenomenon is not easy. Crash is probably the worst because of its exquisitely stochastic nature. No two crash tests are identical. A crash test cannot be repeated. So what are we trying to achieve? A lucky hit?

    Like

Leave a reply to ontonixqcm Cancel reply