This digital twin predicts gear and bearing fatigue and reliability, accounting for all the same influences on component fatigue as are included in state-of-the-art design analysis, but in a process that runs automatically, requiring no human interaction or intervention.

Technological trends tend to attract convenient phrases whose usage “goes viral,” picking up more and more references across the world. In particular, the “Digital Twin” is something that has been spoken about widely, yet it has rarely (if ever) been seen — for gearboxes at least.

For sure, the “Design Twin” exists — a simulation model of the gearbox used during the design process. However, this is simply computer-aided engineering (CAE) as practiced by engineers since the late 1980s.

The real Digital Twin should exist alongside the physical asset in monitoring and informing about the health of the gearbox in service. In this respect, much has been said and promised about Digital Twins, but very little has actually been delivered.

This article describes one of the projects that Hexagon has completed on behalf of a handful of clients for the implementation of gearbox Digital Twins. These clients are all long-standing users of Hexagon’s software for gearbox simulation, Romax, and are globally recognized brands and cover a wide range of different ground vehicles.

Telematic data is uploaded from the vehicles to the Cloud, where it is automatically processed by a simulation model of the gearbox, which gives the same results as the model used to design the gearbox by the design engineers. Fatigue data is converted into predicted reliability data, which can be presented to drivers/operators/owners via apps on common mobile devices, along with recommendations to “maintain” or “replace.” More detailed data analysis allows unparalleled insight into how the vehicles are being driven, especially regarding the variability of operation. The whole process runs without human intervention or interaction.

However, the transition from fatigue damage to reliability is not without its complications. This article studies the data sources from standards and published technical papers and identifies an enormous variation in implied performance of the gears. Practical difficulties in using such data are also identified. The likely source of this variation is discussed and practical solutions proposed.

In the round, this article outlines the potential for Digital Twins in vehicle transmissions and shows that, despite the hype and broken promises of the past, practical implementations are working today and the future looks enormously promising.

1 Defining the Digital Twin

In recent years, few terms in engineering have been given as much airtime as “Digital Twin”; however, it is often applied in so many different contexts that its interpretation has become extremely confused. The concept of a Digital Twin was described in 2001 in a presentation by Michael Grieves, being described as a “Conceptual Idea for PLM” [1]. In this, digital information would be a “twin” of the physical system, and it would be embedded within the physical system itself and be linked with that physical system through its entire lifecycle. Grieves expanded on the idea over subsequent years and, by studying work by the same, originating author, it can be seen how the different interpretations of the term Digital Twin have arisen.

In 2016 [1], Grieves expanded on the definitions, confirming the Digital Twin covers all aspects of the product life cycle, but with different purposes. For example, the Digital Twin Prototype (DTP) covers the design phase, when the physical product does not actually exist. Following from this, the Digital Twin Instance (DTI) is created as a digital representation of a specific physical product, to be linked to that physical product through its entire life. It is interesting that he goes on to talk about Digital Twin Aggregates, where data from multiple DTIs is aggregated within what he calls a Digital Twin Environment (DTE).

It is interesting to note that, using this definition, the Digital Twin Prototype aligns pretty well with another commonly-used term, the “Design Twin.” Within this context, the DTP/Design Twin is clearly a computer simulation of the engineered system for use during the design phase. The author finds this curious since, for all intents and purposes, this is essentially a computer-aided engineering (CAE) model, provided to and used by industry for well over 30 years and promoted within the context of Concurrent Engineering, a theme that rose to prominence in the late 1980s. While CAE models of many different systems exist and may contribute to various Digital Twins, this is not the focus of this article.

For the purposes of this article, the Digital Twin refers to a digital asset that exists alongside the physical asset during its operational life, providing insight into and feedback on the physical asset’s performance and health. Thus, the focus is on the Digital Twin Instance, with the potential to aggregate data into a Digital Twin Aggregate for the gearbox design being considered, and within the Digital Twin Environment set up by Hexagon.

In respect of the physical asset across its life, nothing is more important about its performance than its ability to function, i.e., reliability, and for the purposes of CAE, nothing is of greater importance than to be able to predict the reliability of a product being designed. Thus, for the purpose of this study, while gearbox noise, efficiency, and thermal behavior may be of interest, the primary interest is fatigue and reliability.

2 Working with the OEM

Hexagon has been working on In-service Digital Twins with a number of different manufacturers of ground vehicles, extending across different non-competing industries. This article is based on a collaboration with one company in particular, a world-renowned manufacturer of ground vehicles. For reference, the vehicles are IC engine driven, and the gearboxes have several discrete ratios.

This article contains a wealth of data, all of it derived from the operation of these vehicles and processed by Hexagon using its Digital Twin. To protect the interests of its client (the OEM), the data has been subject to obfuscation by the provision of graphs without labels, and by normalization with respect to the design duty cycle.

Nonetheless, all the data shown is genuine data from real vehicles in operation and the narrative, insight and conclusions from the study are the same as for the OEM, whose principal engineers have read and approved the text of this article prior to its submission.

Figure 1: Flow diagram illustrating the processing of CAN data from the vehicle.

3 Building the Cloud-based Digital Twin

While the Digital Twin Prototype (DTP) is not the focus of this article, the technical methods used owe their origins from the DTP. During the design phase, a designer must carry out calculations to confirm the gearbox is fit for purpose, and, in this respect, durability/reliability is the most important consideration. The prediction of gearbox reliability does not start from a clean sheet. Gearbox fatigue has been a subject of mathematical methods that have been developed over the years and implemented in standards (ISO 6336 for gears, ISO 16281 for bearings [2]) that have become universally implemented.

Additionally, the Digital Twin has included many of the refinements that have been proven to be essential in the prediction of gear and bearing performance over recent decades. Housing stiffness has been shown to have an impact on gear and bearing misalignment and hence life [3], and this is included; likewise, gear micro-geometry is introduced to accommodate misalignment from system deflections, affecting gear stress and fatigue; finally, bearing internal load-sharing, pre-load and misalignment are also included.

All these influences are part of the commercial software package, Romax ENDURO, which has been marketed under various Romax names since its release in 1994 and which was acquired by Hexagon in 2020. Romax ENDURO is principally used during the design phase of the gearbox, i.e. as a DTP.

Both ISO 16281 and ISO 6336 output component fatigue damage. This does not predict failure. Rather, 100 percent damage for the L10 life of bearings indicates 10 percent failure, whereas 100 percent damage (a safety factor of 1.0) for gears indicates 1 percent failure.

Table 1: Gear reliability data from a range of published sources.

Being a long-term user of Romax as a DTP, the OEM had a confirmed model of the gearbox that could be used for the Digital Twin study. This model had been used in the design of the gearbox in question (i.e. a ‘Design Twin’), and a “Design Duty Cycle” had been established by the OEM with the intention of representing the usage pattern that the vehicles would see during an anticipated working life.

However, the OEM has strong ambitions for using data to learn more about their applications and to derive competitive advantage. In recent years, the OEM had installed on its standard production vehicles the instrumentation necessary to transmit to the Cloud significant quantities of CAN (Controlled Area Network) data. It is important to note this setup was established to understand the vehicle as a whole, not just the gearbox — this gearbox Digital Twin work simply worked with data already being downloaded. Of this CAN data being downloaded, the following data were available at a frequency of 1Hz:

  • Engine speed.
  • Engine torque.
  • Selected ratio.
  • Gearbox oil temperature.

It is obvious how this data can be used as an import to define an “in-service duty cycle,” against which component fatigue calculations can be carried out. Indeed, Romax has been able to import time domain data for fatigue calculations for more than 15 years. However, this project provided additional challenges.

The first was to set everything on the Cloud and to have it operating without human intervention or human interaction. This is required because, in the end, the Digital Twin would need to process the data from many thousands of vehicles. The Digital Twin was set up such that the data was transferred and processed daily for each vehicle. Discussions did take place regarding having more (and less) frequent data transfer; however, it was decided daily processing would provide the best balance between cost/complexity and insight.

Data integration — the ability to pull data across the Cloud and integrate different solutions — was also required. In this respect, the team at Romax was able to collaborate with the team at Xalt Solutions, who reside within a different division of Hexagon, and who have extensive experience in integrating diverse data systems. Xalt Integration is the tool that links different systems together, and Xalt Mobility presents the results and provides web-based access to the user.

Hexagon successfully built the Digital Twin and, as of the end of December 2022, it had successfully processed eight months’ worth of data from 10 different test vehicles, with more being processed each day. The fact of the Digital Twin’s successful operation is noteworthy. However, of greater interest is what the acquired data has already revealed in terms of how to approach the reliability engineering of gearboxes. This is to be the focus of the rest of the article.

4 Predicting Gearbox Reliability

4.1 Challenges in Predicting Bearing Reliability

As has been stated, it is possible to calculate the fatigue damage for the gears (contact and bending) according to ISO 6336 and for the bearings according to ISO 16128; 100 percent damage pertains to a 1 percent failure rate for the gears and 10 percent failure rate for the bearings. The conversion from fatigue damage to reliability for bearings was discussed at length in a paper [4] by the same author at CTI International Congress and Expo at Novi, Michigan, May 23-25, 2023.

The article identifies the Digital Twin took the most “official” recommendation for the reliability characteristic of the rolling element bearings, i.e., the reliability factor A1 from ISO. Fundamental to this is a Weibull shape parameter (β) of 1.5.

However, the article also identified there is substantial reason to question this data, owing to large variations in recommended values of β from different sources. GE [5] propose values that range from 0.7 to 3.5, while Bertsche and Lechner [6] propose 1.1 to 1.35. Meanwhile, a globally-renowned aerospace manufacturer [7] claims values as high as 4.0.

The differences in these values are not simply of academic interest — they have a massive impact upon the predicted failure rates of gearboxes. With such a wide range of values of β available, the net ends up being cast so wide that pretty much “any” observed outcome could be said to correlate with one or other of the quoted reliability models.

This article identified and quantified a key source of variability — that of the duty cycle. A fleet of 10 vehicles was monitored for a period of eight months, with their usage data recorded at a rate of 1 Hz. Without the Digital Twin, these vehicles would have been assumed to have identical usage patterns. Through simulation, the article was able to quantify the error, i.e. the reduction, in observed Weibull shape parameter that would have occurred if the failure analysis of the fleet was carried out against time and not damage, that is to say, ignoring the duty cycle of each vehicle against including it in the analysis.

The analysis showed if a value of β =1.5 is the correct descriptor of bearing reliability, failing to account for the variation in vehicle usage would mean the fleet would appear to fail according to a value of β =0.67. Even if bearings are, in fact, much more reliable, with a value of β =4.0, this makes no difference and, again, the fleet would appear to fail according to a value of β =0.68.

This analysis shows the absolute importance of monitoring the usage of each vehicle if the prediction of component reliability is to make any sense. This was impractical/uneconomical in the past; however, with recent advances in data acquisition, a far more rigorous approach to component reliability analysis now becomes possible. Therefore, our attention should turn to what would be used for gear reliability.

4.2 Sources of Data for Gear Reliability

There is a wide range of sources for values on gear reliability, and unfortunately the picture appears even more confusing than it is for bearings. For a start, different studies (using different data sets) conclude that different reliability models provide the best fit. Some references use Weibull shape parameter (β), some use standard deviations in strength, and some use lognormal distributions. Additionally, there is the provision (sometimes applied, sometimes not) of different models for contact and bending, and also for conditions above and below the endurance limit.

Nonetheless, it is useful to see all the different reference data in one place. Much of this has been taken from the Gear Solutions article by Dr. Hein et al., [8] from February 2020, but more have been added from other sources.]

One clear difficulty in comparing these values comes from the fact they are presented in different forms. By presenting the values in the form of a Weibull Shape Parameter, the results point to a variation in failure against time. By presenting a standard deviation in strength, the source is indicating a normal distribution in strength, which implies a variation in failure against time.

4.3 Converting Reliability Guidelines into a common format and comparing them

To make sense of these values, they need to be converted to a “common currency,” and within this study it was decided to convert to Weibull Shape Parameter. The approach taken by Hexagon contained several assumptions, and while alternative assumptions could have been made, Hexagon does not believe it would affect the overall outcome of the study.

First, the data for the standard deviation in strength was converted to an equivalent Weibull Shape Parameter. A graphical representation can be seen in Figure 2. The S-N curve from the standards (ISO and AGMA) refer to the 1 percent failure rate. Assuming a normal distribution and using the standard deviation in strength, a distribution in strength can be implied.

Figure 2: Graphical illustration of the conversion between a distribution in strength to a distribution in life.

A hypothetical population of gears with this distribution in strength was generated numerically, and their resulting failure points calculated. The limited life part of the S-N curve for contact for case-carburized gears was used. This was because case-carburized gears are most common, contact failures are more common than bending failures and the greatest number of gear fatigue tests have focused on this part of the S-N curve, owing to the greater ease in achieving failures.

Figure 3 shows the data arising from the FZG data for the full set of data. This shows a normal distribution in strength, projected onto the time axis using the S-N curve of ISO 6336, does not give a perfect Weibull Distribution. This pattern was also seen for all the other values used.

Figure 3: Derivation of equivalent Weibull Shape Parameter using the data from Stahl [10], for the whole population.
Figure 4: Derivation of equivalent Weibull Shape Parameter using the data from Stahl [10], for the first 50% of the population.

It was decided to concentrate on the earlier sections of the population, since it is unlikely a complete population of gears would be allowed to fail in service. By concentrating on the first 50 percent of failures, a closer match to Weibull and a clearer value for β could be achieved.

At this point, a “sense check” can be made. Given a standard deviation of 3.5 percent of the mean and 2.326 standard deviations between mean (50 percent failure) and 1 percent failure, this means the mean strength of the population (and hence the gear that provides the point that crosses the 50 percent point on the Y axis) is 8.14 percent higher than the strength of the gear that sits at y=1%. Using the slope of the S-N curve for contact from ISO (13.22), this indicates the mean life should be approximately 3.0 times the life for 1 percent. Inspection of the graph indicates the results (5.5e6 cycles for 1%; 1.6e7 cycles for 50%) match what is expected.

The corresponding analysis using the data from AGMA [12] and Gleason [9] can be seen in Figure 5 and Figure 6 respectively.

Figure 5: Derivation of equivalent Weibull Shape Parameter using the data from AGMA [12], for the first 50% of the population.
Figure 6: Derivation of equivalent Weibull Shape Parameter using the data from Gleason [9] for the first 50% of the population.

Table 2: Gear reliability data from a range of published sources, converted to equivalent Weibull Shape Parameters (β), for contact.

Thus, we can take Table 2 and complete it with values that allow comparison on a like-for-like basis, at least for Contact.

4.4 Understanding the Varying Guidelines in Gear Reliability

So, we end up with a wide range of different values for β, starting with 0.85 for Gleason [9], 1.06 for AGMA [12], 1.1 from Bertsche and Lechner [6], extending up to 4.9 from Stahl [10] and as high as 6.0 from GE [5]. This is an extraordinary range. It is not just an aspect of “academic interest” — if a design engineer has a system for which the reliability needs to be calculated, the biggest impact on the result will not be the use of AGMA or ISO, or the method for calculating misalignment, or how KHBeta is calculated. The single main determining factor affecting the reliability result becomes which data a given engineer chooses in order to extrapolate from the predicted life for the 1 percent failure rate to the predicted reliability of the population.

This can be illustrated in Figure 7, where the same nominal life for 1 percent is used and the reliability of the population is predicted using the various values of β derived from the various public sources. The problem with the lower values of β is clear — there is an enormous variation across the population. For the FZG data, the 80 percent failure point (20 percent reliability) is only ~3 times the 1 percent failure point, but for AGMA it is ~120 times and for Gleason it is ~400 times.

Figure 7: Representative plot of the reliability of a whole population of gears using different values for the Weibull Shape Parameter.

Consider the implications of using the AGMA and Gleason data: If the purpose of the calculation is to predict when a gear will fail, the result becomes so uncertain that realistically no insight is possible and the calculation is essentially meaningless. If the purpose is to check observed failure rates against predicted failure rates, the net ends up being cast so wide that pretty much “any” observed outcome could be said to correlate with one or other of the quoted reliability models.

Another problem arises when using some of the data [10] in that different behavior is described above and below the endurance limit. This may accurately reflect the results derived from the testing of gears in laboratory conditions, where each gear is subject to constant loading, which is targeted to be in either limited life or high cycle regimes. However, if a gear is subject to in-service loading that includes both high cycle and limited life regimes, how is this to be considered? What happens if we have a mix of 80%/20%, 20%/80%, 60%/40%? If we have a loading that is just above the endurance limit for 1 percent of the population, do we assume it to be below the endurance limit (and this requiring a different reliability model) for the remaining 99 percent of the population?

The result of this wide variation in results (and the subsequent uncertainty regarding differing behavior above and below the endurance limit) means that engineers tend to take the relatively comfortable approach of predicting “indicative” reliability, saying something “is as durable/reliable as the last design we did.” However, if this is the case then Computer Aided Engineering in general is profoundly guilty of what could be described as “over-promising and under-delivering.” The promise has always been that product performance can be predicted so failure modes can be understood, and “quantified trade-offs” carried out with confidence. In the case of gear reliability, these trade-offs would include:

  • How will my failure rate increase (∆ %) if I decrease my center distance (∆ mm), thereby saving weight and material?
  • What are the chances that the gearbox will fail if I run it for another 1,000 hours?

Instead, design engineers tend to approach these considerations more in the form of:

  • If we go below center distance “X” we will exceed the safety factor that we have typically used for this application.
  • While some of our gearboxes have continued to be used beyond this point, we do not have enough confidence to know whether this is suitable.

This raises two questions:

  • Why has this happened?
  • What should we do about it?

4.4.1 How did this variability arise?

By reflecting on its work with different companies and research organizations over the last 30 years in different industries, Hexagon believes there are two reasons for this range of values and that understanding these can help indicate what the best solution is moving forward:

The first thing to recognize is that, to a certain extent, variability has dropped (and values of β have improved) over time. The Gleason data [9] is from 1965, whereas the higher values are more recent.

It is to be expected that, over time, the cleanliness of steels and the quality control of the manufacturing processes (heat treatment, grit blasting, shot peening, etc.) has improved, leading to a reduction in the variability of the strength as well as an improvement in overall strength. Certain academics have also suggested this as an observation [11].

The other aspect is to look at where these values were derived from. In some instances [10, 11], the data comes from test programs from university laboratories, where carefully designed test rigs with high levels of stiffness and alignment are used to test gears carefully manufactured and inspected, using consistent and carefully monitored loads. The cycles to failure of each gear would have been carefully recorded. Therefore, the variability of the gear lives could be put down to variations in the material properties alone.

In other instances [5, 6], the values would have been based on observations from in-service gears. In these instances, there are many more factors that affect the gear reliability, which are unlikely to have been measured.

Housing stiffness has been known to affect gear misalignment and hence gear stress for many years [3]; however, such studies may well not have been able to account for this factor, and, in any case, production gearbox housings are subject to variations in wall thickness, and this will not have been accounted for.

Gear micro-geometry is known to affect gear stress and hence durability, and most design engineers account for this by using the nominal (target) micro-geometry for their prediction of gear durability.

However, variations from the nominal exist, and while it is possible to measure gear geometry and import the actual tooth profile for use in the stress calculation [13], it is reasonable to assume these influences also were not included.

Thus, we see that while the values arising from university laboratories account for the variation on gear material strength alone, other values account for a much wider set of variables, factors known to affect gear durability but are unlikely to have been known for any given failed gear. These include:

  • Gear micro-geometry variations.
  • Gear misalignment variations, owing to variations in housing stiffness and gear alignment tolerances.
  • Variations in loading.
  • Potential for inconsistent lubrication, lubricant degradation, water ingress, etc.

During the execution of this Digital Twin work, Hexagon was able to acquire usage data from many vehicles subject to nominally identical operating environments. A study [4] was carried out, focusing on bearings (bearings were chosen because these were found to be the components most likely to fail) to account for the impact of not knowing the loading conditions if the Weibull analysis were to be performed to determine the reliability of bearings.

This study showed that, no matter how high the actual value of β is for the bearings (i.e., a predictable failure), if the failure analysis is carried out against time and not damage, the study would deduce a much lower value of β (i.e., an unpredictable failure). This shows how the failure to include in analysis factors that substantially affect a component/system can lead to the incorrect conclusion that the component/system’s reliability is profoundly unpredictable.

Figure 8 gives a graphical representation of what Hexagon believes is happening for gears, and where the variation in values of β comes from. Variability in material strength alone may well correspond to a β of ~5.0, and a quite “peaky” distribution in reliability, but when unknown factors are included, this distribution gets flattened, so that, when many significant factors are left unaccounted for, the observer is left with what appears to be a profoundly unpredictable system.

Figure 8: Illustration of how the inclusion/exclusion of various factors in a Reliability Analysis affects the resulting value of β.

In this situation, it is unsurprising that design engineers will err on the side of caution, basing new designs on old design practices rather than seeking optimum designs with full confidence in quantifiable trade-offs.

4.4.2 How should this data be used?

This insight provides an indication of how to move forward with future developments. Reliability predictions must start somewhere, and the best indication is the variability indicated by the university studies [10, 11] properly identify the variability of strength of modern gear materials. This gives an “upper limit” on the predictability of in-service applications, since it is not possible for any production gearbox to be as tightly controlled, or to have as much data as a research laboratory, in terms of manufacturing, inspection, assembly, and loading.

This article has shown how various influences that affect the durability of gearboxes push the observed failure rate toward randomness if they are not accounted for, and in effect, this determines what data should be used in the prediction of gearbox reliability. Load spectrum, gear geometry, and gear alignment are all examples of these influences.

Therefore, if any of these factors are known, or are tightly controlled, then there will be less scatter imparted on the life of the gearbox components, and a higher value of β can be used. If these factors are not known, or they are poorly controlled, then a lower value of β should be used.

Broadly speaking, any prediction of reliability must account for as many of the variables that affect reliability as possible. When working on Failure Mode Avoidance, Tim Davis [14] extended the work of Taguchi, Miscke, and Phadke to take signal factors and noise factors into account (see Figure 9). In this instance, there is a danger that if the signal-to-noise ratio decreases, it becomes difficult to achieve the ideal function of the system and error states occur.

Figure 9: The p-diagram to describe an engineering system or component [14].

The analogy for gearboxes is clear — the ideal function is reliable operation and component failure is an error state. Hexagon’s Digital Twin work has shown how vehicle usage can be a noise, but with the correct setup can be converted to a sgnal factor. Put another way, it is converted from an unknown to a known.

Within the framework of this Digital Twin, Hexagon has sought to convert one major factor (load spectrum) from an unknown to a known, from a noise to a signal factor. In the absence of reasons to the contrary, the Digital Twin uses the recommendations from Stahl [10] from the limited life for all loads, with different models for contact and bending. Furthermore, any Weibull analysis of failed gearboxes should provide decent clarity regarding the behavior of the population as a whole.

On the other hand, if the load spectrum is assumed and not measured, the analysis carried out by Hexagon implies a very different approach to reliability should be considered, for example Bertsche and Lechner [6]. In this case, it is unlikely a Weibull analysis of failed gearboxes will provide credible insight into the reasons for such failures.

Nonetheless, the Digital Twin, with the ability to record the load spectrum for every gearbox, provides the framework for the future of reliability engineering of gearboxes. Hexagon is working with a number of companies who are routinely harvesting vehicle usage data from substantial quantities of their vehicles. By feeding this data into best practice CAE tools, unprecedented accuracy can be achieved in the prediction of component and system reliability.

However, more could be achieved. As it stands, there would still be assumptions that could be improved upon. For example, every OEM has values for the allowable bending and contact stresses of its materials. These values substantially affect any predictions of reliability, yet even though they are affected by a company’s suppliers, machining, heat treatment and quality control, often the “reference” data from ISO 6336 part 5 is used.

With the cost of data acquisition plummeting, Hexagon believes it will start to see large scale data acquisition and failure recording. When combined in a disciplined manner, this would yield unprecedented insight into the durability and reliability of gearbox systems, insight that could be fed back to engineers for re-use in the design of the next generation of vehicles. Effectively, a “Digital Thread” as shown in Figure 10, would be complete.

Figure 10: An illustration of the Complete Digital Thread for Gears.

This article talks about Digital Twins, but there are other fashionable terms, too. The “Internet of Things,” “Big Data,” and “AI” are often discussed, yet it is not always clear how they will deliver benefits. The proposal for data acquisition, failure recording, and correlation is an example of how such fashionable ideas can be applied in an achievable manner to gearbox engineering to deliver benefits in terms of reduced cost, reduced risk, faster development times, reduced material usage, longer gearbox life, and improved sustainability.

Hexagon does not propose discarding the wealth of knowledge encapsulated in standards such as ISO 6336, but rather to build upon these standards and to refine them based on the evidence provided by the usage and failure patterns identified in each application. Some major OEMs already make changes to the standards for their own benefit and based on their own data, thereby giving them commercial advantage. This is set to accelerate massively.

Further comparisons are intriguing. Stahl’s data [8] was based on 509 gear tests. Soon, Hexagon’s Digital Twin will be extended to cover more than 10,000 gearboxes, each with more than 10 gears — each one effectively an individual test sample since its usage profile will be monitored throughout its life.

On one hand, Stahl’s data is more useful since it is available today, and all 509 gears are assumed to have failed and contributed to the statistical analysis, whereas the vast majority of the 100,000-plus gears under study in the Digital Twin will not fail. Additionally, even the early failures are not expected for a considerable period.

However, the failure analysis arising from such a Digital Twin would be focused on a specific application, a specific supply chain, a specific production line with its heat treatment, lubricant selection, etc.

Furthermore, over time, the number of gearboxes under study will expand another one or two orders of magnitude (perhaps more), and the failure data will start to stack up. When this happens, the statistical significance of the data arising from such a process, and hence its validity and accuracy, will start to dwarf that available from the university research programs that have supplied it to date. All of this will deliver a clear benefit to industry, to the users of such vehicles, and to the economy and environment.

Stahl’s data is already 24 years old. Compared to the alternatives, the time to results for the Digital Twin described here is probably not so onerous.

5 Conclusion

The term “Digital Twin” has often been spoken about, but rarely has it been seen to be put into operation and to deliver value in practice. Likewise, terms such as the “Internet of Things,” and “Big Data” may have delivered value when tracking people’s browsing habits and targeting advertising, but there has seemed to be little relevance to gearbox engineering.

Recent work by Hexagon shows that this is changing. In collaboration with a world-renown vehicle manufacturer, it has developed and put into service a cloud-based Digital Twin for a multi-speed gearbox.

This digital twin predicts gear and bearing fatigue and reliability, accounting for all the same influences on component fatigue as are included in state-of-the-art design analysis, but in a process that runs automatically, requiring no human interaction or intervention.

In converting from durability (percentage damage, safety factor) to reliability (failure rate), data has been referenced from various sources. However, these different sources have been shown to have widely varying values,which, if used in the prediction of gearbox reliability, would lead to huge variations in the results depending on which reference was used.

It has been argued that part of this variation is due to the passage of time, part of it due to the difference between research laboratory testing programs and observation from industrial usage patterns. The former tightly controls many of the factors that affect gear life (loading conditions, alignment, etc.), which are either assumed or left unknown in industrial applications.

The key to useful gear reliability predictions is to maximize the “signal-to-noise ratio,” and in practice this means measuring as many of the factors that affect gear life as possible. In the past, this has not been practical. However, due to recent developments, we are starting to see large programs of data acquisitions by leading vehicle companies. In the short term, this work will use current reference data on reliability to predict when gearboxes are likely to fail; over time, once proportions of the gearbox population have actually failed, the observed failure rate can be compared with the predicted failure rate and adjustments made.

The future of gear reliability prediction is starting to take shape. Manufacturers will be able to collect precise usage data on tens of thousands, probably millions of gears, and match this against failure data. Even though the failure rates will be small, the number of samples will be statistically significant, and will eventually be orders of magnitude greater than the number of samples that have been used to derive current reference data. Furthermore, each company’s data will be based on its application and production processes, rather than general gear applications. Rather than fresh attempts to re-write gear analysis methods, these will be minor tweaks to the current gear rating standards but will provide vastly greater confidence in the results.

Bibliography

  1. Grieves, M., 2016, “Origins of the Digital Twin Concept”, DOI: 10.13140/RG.2.2.26367.61609.
  2. ISO/TS 16281:2008, Rolling bearings — Methods for calculating the modified reference rating life for universally loaded bearings.
  3. James, B. et al., May 2000, “Predicting the effects of housing flexibility and bearing stiffness on gear misalignment and transmission noise using a fully-coupled non-linear hyperstatic analysis”, C577/005/2000, Proceedings of the Institution of Mechanical Engineers, May 2000.
  4. James, B. and Long, L., May 2023 “Demonstrating a Cloud-based Gearbox Digital Twin using telematics data for reliability predictions”, CTI USA Symposium, 24 – 25 May 2023, Novi, MI.
  5. Distribution Types | Predix APM | GE Digital.
  6. Bertsche, B. and Lechner, G., 2004, “Reliability in automotive and mechanical engineering – determination of reliability levels for components and systems” (In German: “Zuverlässigkeit im Fahrzeug- und Maschinenbau – Ermittlung von Bauteil- und System-Zuverlässigkeiten”), 3rd edition, Springer (publisher), Berlin, Germany.
  7. Private conversation during technical and commercial discussions.
  8. Hein, M. et al. “Reliability of Gears: Determination of Statistically Validated Material Strength Numbers”, Gear Solutions, February 2020.
  9. “The Strength of Bevel and Hypoid Gears” – Gleason, 1965.
  10. Stahl, K., 1999, “Life-time statistics (final report) – Statistical methods for estimation of the life- time and reliability of components and exemplary application for gears”;  in German: “Lebensdauerstatistik (Abschlussbericht) – Statistische Methoden zur Beurteilung von Bauteillebensdauer und Zuverlässigkeit und ihre beispielhafte Anwendung auf die Zahnräder”, IGF Nr. 11154, FVA-Nr. 304/I, Issue 580, Research Association for Drive Technology (FVA), Frankfurt/Main, Germany.
  11. Private conversation with Professor Dieter Hofmann, Newcastle University, circa 2003.
  12. ANSI/AGMA, 2004, “Fundamental Rating Factors and Calculation Methods for Involute Spur and Helical Gear Teeth,” ANSI/AGMA 2101-D04.
  13. James, B., Hughes, S., 2022, “Developing a digital thread linking gear design to manufacturing simulation and metrology”, VDI Wissenforum, Munich, 12th September 2022.
  14. Davis, T., 2004, “Measuring Robustness as a Parameter in a Transfer Function”, SAE 2004-01- 1130.