WHY DO TECHNOLOGY IMPROVEMENT RATES MATTER?
Technology decision makers often face the choice of investing in one core technology (believing it to be crucial to their mission/ existence), an alternate technology (a key “disruptor”) or multiple technologies to hedge against failure. Information on improvement rates is crucial in making best choices in such allocation decisions. Although such decisions are made at a point-in-time, it needs to be recognized that they deal with outcomes that are changing (sometime rapidly) with time.
Often times for a key alternate technology (“disruptor”), subjective expert opinions assess when the specific technology will be ready. There are fundamental challenges with this approach in addition to the normal issues with subjectivity. New technologies are often over-hyped (to the extent that there is now a well-known model called the Gartner hype cycle). The creators of the technology and those covering them often have an incentive in overstating when they will be ready. Decision-makers can be led astray by lab vs real product behavior and leadership can begin to dangerously believe in “reality distortion” due to their or their team’s overly high self-assessment.
This can lead to large losses for multiple stakeholders including investors, consumers/users, employees and often times general public. Some examples include-probable over-valuation of Uber and others due to very wrong estimates of viability timing for autonomous cars, medical harm due to Theranos, busts in virtual reality and other cases which are essentially about incorrect, overly eager, estimation of timing for potentially valuable technologies.
All technologies require relatively long periods of technology development before they become market ready and an under-estimation of the time to completion can lead to loss of trust, serious disillusionment and withdrawal of a support causing collapse of critical mass of knowhow (for instance, the Artificial intelligence “winters”, the collapse of cleantech in US). Furthermore, investing in a new technology too early can lead to substantial costs (due to long investment periods without revenues) and ultimately unrealized benefits as the competitors may catch-up quickly and be more effective at marketing as well as establishing a “dominant” design (the innovators of the smart phone -IBM Simon and other innovators only received losses for their innovative behavior while the well designed and well-timed iPhone made a fabulous fortune for Apple). On the other hand, there is also a serious risk of under-estimating an upcoming technology. This can be especially critical if the upcoming technology threatens the core-technology underlying the main product of the incumbent- examples include Blockbuster, Nokia, Research in Motion(Black Berry) Barnes & Noble and others.
Thus, it is important for technology managers of all kinds to make accurate estimates of technology timeliness/readiness. It follows that most technology managers consider the rate of improvement of performance for a technology as an important indicator of the potential future importance of that technology. Limitations on acting on that correct intuition include the lack of objective information about rates of improvement and the lack of clarity about the methodology to use that information most effectively.
 Hoisl, Karin, Tobias Stelzer, and Stefanie Biala. 2015. “Forecasting Technological Discontinuities in the ICT Industry.” Research Policy 44 (2): 522–32. https://doi.org/10.1016/j.respol.2014.10.004.