Creating a SIPOC Chart
February 12, 2025
The world seems to have turned on its head. Countries like the United States and the United Kingdom were earlier frontrunners in the race for globalization. They were increasingly pressurizing developing countries to open their markets. They thought this would allow them to have increased sales for their products. They never thought the reverse could […]
Why Data is the New Oil Data is the New Oil, proclaimed, India’s Richest Person and the one of the World’s leading Businessperson, Mukesh Ambani. He was referring to the fights over data in the digital age being similar to the battles over Oil in the Industrial Age. As Oil was the Lubricant that made […]
Mergers and acquisitions used to be fairly straightforward in the yesteryears. This is because companies would only make acquisitions within their own industry. This meant that if a technology-based startup were up for grabs, the list of potential suitors would only include companies like Google and Microsoft. This has completely changed now. Traditional companies like […]
Imagine one fine evening your nation’s top leader suddenly appeared on television to make an address. In this speech, he stated that the money which currently accounts for over 80% by value and 20% by volume of the entire money supply are suddenly going to be invalid i.e. worthless. They would cease to be legal […]
Trade Wars and Zero Sum Thinking The ongoing trade wars between the United States, on one hand, and the rest of its trading partners, including China, Canada, and the EU or the European Union stem from a basic misinterpretation of global trade in Zero Sum terms or I win, You Lose kind of thinking. As […]
One of the greatest criticisms that have been mounted against the six sigma methodology is the fact that there is a possibility that the entire system is built on fudged numbers. Statisticians have claimed that the name six sigma is misleading. Here are the reasons why:
The statistical term Six Sigma actually refers to a process in which there will be 2 defects per billion times the process is run. However, the definition of Six Sigma accepted by modern day practitioners is a much easier to follow, 3.4 defects per million. Although even achieving efficiency of 3.4 defects per million, makes the process achieve near zero and therefore negligible defects, the statistical name 6 sigma is misleading. The values 3.4 defects per million, in reality, correspond to 4.5 sigma levels. The balance is accounted for by the 1.5 sigma shift.
The logic behind the 1.5 sigma shift is rooted in empirical studies. Empirical studies have shown that processes tend to fare better in the short term than they actually do in the long term. This is because in the short term, there is only normal process variation that needs to be dealt with. However in the long term cases of special process variation also occur. This results in the process performing at 6 sigma levels in the short run but at 4.5 sigma levels in the long run.
The long term variation in the process variation is accounted for by one of the two reasons:
As a result of either of the above reasons, or a combination of both, the process fails to meet its Six Sigma objectives. This phenomenon is called long term dynamic mean variation.
Now, we know that the Six Sigma criteria are not met because of long term dynamic mean variation. But how do we know that we need to remove 1.5 sigma from both sides of the normal curve. Well, it isn’t a statistical reality but just an industry convention.
Motorola was the pioneer of Six Sigma methodology worldwide. They have made empirical studies about the processes that they have improved and concluded that a 1.5 sigma shift occurs. While many statisticians have called this 1.5 sigma shift arbitrary, the industry wants to go the Motorola Way and 3.4 defects per million which define 4.5 Sigma have become an industry wide accepted definition of a Six Sigma process.
Your email address will not be published. Required fields are marked *