The Pareto principle (also known as the 80/20 rule, the law of the vital few and the principle of factor sparsity[1][2]) states that for many outcomes, roughly 80% of consequences come from 20% of causes (the "vital few").[1]
In 1941, management consultant Joseph M. Juran developed the concept in the context of quality control and improvement after reading the works of Italian sociologist and economist Vilfredo Pareto, who wrote in 1906 about the 80/20 connection while teaching at the University of Lausanne.[3] In his first work, Cours d'économie politique, Pareto showed that approximately 80% of the land in the Kingdom of Italy was owned by 20% of the population. The Pareto principle is only tangentially related to the Pareto efficiency.
Mathematically, the 80/20 rule is roughly described by a power law distribution (also known as a Pareto distribution) for a particular set of parameters. Many natural phenomena are distributed according to power law statistics.[4] It is an adage of business management that "80% of sales come from 20% of clients."[5]
In 1941, Joseph M. Juran, a Romanian-born American engineer, came across the work of Italian polymath Vilfredo Pareto. Pareto noted that approximately 80% of Italy's land was owned by 20% of the population.[6][4] Juran applied the observation that 80% of an issue is caused by 20% of the causes to quality issues. Later during his career, Juran preferred to describe this as "the vital few and the useful many" to highlight that the contribution of the remaining 80% should not be discarded entirely.[7]
The demonstration of the Pareto principle is explained by a large proportion of process variation being associated with a small proportion of process variables.[2] This is a special case of the wider phenomenon of Pareto distributions. If the Pareto index α, which is one of the parameters characterizing a Pareto distribution, is chosen as α = log45 ≈ 1.16, then one has 80% of effects coming from 20% of causes.[8]
The term 80/20 is only a shorthand for the general principle at work. In individual cases, the distribution could be nearer to 90/10 or 70/30. There is also no need for the two numbers to add up to the number 100, as they are measures of different things. The Pareto principle is an illustration of a "power law" relationship, which also occurs in phenomena such as bush fires and earthquakes.[9] Because it is self-similar over a wide range of magnitudes, it produces outcomes completely different from Normal or Gaussian distribution phenomena. This fact explains the frequent breakdowns of sophisticated financial instruments, which are modeled on the assumption that a Gaussian relationship is appropriate to something like stock price movements.[10]
Using the "A:B" notation (for example, 0.8:0.2) and with A + B = 1, inequality measures like the Gini index (G) and the Hoover index (H) can be computed. In this case both are the same:
Pareto analysis is a formal technique useful where many possible courses of action are competing for attention. In essence, the problem-solver estimates the benefit delivered by each action, then selects a number of the most effective actions that deliver a total benefit reasonably close to the maximal possible one.
Pareto analysis is a creative way of looking at causes of problems because it helps stimulate thinking and organize thoughts. However, it can be limited by its exclusion of possibly important problems which may be small initially, but will grow with time. It should be combined with other analytical tools such as failure mode and effects analysis and fault tree analysis for example.[citation needed]
This technique helps to identify the top portion of causes that need to be addressed to resolve the majority of problems. Once the predominant causes are identified, then tools like the Ishikawa diagram or Fish-bone Analysis can be used to identify the root causes of the problems. While it is common to refer to pareto as "80/20" rule, under the assumption that, in all situations, 20% of causes determine 80% of problems, this ratio is merely a convenient rule of thumb and is not, nor should it be considered, an immutable law of nature.
The application of the Pareto analysis in risk management allows management to focus on those risks that have the most impact on the project.[11]
Steps to identify the important causes using 80/20 rule:[12]
Pareto's observation was in connection with population and wealth. Pareto noticed that approximately 80% of Italy's land was owned by 20% of the population.[6] He then carried out surveys on a variety of other countries and found to his surprise that a similar distribution applied.[citation needed]
A chart that demonstrated the effect appeared in the 1992 United Nations Development Program Report, which showed that the richest 20% of the world's population receives 82.7% of the world's income.[13] However, among nations, the Gini index shows that wealth distributions vary substantially around this norm.[14]
Quintile of population | Income |
---|---|
Richest 20% | 82.70% |
Second 20% | 11.75% |
Third 20% | 2.30% |
Fourth 20% | 1.85% |
Poorest 20% | 1.40% |
The principle also holds within the tails of the distribution. The physicist Victor Yakovenko of the University of Maryland, College Park and AC Silva analyzed income data from the US Internal Revenue Service from 1983 to 2001 and found that the income distribution of the richest 1–3% of the population also follows Pareto's principle.[16]
In Talent: How to Identify Entrepreneurs, economist Tyler Cowen and entrepreneur Daniel Gross suggest that the Pareto Principle can be applied to the role of the 20% most talented individuals in generating the majority of economic growth.[17] According to the New York Times in 1988, many video rental shops reported that 80% of revenue came from 20% of videotapes (although rarely rented classics such as Gone with the Wind must be stocked to appear to have a good selection).[18]
In computer science the Pareto principle can be applied to optimization efforts.[19] For example, Microsoft noted that by fixing the top 20% of the most-reported bugs, 80% of the related errors and crashes in a given system would be eliminated.[20] Lowell Arthur expressed that "20% of the code has 80% of the errors. Find them, fix them!"[21] It was also discovered that, in general, 80% of a piece of software can be written in 20% of the total allocated time. Conversely, the hardest 20% of the code takes 80% of the time. This factor is usually a part of COCOMO estimating for software coding.[citation needed]
Occupational health and safety professionals use the Pareto principle to underline the importance of hazard prioritization. Assuming 20% of the hazards account for 80% of the injuries, and by categorizing hazards, safety professionals can target those 20% of the hazards that cause 80% of the injuries or accidents. Alternatively, if hazards are addressed in random order, a safety professional is more likely to fix one of the 80% of hazards that account only for some fraction of the remaining 20% of injuries.[22]
Aside from ensuring efficient accident prevention practices, the Pareto principle also ensures hazards are addressed in an economical order, because the technique ensures the utilized resources are best used to prevent the most accidents.[23]
The Pareto principle is the basis for the Pareto chart, one of the key tools used in total quality control and Six Sigma techniques. The Pareto principle serves as a baseline for ABC-analysis and XYZ-analysis, widely used in logistics and procurement for the purpose of optimizing stock of goods, as well as costs of keeping and replenishing that stock.[24] In engineering control theory, such as for electromechanical energy converters, the 80/20 principle applies to optimization efforts.[19]
The remarkable success of statistically based searches for root causes is based upon a combination of an empirical principle and mathematical logic. The empirical principle is usually known as the Pareto principle.[25] With regard to variation causality, this principle states that there is a non-random distribution of the slopes of the numerous (theoretically infinite) terms in the general equation.
All of the terms are independent of each other by definition. Interdependent factors appear as multiplication terms. The Pareto principle states that the effect of the dominant term is very much greater than the second-largest effect term, which in turn is very much greater than the third, and so on.[26] There is no explanation for this phenomenon; that is why we refer to it as an empirical principle.
The mathematical logic is known as the square-root-of-the-sum-of-the-squares axiom. This states that the variation caused by the steepest slope must be squared, and then the result added to the square of the variation caused by the second-steepest slope, and so on. The total observed variation is then the square root of the total sum of the variation caused by individual slopes squared. This derives from the probability density function for multiple variables or the multivariate distribution (we are treating each term as an independent variable).
The combination of the Pareto principle and the square-root-of-the-sum-of-the-squares axiom means that the strongest term in the general equation totally dominates the observed variation of effect. Thus, the strongest term will dominate the data collected for hypothesis testing.
In the systems science discipline, Joshua M. Epstein and Robert Axtell created an agent-based simulation model called Sugarscape, from a decentralized modeling approach, based on individual behavior rules defined for each agent in the economy. Wealth distribution and Pareto's 80/20 principle emerged in their results, which suggests the principle is a collective consequence of these individual rules.[27]
In 2009, the Agency for Healthcare Research and Quality said 20% of patients incurred 80% of healthcare expenses due to chronic conditions.[28] A 2021 analysis showed unequal distribution of healthcare costs, with older patients and those with poorer health incurring more costs.[29] The 80/20 rule has been proposed as a rule of thumb for the infection distribution in superspreading events.[30][31] However, the degree of infectiousness has been found to be distributed continuously in the population.[31] In epidemics with super-spreading, the majority of individuals infect relatively few secondary contacts.