Quantcast
Channel: Minitab | Minitab
Viewing all 828 articles
Browse latest View live

Strangest Capability Study: Super-Zooper-Flooper-Do Broom Boom

$
0
0

by Matthew Barsalou, guest blogger

The great Dr. Seuss tells of Mr. Plunger who is the custodian at Diffendoofer School on the corner of Dinkzoober and Dinzott in the town of Dinkerville. The good Mr. Plunger “keeps the whole school clean” using a supper-zooper-flooper-do.

Unfortunately, Dr. Seuss fails to tell us where the supper-zooper-flooper-do came from and if the production process was capable.

supper-zooperLet’s assume the broom boom length was the most critical dimension on the supper-zooper-flooper-do. The broom boom length drawing calls for a length of 55.0 mm with a tolerance of +/- 0.5 mm. The quality engineer has checked three supper-zooper-flooper-do broom booms and all were in specification, so he concludes that there is no reason to worry about the process producing out of specification parts. But we know this not true. Perhaps the fourth supper-zooper-flooper-do broom boom will be out of specification. Or maybe the 1,000th.

It’s time for a capability study, but don’t fire up your Minitab Statistical Software just yet. First we need to plan the capability study. Each day the supper-zooper-flooper-do factory produces supper-zooper-flooper-do broom booms with a change in broom boom material batch every 50th part. A capability study should have a minimum of 100 values and 25 subgroups. The subgroups should be rational: that means the variability within each subgroup should be less than the variability between subgroups. We can anticipate more variation between material batches than within a material batch so we will use the batches as subgroups, with a sample size of four.

Once the data has been collected, we can crank up our Minitab and perform a capability study by going to Stat > Quality Tools > Capability Analysis > Normal. Enter the column containing the measurement values. Then either enter the column containing the subgroup or type the size of the subgroup. Enter the lower specification limit and the upper specification limit, and click OK.

Process Capability Report for Broom Boom Length

We now have the results for the supper-zooper-flooper-do broom boom lengths, but can we trust our results? A capability study has requirements that must be met. We should have a minimum of 100 values and 25 subgroups, which we have. But the data should also be normally distributed and in a state of statistical control; otherwise, we either need to transform the data, or identify the distribution of the data and perform capability study for nonnormal data.

Dr. Seuss has never discussed transforming data so perhaps we should be hesitant if the data do not fit a distribution. Before performing a transformation, we should determine if there is a reason the data do not fit any distribution.

We can use the Minitab Capability Sixpack to determine if the data is normally distributed and in a state of statistical control. Go to Stat > Quality Tools > Capability Sixpack > Normal. Enter the column containing the measurement values. Then either enter the column containing the subgroup or type the size of the subgroup. Enter the lower specification limit and the upper specification limit and click OK.

Process Capability Sixpack Report for Broom Boom Length

There are no out-of-control points in the control chart and the P value is greater than 0.05 so we can reject the null hypothesis of “The data is not normally distributed.” The data is suitable for a capability study.

The within subgroup variation is also known as short term capability and is indicated by Cp and Cpk. The between subgroup variability is also known as long term capability is given as Pp and Ppk. The Cp and Cpk fail to account for the variability that will occur between batches;  Pp and Ppk tell us what we can expect from the process over time.

Both Cp and Pp tell us how well the process conforms to the specification limits. In this case, a Cp of 1.63 tells us the spread of the data is much narrower than the width of the specification limits, and that is a good thing. But Cp and Pp alone are not sufficient. The Cpk and Ppk indicate how spread out the data is relative to the center of the specification limits. There is an upper and lower Cpk and Ppk; however, we are generally only concerned with the lower of the two values.

In the supper-zooper-flooper-do broom boom length example, a Cpk of 1.10 is an indication that the process is off center. The Cpk is 1.63, so we can reduce the number of potentially out-of-specification supper-zooper-flooper-do broom booms if we shift the process mean down to center the process while maintaining the current variation. This is a fortunate situation as it is often easier to shift the process mean than to reduce the process variation.

Once improvements are implemented and verified, we can be sure that the next supper-zooper-flooper-do the Diffendoofer School purchases for Mr. Plunger will have a broom boom that is in specification if only common cause variation is present.

 

About the Guest Blogger

Matthew Barsalou is a statistical problem resolution Master Black Belt at BorgWarner Turbo Systems Engineering GmbH. He is a Smarter Solutions certified Lean Six Sigma Master Black Belt, ASQ-certified Six Sigma Black Belt, quality engineer, and quality technician, and a TÜV-certified quality manager, quality management representative, and auditor. He has a bachelor of science in industrial sciences, a master of liberal studies with emphasis in international business, and has a master of science in business administration and engineering from the Wilhelm Büchner Hochschule in Darmstadt, Germany. He is author of the books Root Cause Analysis: A Step-By-Step Guide to Using the Right Tool at the Right TimeStatistics for Six Sigma Black Belts and The ASQ Pocket Guide to Statistics for Six Sigma Black Belts.

 

Five Reasons Why a Value Stream Map Is Like a Pirate’s Treasure Map (and One Reason Why It Is ...

$
0
0

Ahoy, matey! Ye’ve come to the right place to learn about Value Stream Maps (VSM).  Just as a treasure map can lead a band o’ pirates to buried treasures, so too can the VSM lead a process improvement bilge rat to the loot buried deep inside a process! Minitab’s Quality Companion has an easy-to-use VSM tool to guide yer way.Skull and Crossbones

Use a value stream map to illustrate the flow of materials and information as a product or service moves through the value stream. A value stream is the collection of all activities, both value-added and non-value added, that generate a product or service.  The VSM is one of the most useful tools in your tool bag because it helps document the process, identify wasteful activities and spot opportunities to improve.

In this blog post, I look at five reasons why a VSM is like a pirate’s treasure map and one reason why it is not!

Reason 1: The Map Starts and Ends at the GEMBA!

GembaMaps have been around since ancient times and are often associated with pirates.  When pirates buried their treasure on a remote island, they relied on the treasure map to lead them back to that crucial spot. 

In Lean, the crucial spot is called the GEMBA.  It’s where the ‘activity is happening,’ so it is the ‘place to be’!  When pirates created their map, they usually started where they buried the gold and drew the map backward. Process improvement practitioners often do just that when creating a value stream map: start at the end of the process and work backward. But whether you start at the end or the beginning, to accurately capture the process and all rework loops, you must go to the GEMBA, walk the process and talk to the operators.

Reason 2: The Maps Are Hand-Drawn

When pirates created a map, they used a scrap of parchment and an ink quill to draw the most important landmarks leading them back to their treasure. Take a cue from the pirates and use pencil and paper to start your VSM. That way you can draw the map fairly quickly and change it easily as you learn more about the process. If you do projects with Quality Companion, you can then create the final map in the VSM tool so you can capture data pertaining to each process step, such as inventory levels, defect rate, and cycle or takt time.

Reason 3: The Maps Use Standard Symbols

Pirate MapUnless you’re a scallywag, you will notice that pirates use symbols to identify important landmarks on their maps. Mountains are upside-down V’s, a wavy circle is a lake, a skull-and-crossbones represents danger, a dashed line charts the path to follow, and so on, until X marks the spot. 

Similarly, a VSM uses symbols to illustrate the important parts of the value stream such as the process steps, suppliers/customers, inventory, transportation, product/information flow, and so on.

Quality Companion uses these standard, industry-recognized symbols so that everyone in your organization will be able to read and understand your VSM. For a full listing of the symbols available in the Quality Companion VSM tool, press the F1 key to open Help on the web and navigate to the Value Stream Map Shapes section.

Reason 4: Maps Contain Arcane Clues to Follow to Find the Hidden Treasure

Treasure ChestPirates worried that someone could steal their maps and find their loot before they returned to retrieve it. They set traps and used arcane clues to mislead potential thieves.

Like a pirate map, a VSM may seem difficult to decipher at first. But if you pull out your spyglass and look hard for clues, you will find the hidden gold. As you follow the map, keep on the lookout for the dangers: process waste! 

The easiest signs of waste to decipher include the piles of inventory just prior to the bottleneck step, excessive non-value added time, push instead of pull hand-offs, defect rate, scrap rate, equipment downtime, and excessive set-up times—to name a few. Look hard and ask lots of questions of the operators in the process. Often you don’t need to dig deep to find these improvement opportunities.

Reason 5: X Marks the Spot!

X marks the SpotShiver me timbers! On the pirate’s treasure map you will find the gold hidden under the big X! Same with VSMs: once you identify the improvement opportunities, use a Kaizen burst symbol to mark that spot. Gather your team of knowledgeable folks and start digging into the process. Look for the sources of waste, broken hand-offs, unclear decision points, rework loops, excessive activities and opportunities for simplification. 

One Important Reason a VSM Is Not Like a Pirate Map: Cooperation!

Pirates on a ship“Dead men tell no tales!” When pirates journeyed back to the site of the hidden treasure, a lot of backstabbing and trickery ensued. Typically, only one pirate arrived alive to claim the loot.

This should not be the case for your VSM effort.

Rarely does one person know all the details about a process!  Work as a team to document the process, collect and validate the data, and then interpret the map and brainstorm solutions together. Finding gold with a VSM is a team effort, not an individual effort.

Next time your process makes you feel like you want to walk the plank, pull out your VSM tools, weigh anchor, and hoist the mizzen! You’ll be glad you did!

Statistical Tools for Process Validation, Stage 1: Process Design

$
0
0

Process validation is vital to the success of companies that manufacture drugs and biological products for people and animals. According to the FDA guidelines published by the U.S. Department of Health and Human Services:Process Validation Stages

“Process validation is defined as the collection and evaluation of data, from the process design state through commercial production, which establishes scientific evidence that a process is capable of consistently delivering quality product.”
— Food and Drug Administration

The FDA recommends three stages for process validation. In this 3-part series, we will briefly explore the stage goals and the types of activities and statistical techniques typically conducted within each. For complete FDA guidelines, see www.fda.gov

Stage 1: Process Design

The goal of this stage is to design a process suitable for routine commercial manufacturing that can consistently deliver a product that meets its quality attributes. It is important to demonstrate an understanding of the process and characterize how it responds to various inputs within Process Design.

Example: Identify Critical Process Parameters with DOE

Suppose you need to identify the critical process parameters for an immediate-release tablet. There are three process input variables that you want to examine: filler%, disintegrant%, and particle size. You want to find which inputs and input settings will maximize the dissolution percentage at 30 minutes.

To conduct this analysis, you can use design of experiments (DOE). DOE provides an efficient data collection strategy, during which inputs are simultaneously adjusted, to identify if relationships exist between inputs and output(s). Once you collect the data and analyze it to identify important inputs, you can then use DOE to pinpoint optimal settings.

Running the Experiment

The first step in DOE is to identify the inputs and corresponding input ranges you want to explore. The next step is to use statistical software, such as Minitab, to create an experimental design that serves as your data collection plan.

According to the design shown below, we first want to use a particle size of 10, disintegrant of 1%, and MCC at 33.3%, and then record the corresponding average dissolution% using six tablets from a batch:

DOE Experiment

Analyzing the Data

Using Minitab’s DOE analysis and p-values, we are ready to identify which X's are critical. Based on the bars that cross the red significance line, we can conclude that particle size and disintegrant% significantly affect the dissolution%, as does the interaction between these two factors. Filler% is not significant.

Pareto chart

Optimizing Product Quality

Now that we've identified the critical X's, we're ready to determine the optimal settings for those inputs. Using a contour plot, we can easily identify the process window for the particle size and disintegrant% settings needed to achieve a percent dissolution of 80% or greater.

Contour plot

And that's how you can use design of experiments to conduct the Process Design stage. Next in this series, we'll look at the statistical tools and techniques commonly used for Process Qualification!

How to Use Data to Understand and Resolve Differences in Opinion, Part 1

$
0
0

Opinions, they say, are like certain anatomical features: everybody has one. Usually that's fine—if everybody thought the same way, life would be pretty boring—but many business decisions are based on opinion. And when different people in an organization reach different conclusions about the same business situation, problems follow. 

difference of opinion

Inconsistency and poor quality result when people being asked to make yes / no, pass / fail, and similar decisions don't share the same opinions, or base their decisions on divergent standards. Consider the following examples. 

Manufacturing: Is this part acceptable? 

Billing and Purchasing: Are we paying or charging an appropriate amount for this project? 

Lending: Does this person qualify for a new credit line? 

Supervising: Is this employee's performance satisfactory or unsatisfactory? 

Teaching: Are essays being graded consistently by teaching assistants?

It's easy to see how differences in judgment can have serious impacts. I wrote about a situation encountered by the recreational equipment manufacturer Burley. Pass/fail decisions of inspectors at a manufacturing facility in China began to conflict with those of inspectors at Burley's U.S. headquarters. To make sure no products reached the market unless the company's strict quality standards were met, Burley acted quickly to ensure that inspectors at both facilities were making consistent decisions about quality evaluations. 

Sometimes We Can't Just Agree to Disagree

The challenge is that people can have honest differences of opinion about, well, nearly everything—including different aspects of quality. So how do you get people to make business decisions based on a common viewpoint, or standard?

Fortunately, there's a statistical tool that can help businesses and other organizations figure out how, where, and why people evaluate the same thing in different ways. From there, problematic inconsistencies can be minimized. Also, inspectors and others who need to make tough judgment calls can be confident they are basing their decisions on a clearly defined, agreed-upon set of standards. 

That statistical tool is called "Attribute Agreement Analysis," and using it is easier than you might think—especially with data analysis software such as Minitab

What Does "Attribute Agreement Analysis" Mean? 

Statistical terms can be confusing, but "attribute agreement analysis" is exactly what it sounds like: a tool that helps you gather and analyze data about how much agreement individuals have on a given attribute.

So, what is an attribute? Basically, any characteristic that entails a judgment call, or requires us to classify items as this or that. We can't measure an attribute with an objective scale like a ruler or thermometer. The following statements concern such attributes:

  • This soup is spicy.
  • The bill for that repair is low
  • That dress is red
  • The carpet is rough
  • That part is acceptable
  • This candidate is unqualified

Attribute agreement analysis uses data to understand how different people assess a particular item's attribute, how consistently the same person assesses the same item on multiple occasions, and compares both to the "right" assessment. 

pass-fail

    This method can be applied to any situation where people need to appraise or rate things. In a typical quality improvement scenario, you might take a number of manufactured parts and ask multiple inspectors to assess each part more than once. The parts being inspected should include a roughly equal mix of good and bad items, which have been identified by an expert such as a senior inspector or supervisor. 

    In my next post, we'll look at an example from the financial industry to see how a loan department used this statistical method to make sure that applications for loans were accepted or rejected appropriately and consistently. 

    DMAIC Tools and Techniques: The Measure Phase

    $
    0
    0

    In my last post on DMAIC tools for the Define phase, we reviewed various graphs and stats typically used to define project goals and customer deliverables. Let’s now move along to the tools you can use in Minitab Statistical Software to conduct the Measure phase.

    Measure Phase Methodology

    The goal of this phase is to measure the process to determine its current performance and quantify the problem. This includes validating the measurement system and establishing a baseline process capability (i.e., sigma level).

    I. Tools for Continuous Data Gage RandR Gage R&R

    Before you analyze your data, you should first make sure you can trust it, which is why successful Lean Six Sigma projects begin the Measure phase with Gage R&R. This measurement systems analysis tool assesses if measurements are both repeatable and reproducible. And there are Gage R&R studies available in Minitab for both destructive and non-destructive tests.

    Minitab location:Stat > Quality Tools > Gage Study > Gage R&R Study OR Assistant > Measurement Systems Analysis.

    Gage Linearity and Bias

    When assessing the validity of our data, we need to consider both precision and accuracy. While Gage R&R assesses precision, it’s Gage Linearity and Bias that tells us if our measurements are accurate or are biased.

    Minitab location: Stat > Quality Tools > Gage Study > Gage Linearity and Bias Study.

    Gage Linearity and Bias

    Distribution Identification

    Many statistical tools and p-values assume that your data follow a specific distribution, commonly the normal distribution, so it’s good practice to assess the distribution of your data before analyzing it. And if your data don’t follow a normal distribution, do not fear as there are various techniques for analyzing non-normal data.

    Minitab location: Stat > Basic Statistics > Normality Test OR Stat > Quality Tools > Individual Distribution Identification.

    Distribution Identification

    Capability Analysis

    Capability analysis is arguably the crux of “Six Sigma” because it’s the tool for calculating your sigma level. Is your process at a 1 Sigma, 2 Sigma, etc.? It reveals just how good or bad a process is relative to specification limit(s). And in the Measure phase, it’s important to use this tool to establish a baseline before making any improvements.

    Minitab location: Stat > Quality Tools > Capability Analysis/SixpackOR Assistant > Capability Analysis.

    Process Capability Analysis

    II. Tools for Categorical (Attribute) Data Attribute Agreement Analysis Attribute Agreement Analysis

    Like Gage R&R and Gage Linearity and Bias studies mentioned above for continuous measurements, this tool helps you assess if you can trust categorical measurements, such as pass/fail ratings. This tool is available for binary, ordinal, and nominal data types.

    Minitab location: Stat > Quality Tools > Attribute Agreement AnalysisOR Assistant > Measurement Systems Analysis.

    Capability Analysis (Binomial and Poisson)

    If you’re counting the number of defective items, where each item is classified as either pass/fail, go/no-go, etc., and you want to compute parts per million (PPM) defective, then you can use binomial capability analysis to assess the current state of the process.

    Or if you’re counting the number of defects, where each item can have multiple flaws, then you can use Poisson capability analysis to establish your baseline performance.

    Minitab location:Stat > Quality Tools > Capability Analysis OR Assistant > Capability Analysis.

    Binomial Process Capability

    Variation is Everywhere

    As I mentioned in my last post on the Define phase, Six Sigma projects can vary. Every project does not necessarily use the same identical tool set every time, so the tools above merely serve as a guide to the types of analyses you may need to use. And there are other tools to consider, such as flowcharts to map the process, which you can complete using Minitab’s cousin, Quality Companion.

    Creating a Waterfall Chart in Minitab

    $
    0
    0

    While there are many graph options available in Minitab’s Graph menu, there is no direct option to generate a waterfall chart. This type of graph helps visualize the cumulative effect of sequentially introducing positive or negative values.

    In this post, I’ll show you the steps to follow to make Minitab display a waterfall chart even without a "waterfall chart" tool. If you don’t already have Minitab 17, you can download a free 30-day trial here.

    For the purpose of this post, I’ll replicate this sample waterfall chart that I found in Wikipedia:

    1

    In Minitab, we’ll need to set up the data in table form. Here is how I’ve set up the data in my Minitab 17 worksheet:

    2

    The tricky part is adding and subtracting to make sure that each section on each of the five bars is the right height. For example, the height of the Services Revenue bar is 630 (that’s 420 + 210 = 630).

    To make the Fixed Costs bar reflect a $170 decrease from $630, we enter Fixed Costs twice in the worksheet with values of 460 and 170 (that’s 630 - 170 = 460). We will sum the two values together when we create the bar chart, and we will use column C3 to make one bar with two sections representing those two values.

    To make the graph, go to Graph> Bar chart> A function of a variable> One Y Stack:

    3

    Complete the new window like this:

    1

    When you click OK in the window above, Minitab will create a graph that looks similar to the one below:

    5

    To get the final waterfall chart, the graph above will need to be manually edited. In the example below, I’ve hidden the sections of the bar chart that I don’t want to see. To hide a section of the bar chart, make sure only that section is selected (single-click on the bar you want to edit until only that section is selected) and then double-click to bring up the Edit Bars window. Next, make the selections show in the image below:

    6

    In the example below, I repeated the steps above to remove the section of each bar that I wanted to hide. I’ve also manually deleted the legend:

    7

    The graph above is almost ready, but to match our initial example I’ll make a few more manual edits as detailed below:

    1. Delete the Y-axis Label Sum of Profits by clicking on the label and using the Delete key on the keyboard. I also changed the title of the graph by clicking on the current title and typing in the title I wanted to see.
    2. Adjust the Y-axis tick labels to the values in the example by double-clicking on the Y-axis scale to bring up the Edit Scale window, selecting Position of ticks and typing in the values I’d like to see on the Y-axis: 0, 140, 280, 420, 560, 700.
    3. Manually change the colors of the Product Revenue and Services Revenue to green by selecting each bar individually, double-clicking to bring up the Edit Bars window and changing the Fill Pattern Background color.
    4. Add horizontal gridlines by right-clicking on the graph and choosing Add> Gridlines and selecting Y major ticks. I also double-clicked on one of the gridlines to bring up the Edit Gridlines window and changed the Major Gridlines to Custom and selected a solid line (instead of the default dotted line) so I could match the example more closely.
    5. Use the Graph Annotation toolbar to insert a text box (look for a button that looks like a capital T in the toolbars at the top), placing the text box for each bar where I want to see it, typing in the value I want to display ($420K, $210K, $-170K, etc.), and clicking OK. Finally, I double-clicked on each label to bring up the Edit Text window where I used the Font tab to change the font color from black to white.

    The final result looks very much like the example shown at the beginning of this post:

    8

    I hope you’ve enjoyed reading this post! For more information about editing graphs in Minitab 17, take a look at this online support page.

    How to Use Data to Understand and Resolve Differences in Opinion, Part 2

    $
    0
    0

    Previously, I discussed how business problems arise when people have conflicting opinions about a subjective factor, such as whether something is the right color or not, or whether a job applicant is qualified for a position. The key to resolving such honest disagreements and handling future decisions more consistently is a statistical tool called attribute agreement analysis. In this post, we'll cover how to set up and conduct an attribute agreement analysis. 

    Does This Applicant Qualify, or Not? 

    A busy loan office for a major financial institution processed many applications each day. A team of four reviewers inspected each application and categorized it as Approved, in which case it went on to a loan officer for further handling, or Rejected, in which case the applicant received a polite note declining to fulfill the request. filling out an application

    The loan officers began noticing inconsistency in approved applications, so the bank decided to conduct an attribute agreement analysis on the application reviewers.

    Two outcomes were possible: 

    1. The reviewers make the right choice most of the time. If this is the case, loan officers can be confident that the reviewers do a good job, rejecting risky applicants and approving applicants with potential to be good borrowers. 

    2. The reviewers too often choose incorrectly. In this case, the loan officers might not be focusing their time on the best applications, and some people who may be qualified may be rejected incorrectly. 

    One particularly useful thing about an attribute agreement analysis: even if reviewers make the wrong choice too often, the results will indicate where the reviewers make mistakes. The bank can then use that information to help improve the reviewers' performance. 

    The Basic Structure of an Attribute Agreement Analysis 

    A typical attribute agreement analysis asks individual appraisers to evaluate multiple samples, which have been selected to reflect the range of variation they are likely to observe. The appraisers review each sample item several times each, so the analysis reveals how not only how well individual appraisers agree with each other, but also howl consistently each appraiser evaluates the same item. 

    For this study, the loan officers selected 30 applications, half of which the officers agreed should receive approval and half which should be rejected. These included both obvious and borderline applications. 

    Next, each of the four reviewers was asked to approve or reject the 30 applications two times. These evaluation sessions took place one week apart, to make it less likely they would remember how they'd classified them the first time. The applications were randomly ordered each time.

    The reviewers did not know how the applications had been rated by the loan officers. In addition, they were asked not to talk about the applications until after the analysis was complete, to avoid biasing one another. 

    Using Software to Set Up the Attribute Agreement Analysis

    You don't need to use software to perform an Attribute Agreement Analysis, but a program like Minitab does make it easier both to plan the study and gather the data, as well as to analyze the data after you have it. There are two ways to set up your study in Minitab. 

    The first way is to go to Stat > Quality Tools > Create Attribute Agreement Analysis Worksheet... as shown here: 

    create attribute agreement analysis worksheet

    This option calls up an easy-to-follow dialog box that will set up your study, randomize the order of reviewer evaluations, and permit you to print out data collection forms for each evaluation session. 

    But it's even easier to use Minitab's Assistant. In the menu, select Assistant > Measurement Systems Analysis..., then click the Attribute Agreement Worksheet button:

    Assistant MSA Dialog

    That brings up the following dialog box, which walks you through setting up your worksheet and printing out data collection forms, if desired. For this analysis, the Assistant dialog box is filled out as shown here: 

    Create Attribute Agreement Analysis Worksheet

    After you press OK, Minitab creates a worksheet for you and gives you the option to print out data collection forms for each reviewer and each trial. As you can see in the "Test Items" column below, Minitab randomizes the order of the observed items in each trial automatically, and the worksheet is arranged so you need only enter the reviewers' judgments in the the "Results" column. 

    attribute agreement analysis worksheet

    In my next post, we'll analyze the data collected in this attribute agreement analysis. 

    The Empirical CDF, Part 1: What's a CDF?

    $
    0
    0

    T'was the season for toys recently, and Christmas day found me playing around with a classic, the Etch-a-Sketch. As I noodled with the knobs, I had a sudden flash of recognition: my drawing reminded me of the Empirical CDF Plot in Minitab Statistical Software. Did you just ask, "What's a CDF plot? And what's so empirical about it?" Both very good questions. Let's start with the first, and we'll save that second question for a future post.etch-a-sketch

    The acronym CDF stands for Cumulative Distribution Function. If, like me, you're a big fan of failures, then you might be familiar with the cumulative failure plot that you can create with some Reliability/Survival tools in Minitab. (For an entertaining and offbeat example, check out this excellent post, What I Learned from Treating Childbirth as a Failure.) The cumulative failure plot is a CDF.

    Even if you're not a fan of failure plots and CDFs, you're likely very familiar with the CDF's famous cousin, the PDF or Probability Density Function. The classic "bell curve" is no more (and no less) than a PDF of a normal distribution.

    For example, here's a histogram with a fitted normal PDF for PinLength.MTW, from Minitab's online Data Set Library.

    zzz

    To create this plot, do the following:

    1. Download the data file, PinLength.MTW, and open it in Minitab.
    2. Choose Graph > Histogram > With Fit, and click OK.
    3. In Graph variables, enter Length.
    4. Click the Scale button.
    5. On the Y-Scale Type tab, choose Percent.
    6. Click OK in each dialog box.

    The data are from a sample of 100 connector pins. The histogram and fitted line show that the lengths of the pins (shown on the x-axis) roughly follow a normal distribution with a mean of 19.26 and a standard deviation of 2.154. You can get the specifics for each bin of the histogram by hovering over the corresponding bar.

    zzz

    The height of each bar represents the percentage of observations in the sample that fall within the specified lengths. For example, the fifth bar is the tallest. Hovering over the fifth bar reveals that 18% of the bins have lengths that fall between 18.5 mm to 19.5 mm. Remember that for a moment.

    Now let's try something a little different.

    1. Double-click the y-axis.
    2. On the Type tab, select Accumulate values across bins.
    3. Click OK.

    zzz 

    zzz

    It looks very different, but it's the exact same data. The difference is that the bar heights now represent cumulative percentages. In other words, each bar represents the percentage of pins with the specified lengths or smaller.

    zzzzzz

    For example, the height of the fifth bar indicates that 55% of the pin lengths are less than 19.5 mm. The height of the fourth bar indicates that 37% of pin lengths are 18.5 or less. The difference in height between the 2 bars is 18, which tells us that 18% of the pins have lengths between 18.5 and 19.5. Which, if you remember, we already knew from our first graph. So the cumulative bars look different, but it's just another way of conveying the same information.

    You may have also noticed that the fitted line no longer looks like a bell curve. That's because when we changed to a cumulative y-axis, Minitab changed the fitted line from a PDF to... you guessed it, a cumulative distribution function (CDF). Like the cumulative bars, the cumulative distribution function represents the cumulative percentage of observations that have values less than or equal to X. Basically, the CDF of a distribution gives us the cumulative probabilities from the PDF of the same distribution.

    I'll show you what I mean. Choose Graph > Probability Distribution Plot > View Probability, and click OK. Then enter the parameters and x-value as shown here, and click OK.

    zzzzzz

    zzz

    The "Left Tail" probabilities are cumulative probabilities. The plot tells us that the probability of obtaining a random value that is less than or equal to 16 is about 0.065. That's another way of saying that 6.5% of the values in this hypothetical population are less than or equal to 16.

    Now we can create a CDF using the same parameters:

    1. Choose Graph > Empirical CDF > Single and click OK.
    2. In Graph variables, enter Length.
    3. Click the Distribution button.
    4. On the Data Display tab, select Distribution fit only.
    5. Click OK, then click the Scale button.
    6. On the Percentile Lines tab, under Show percentile lines at data values, enter 16.

    zzz

    The CDF tells us that 6.5% of the values in this distribution are less than or equal to 16, as did the PDF.

    Let's try another. Double-click the shaded area on the PDF and change x to 19.26, which is the mean of the distribution.

    zzz

    Naturally, because we're dealing with a perfect theoretical normal distribution here, half of the values in the hypothetical population are less than or equal to the mean. You can also visualize this on the CDF by adding another percentile line. Click the CDF and choose Editor > Add > Percentile Lines. Then enter 19.26 under Show percentile lines at data values.

    zzz

    There's a little bit of rounding error, but the CDF tells us the same thing that we learned from the PDF, namely that 50% of the values in the distribution are less than or equal to the mean.

    Finally, let's input a probability and determine the associated x-value. Double-click the shaded area on the PDF, but this time enter a probability of 0.95 as shown:

    zzz

    zzz

    The PDF shows that the x-value that is associated with a cumulative probability of 0.5 is 22.80. Now right-click the CDF and choose Add > Percentile Lines. This time, under Show percentile lines at Y values, enter 95 for 95%.  

    zzz

    Once again, other than a little rounding error, the CDF tells us the same thing as the PDF.

    For most people (maybe everyone?), the PDF is an easier way to visualize the shape of a distribution. But the nice thing about the CDF is that there's no need to look up probabilities for each x-value individually: all of the x-values in the distribution and the associated cumulative probabilities are right there on the curve.


    So Why Is It Called "Regression," Anyway?

    $
    0
    0

    Did you ever wonder why statistical analyses and concepts often have such weird, cryptic names?

    One conspiracy theory points to the workings of a secret committee called the ICSSNN. The International Committee for Sadistic Statistical Nomenclature and Numerophobia was formed solely to befuddle and subjugate the masses. Its mission: To select the most awkward, obscure, and confusing name possible for each statistical concept.

    A whistle-blower recently released the following transcript of a secretly recorded ICSSNN meeting:

    "This statistical analysis seems pretty straightforward…"

    “What does it do?”

    “It describes the relationship between one or more 'input' variables and an 'output' variable. It gives you an equation to predict values for the 'output' variable, by plugging in values for the input variables."

    “Oh dear. That sounds disturbingly transparent.”

    “Yes. We need to fix that—call it something grey and nebulous. What do you think of 'regression'?”

    “What’s 'regressive' about it? 

    “Nothing at all. That’s the point!”

    Re-gres-sion. It does sound intimidating. I’d be afraid to try that alone.”

    “Are you sure it’s completely unrelated to anything?  Sounds a lot like 'digression.' Maybe it’s what happens when you add up umpteen sums of squares…you forget what you were talking about.”

    “Maybe it makes you regress and relive your traumatic memories of high school math…until you  revert to a fetal position?”

    “No, no. It’s not connected with anything concrete at all.”

    “Then it’s perfect!”

     “I don’t know...it only has 3 syllables. I’d feel better if it were at least 7 syllables and hyphenated.”

    “I agree. Phonetically, it’s too easy…people are even likely to pronounce it correctly. Could we add an uvular fricative, or an interdental retroflex followed by a sustained turbulent trill?”

    The Real Story: How Regression Got Its Name

    Conspiracy theories aside, the term “regression” in statistics was probably not a result of the workings of the ICSSNN. Instead, the term is usually attributed to Sir Francis Galton.

    Galton was a 19th century English Victorian who wore many hats: explorer, inventor, meteorologist, anthropologist, and—most important for the field of statistics—an inveterate measurement nut. You might call him a statistician’s statistician. Galton just couldn’t stop measuring anything and everything around him.

    During a meeting of the Royal Geographical Society, Galton devised a way to roughly quantify boredom: he counted the number of fidgets of the audience in relation to the number of breaths he took (he didn’t want to attract attention using a timepiece). Galton then converted the results on a time scale to obtain a mean rate of 1 fidget per minute per person. Decreases or increases in the rate could then be used to gauge audience interest levels. (That mean fidget rate was calculated in 1885. I’d guess the mean fidget rate is astronomically higher today—especially if glancing at an electronic device counts as a fidget.)

    Galton also noted the importance of considering sampling bias in his fidget experiment:

    “These observations should be confined to persons of middle age. Children are rarely still, while elderly philosophers will sometimes remain rigid for minutes.”

    But I regress…

    Galton was also keenly interested in heredity. In one experiment, he collected data on the heights of 205 sets of parents with adult children. To make male and female heights directly comparable, he rescaled the female heights, multiplying them by a factor 1.08. Then he calculated the average of the two parents' heights (which he called the “mid-parent height”) and divided them into groups based on the range of their heights. The results are shown below, replicated on a Minitab graph.

    For each group of parents, Galton then measured the heights of their adult children and plotted their median heights on the same graph.

    Galton fit a line to each set of heights, and added a reference line to show the average adult height (68.25 inches).

    Like most statisticians, Galton was all about deviance. So he represented his results in terms of deviance from the average adult height.

    Based on these results, Galton concluded that as heights of the parents deviated from the average height (that is as they became taller or shorter than the average adult), their children tended to be less extreme in height. That is, the heights of the children regressed to the average height of an adult.

    He calculated the rate of regression as 2/3 of the deviance value. So if the average height of the two parents was, say, 3 inches taller than the average adult height, their children would tend to be (on average) approximately 2/3*3 = 2 inches taller than the average adult height.

    Galton published his results in a paper called “Regression towards Mediocrity in Hereditary Stature.

    So here’s the irony: The term regression, as Galton used it, didn't refer to the statistical procedure he used to determine the fit lines for the plotted data points. In fact, Galton didn’t even use the least-squares method that we now most commonly associate with the term “regression.” (The least-squares method had already been developed some 80 years previously by Gauss and Legendre, but wasn’t called “regression” yet.) In his study, Galton just "eyeballed" the data values to draw the fit line.

    For Galton, “regression” referred only to the tendency of extreme data values to "revert" to the overall mean value. In a biological sense, this meant a tendency for offspring to revert to average size ("mediocrity") as their parentage became more extreme in size. In a statistical sense, it meant that, with repeated sampling, a variable that is measured to have an extreme value the first time tends to be closer to the mean when you measure it a second time. 

    Later, as he and other statisticians built on the methodology to quantify correlation relationships and to fit lines to data values, the term “regression” become associated with the statistical analysis that we now call regression. But it was just by chance that Galton's original results using a fit line happened to show a regression of heights. If his study had showed increasing deviance of childrens' heights from the average compared to their parents, perhaps we'd be calling it "progression" instead.

    So, you see, there’s nothing particularly “regressive” about a regression analysis.

    And that makes the ICSSNN very happy.

    Don't Regress....Progress

    Never let intimidating terminology deter you from using a statistical analysis. The sign on the door is often much scarier than what's behind it. Regression is an intuitive, practical statistical tool with broad and powerful applications.

    If you’ve never performed a regression analysis before, a good place to start is the Minitab Assistant. See Jim Frost’s post on using the Assistant to perform a multiple regression analysis. Jim has also compiled a helpful compendium of blog posts on regression.

    And don’t forget Minitab Help. In Minitab, choose Help > Help. Then click Tutorials > Regression, or  Stat Menu >  Regression.

    Sources

    Bulmer, M. Francis Galton: Pioneer or Heredity and Biometry. Johns Hopkins University Press, 2003.

    Davis, L. J. Obsession: A History. University of Chicago Press, 2008.

    Galton, F. “Regression towards Mediocrity in Hereditary Stature.”  http://galton.org/essays/1880-1889/galton-1886-jaigi-regression-stature.pdf

    Gillham, N. W. A  Life of Sir Francis Galton. Oxford University Press, 2001.

    Gould, S. J. The Mismeasure of Man. W. W. Norton, 1996.

    How to Use Data to Understand and Resolve Differences in Opinion, Part 3

    $
    0
    0

    In the first part of this series, we saw how conflicting opinions about a subjective factor can create business problems. In part 2, we used Minitab's Assistant feature to set up an attribute agreement analysis study that will provide a better understanding of where and when such disagreements occur. 

    We asked four loan application reviewers to reject or approve 30  selected applications, two times apiece. Now that we've collected that data, we can analyze it. If you'd like to follow along, you can download the data set here.

    As is so often the case, you don't need statistical software to do this analysis—but with 240 data points to contend with, a computer and software such as Minitab will make it much easier. 

    Entering the Attribute Agreement Analysis Study Data

    Last time, we showed that the only data we need to record is whether each appraiser approved or rejected the sample application in each case. Using the data collection forms and the worksheet generated by Minitab, it's very easy to fill in the Results column of the worksheet. 

    attribute agreement analysis worksheet data entry

    Analyzing the Attribute Agreement Analysis Data

    The next step is to use statistics to better understand how well the reviewers agree with each others' assessments, and how consistently they judge the same application when they evaluate it again. Choose Assistant > Measurement Systems Analysis (MSA)... and press the Attribute Agreement Analysis button to bring up the appropriate dialog box: 

    attribute agreement analysis assistant selection

    The resulting dialog couldn't be easier to fill out. Assuming you used the Assistant to create your worksheet, just select the columns that correspond to each item in the dialog box, as shown: 

    attribute agreement analysis dialog box

    If you set up your worksheet manually, or renamed the columns, just choose the appropriate column for each item. Select the value for good or acceptable items—"Accept," in this case—then press OK to analyze the data.  

    Interpreting the Results of the Attribute Agreement Analysis

    Minitab's Assistant generates four reports as part of its attribute agreement analysis. The first is a summary report, shown below: 

    attribute agreement analysis summary report

    The green bar at top left of the report indicates that overall, the error rate of the application reviewers is 15.8%. That's not as bad as it could be, but it certainly indicates that there's room for improvement! The report also shows that 13% of the time, the reviewers rejected applications that should be accepted, and they accepted applications that should be rejected 18% of the time. In addition, the reviewers rated the same item two different ways almost 22% of the time.

    The bar graph in the lower left indicates that Javier and Julia have the lowest accuracy percentages among the reviewers at 71.7% and 78.3%, respectively. Jim has the highest accuracy, with 96%, followed by Jill at 90%.

    The second report from the Assistant, shown below, provides a graphic summary of the accuracy rates for the analysis.

    attribute agreement analysis accuracy report

    This report illustrates the 95% confidence intervals for each reviewer in the top left, and further breaks them down by standard (accept or reject) in the graphs on the right side of the report. Intervals that don't overlap are likely to be different. We can see that overall, Javier and Jim have different overall accuracy percentages. In addition, Javier and Jim have different accuracy percentages when it comes to assessing those applications that should be rejected. However, most of the other confidence intervals overlap, suggesting that the reviewers share similar abilities. Javier clearly has the most room for improvement, but none of the reviewers are performing terribly when compared to the others. 

    The Assistant's third report shows the most frequently misclassified items, and individual reviewers' misclassification rates:

    attribute agreement analysis misclassification report

    This report shows that App 9 gave the reviewers the most difficulty, as it was misclassified almost 80% of the time. (A check of the application revealed that this was indeed a borderline application, so the fact that it proved challenging is not surprising.) Among the reject applications that were mistakenly accepted, App 5 was misclassified about half of the time. 

    The individual appraiser misclassification graphs show that Javier and Julia both misclassified acceptable applications as rejects about 20% of the time, but Javier accepted "reject" applications nearly 40% of the time, compared to roughly 20% for Julia. However, Julia rated items both ways nearly 40% of the time, compared to 30% for Javier. 

    The last item produced as part of the Assistant's analysis is the report card:

    attribute agreement analysis report card

    This report card provides general information about the analysis, including how accuracy percentages are calculated. It also can alert you to potential problems with your analysis (for instance, if there were an imbalance in the amount of acceptable to rejectable items being evaluated); in this case, there are no alerts we need to be concerned about. 

    Moving Forward from the Attribute Agreement Analysis

    The results of this attribute agreement analysis give the bank a clear indication of how the reviewers can improve their overall accuracy. Based on the results, the loan department provided additional training for Javier and Julia (who also were the least experienced reviewers on the team), and also conducted a general review session for all of the reviewers to refresh their understanding about which factors on an application were most important. 

    However, training may not always solve problems with inconsistent assessments. In many cases, the criteria on which decisions should be based are either unclear or nonexistent. "Use your common sense" is not a defined guideline! In this case, the loan officers decided to create very specific checklists that the reviewers could refer to when they encountered borderline cases. 

    After the additional training sessions were complete and the new tools were implemented, the bank conducted a second attribute agreement analysis, which verified improvements in the reviewers' accuracy. 

    If your organization is challenged by honest disagreements over "judgment calls," an attribute agreement analysis may be just the tool you need to get everyone back on the same page. 

    Five Ways to Make Your Control Charts More Effective

    $
    0
    0

    Have you ever wished your control charts were better?  More effective and user-friendly?  Easier to understand and act on?  In this post, I'll share some simple ways to make SPC monitoring more effective in Minitab.

    Common Problems with SPC Control Charts

    manufacturing line SPCI worked for several years in a large manufacturing plant in which control charts played a very important role. Virtually thousands of SPC (Statistical Process Control) charts were used to monitor processes, contamination in clean rooms, monitor product thicknesses and shapes as well as critical equipment process parameters. Process engineers regularly checked the control charts of the processes they were responsible for. Operators were expected to stop using equipment as soon as an out of control alert appeared and report this incident back to their team leader.

    But some of the problems we faced had little to do with statistics. For example, comments entered by the operators were often not explicit at all. Control chart limits were not updated regularly and were sometimes not appropriate due to process changes in time. Also, there was confusion about the difference between control limits and specification limits, so even when drifts from the target were clearly identifiable, some process engineers were reluctant to take action as long as their data remained within specifications.

    Other problems could be solved with a better knowledge of statistics. For example, some processes were cyclical in nature, and therefore the way subgroups were defined was critical. Also, since the production was based on small batches of similar parts, the within-batch variability was often much smaller than the between-batch variability (simply because the parts within a batch had been processed in very similar conditions). This lead to inappropriate control limits when standard X-bar control charts were used.

    Red chart

    Five Ways to Make SPC Monitoring Control Charts More Effective

    Let's look at some simple ways to make SPC monitoring more effective in Minitab. In addition to creating standard control charts, you can use Minitab to:

    1. Import data quickly to identify drifts as soon as possible.
    2. Create Pareto charts to prevent special causes from reoccurring.
    3. Account for atypical periods to avoid inflating your control limits.
    4. Visually manage SPC alerts to quickly identify the out-of-control points.
    5. Choose the right type of charts for your process.
    1. Identify drifts as soon as possible.

    To ensure that your control charts are up to date in Minitab, you can right click on them and choose “Automatically update Graphs.” However, Minitab is not always available on the shop floor, so the input data often must be saved in an Excel file or in a database.

    Suppose that the measurement system generates an XML, Excel or text file, and that this data needs to be reconfigured and manipulated in order to be processed in an SPC chart in Minitab. You can automate these using a Minitab macro.

    This macro might automatically retrieve data from an XML or a Text file or from a database (using Minitab's ODBC “Open Data Base Connectivity” functionality) into a Minitab worksheet, or transpose rows into columns, stack columns, or merge several files into one etc. This macro would enable you to obtain a continuously updated Minitab worksheet -- and consequently a continuously updated control chart.

    You could easily launch the macro just by clicking on a customized icon or menu in Minitab (see the graph below) in order to update the resulting control chart.

    SPC Tool Bar

    Alternatively, if the macro is named Startup.mac, it will launch whenever you launch Minitab. If you're using Minitab to enable process operators or engineers to monitor control charts, you could also customize Minitab's toolbars and icons in order to show only the relevant toolbars and icons and focus on SPC.

    The product support section of our website has information on adding a button to a menu or toolbar that will update data from a file or a database.

    2. Create Pareto charts to prevent special causes from reoccurring.

    Statistical Process Control may be used to identify the true root causes (the so-called special causes) of quality problems from the surrounding process noise (the so-called common causes). The root causes of quality issues need to be truly understood in order to prevent reoccurrence.

    A Pareto chart of the causes for out-of-control points might be very useful to identify which special causes occur most frequently.

    Comments can be entered in a column of the Minitab worksheet for each out-of-control point. These comments should be standardized for each type of problem. A list of keywords displayed in the Minitab worksheet would help operators enter meaningful keywords, instead of comments that differ each time. Then a Pareto chart could be used to identify the 20% causes that generate 80% of your problems, based on the (standardized) comments entered in the worksheet.

    Pareto

    Comments can even be displayed in the SPC chart by using the annotation toolbar.  Click on the T (text) icon of the Graph Annotation toolbar.

    3. Account for atypical periods to avoid inflating your control limits.

    Atypical periods (due to measurement issues, outliers, or a quality crisis) may artificially inflate your control chart limits. In Minitab, control limits may be calculated according to a reference period (one with standard, stable /predictable behavior), or the atypical period may be omitted so that control limits are not affected.

    In Minitab, go to Options in the control chart dialogue box, look for the Estimate Tab and select the subgroups to be omitted (untypical behavior, outliers), or use only some specified sub-groups to set reference periods. Although the atypical period will still get displayed on the control chart, it won't affect the way your control limits are estimated.

    Untypical

    If a reference period has been selected, you will probably need to update it after a certain period of time to ensure that this selection is still relevant.

    4) Visually manage SPC alerts to quickly identify out-of-control points.

    If the number of control charts you deal with is very large, and you need to quickly identify processes that are drifting away from the target, your could display all control charts in a Tile format (go to Window > Tile). When the latest data (i.e., the last row of the worksheet) generates an out-of-control warning, you can have the control chart become completely red, as shown in the picture below:

    Red chart

    You can do this by going to Tools > Options. Select “Control Charts and Quality Tools” on the list, then choose Other. Under the words “When last row of data causes a new test failure for any point,” check the box that says "Change color of chart." Note that the color will change according to the last row (latest single value) not according to the latest subgroup, so this option is more effective when collecting individual values.

    5. Choose the right type of charts for your process.

    When it comes to control charts, one size does not fit all. That's why you'll see a wide array of options when you select Stat > Control Charts. Be sure that you're matching the control chart you're using to the type of data and information you want to monitor. For example, if your subgroups are based on batches of products, I-MR-R/S (within/between) charts are probably best suited to monitor your process.

    If you're not sure which control chart to use, you can get details about each type from the Help menu in Minitab, or try using the Assistant menu to direct you to the best test for your situation.

     

     

     

    Statistical Tools for Process Validation, Stage 2: Process Qualification

    $
    0
    0

    In its industry guidance to companies that manufacture drugs and biological products for people and animals, the Food and Drug Administration (FDA) recommends three stages for process validation.Process Validation Stages While my last post covered statistical tools for the Process Design stage, here we will focus on the statistical techniques typically utilized for the second stage, Process Qualification.

    Stage 2: Process Qualification

    During this stage, the process design is evaluated to determine if it is capable of reproducible commercial manufacture. Successful completion of Stage 2 is necessary before commercial distribution.

    Example: Evaluate Acceptance Criteria with Capability Analysis

    Suppose the active ingredient amount in a tranquilizer needs to be between 360 and 370 mg/mL and you need to assess the quality level, where a minimum Cpk of 1.33 is defined as the acceptance criteria. To assess process performance and determine if measurements are within specification, you can use capability analysis, available in Minitab Statistical Software.

    Five samples are randomly selected from 50 batches and the amount of active ingredient is measured. The data is then analyzed relative to the 360 mg/mL minimum and 370 mg/mL maximum.

    Process Capability

    The capability analysis reveals a Cpk of 0.53, which fails to meet the acceptance criteria of 1.33. The active ingredient amounts for this tranquilizer are not acceptable. So how can we improve it? The Cp value of 1.41 and the graph both reveal that, although the variability is acceptable with respect to the width of the specification limits, the process average needs to be shifted to a higher mg/mL in order to achieve an acceptable Cpk.

    Example: Conduct Variation Analysis across Batches

    Suppose we want to assess content uniformity, a critical quality characteristic, across 3 batches at 10 locations. To visualize the intra-batch (within-batch) variation and the inter-batch (between-batch) variation, we can create boxplots for each batch.

    A boxplot can help us visually assess both the intra- and inter-batch variation, and identify any outliers. This specific graph shows a homogeneous dispersion of measurements both within each batch and between batches. And there are no outliers, which Minitab would flag with an asterisk (*). 

    Boxplot

    Although boxplots are useful tools to conduct a visual assessment, we can also statistically assess if there is a significant difference in the between batch variation using an equal variances test. The test reveals a p-value greater than an alpha-level of 0.05 (or whatever alpha-level you prefer), which supports the conclusion that there is consistency between batches.

    Example: Various Applications for Tolerance Intervals

    Another useful tool for Process Qualification is the tolerance interval. This tool has multiple applications. For example, tolerance intervals can be used to compare your process to specifications, profile the outcome of a process, or establish acceptance criteria.

    For a given product characteristic, a tolerance interval provides a range of values that likely covers a specified proportion of the population (for example, 95%) for a specified confidence level (like 99%).

    For example, suppose we want to know how the active ingredient values in the manufacturing process compare to our specification limits. Based on a dose-response study, the limits are 360 to 370 mg/mL.

    Tolerance Interval

    For this particular data set, Minitab reveals that we can be 99% confident that 95% of the units will be between 362.272 and 367.468 mg/mL. The process bounds therefore indicate that we can meet the requirements of 360 to 370, and we can conclude with high confidence that the process variation is less than the allowable variation, defined by the specification limits.

    Or perhaps we need to assess content uniformity using 99% confidence and 99% coverage. We sample 30 tablets and calculate a tolerance interval, revealing that we can be 99% certain that 99% of the tablets will have a content uniformity within some range, calculated using Minitab.

    And that’s how you can use various statistical tools to support Process Qualification. In the final post in this series, we’ll explore the Continued Process Verification stage!

    How Taguchi Designs Differ from Factorial Designs

    $
    0
    0

    Genichi Taguchi is famous for his pioneering methods of robust quality engineering. One of the major contributions that he made to quality improvement methods is Taguchi designs.Genichi Taguchi

    Designed experiments were first used by agronomists during the last century. This method seemed highly theoretical at first, and was initially restricted to agronomy. Taguchi made the designed experiment approach more accessible to practitioners in the manufacturing industry.

    Thanks partly to him, Design of Experiments (DOE) has become quite popular in many companies, and these methods are widely taught in universities and engineering school. In this blog post, I would like to describe differences between Taguchi DOEs and standard Factorial DOEs.

    Taguchi Designs

    Both Taguchi designs and Factorial designs are are available in the DOE menu in Minitab Statistical Software. To select a design go to Stat > DOE.

    Many Taguchi designs are based on Factorial designs (2-level designs and Plackett & Burman designs, as well as factorial designs with more than 2 levels). Taguchi’s L8 design, for example, is actually a standard 23 (8-run) factorial design.

    Taguchi's designs are usually highly fractionated, which makes them very attractive to practitioners. Doing a half-fraction, quarter-fraction or eighth-fraction of a full factorial design greatly reduces costs and time needed for a designed experiment.

    The drawback of a fractionated design is that some interactions may be confounded with other effects. It is important to consider carefully the role of potential confounders and aliases. Failure to take account of such confounded effects can result in erroneous conclusions and misunderstandings.

    When using a Taguchi design, one needs to guess which interactions are most likely to be significant—even before any experiment is performed. Taguchi created several linear graphs to help practitioners select the interactions they want to study, based on their prior process knowledge.

    Example from a two-level, eight-factor L16 Taguchi design:

    Linear graphs are not displayed in Minitab, but factor allocation and interaction selection are based on Taguchi linear graphs. Suppose that factor A is allocated to column 1 of the orthogonal array, factor B to column 2, C to column 4, D to column 8, E to column 7, F to column 11, G to column 13, and H to column 14 (as described in the Minitab dialog box above and this matches with the corresponding Taguchi linear graph below). With this design, one may select the AB, AC, AD, AE, AF, AG, AH interactions. It is not possible to analyze the remaining interactions, since they are confounded with the selected interactions.

    Taguchi suggested several other linear graphs for an L16 design (a 16-run factorial design):

    Standard Fractional Factorial Designs

    In a standard factorial (non-Taguchi) design, identifying the interactions most likely to be significant is based on alias / confounding "chains." The same alias chains apply to Taguchi designs, but are not displayed. Practitioners may not necessarily be aware that some interaction effects are confounded. However, when you use the factorial design functionality in Minitab the alias chains are clearly displayed showing the confounding pattern:

    AB + CG + DH + EF

    AC + BG + DF + EH

    AD + BH + CF + EG

    AE + BF + CH + DG

    AF + BE + CD + GH

    AG + BC + DE + FH

    AH + BD + CE + FG

    Confounding patterns are a lot more complex for 3-level and 4-level designs.

    In the factorial design menu, the diagram below displays the designs that are available and their resolution (level of confounding). In Minitab, you can quickly access the table of factorial designs shown below by selecting Stat > DOE > Factorial > Create Factorial Design... and clicking "Display Available Designs."

    table of available designs

    Red (Resolution III) designs should be avoided (because main effects are confounded with two-factor interactions). In experiments that use Yellow (Resolution IV) designs, two-factor interactions are confounded with other two-factor interactions. These popular designs provide a good compromise between the amount of information obtained and costs (number of experimental runs). The higher resolution designs (in green) offer high quality—with limited or no confounding—at higher costs.

    The Pareto and “Heredity” Principles

    In a Resolution IV (yellow region) design, main effects are not confounded with two-factor interactions. Often, after an experiment has been performed, the experimenter discovers that only a few of the many effects investigated turn out to be important (the “Pareto rule”).

    When two interactions are confounded with one another, the interaction that is the most likely to be significant is the one containing factors whose main effects are themselves significant (based on the so called “heredity” or “hierarchy” principle). These principles are extremely useful to identify the interactions most likely to be important. We can expect only a few effects to be statistically significant, and we can focus on the interactions containing factors whose main effects are themselves significant.

    Two Approaches to Selecting Which Interactions Are Important

    Taguchi designs are based on prior selection of the most likely interactions, whereas in standard fractional factorial designs, the interactions are selected later on, after the initial results from the designed experiments have been analyzed. The way in which interactions are selected clearly differs between the two approaches.

    According to Taguchi, optimizing a process is not sufficient: making processes and products more robust to quality issues and environmental noises is crucial. In this strategy, designed experiments clearly play a central role.

    How to Compute Probabilities

    $
    0
    0

    Have you ever wanted to know the odds of something happening, or not happening? 

    It's the kind of question that students are frequently asked to calculate by hand in introductory statistics classes, and going through that exercise is a good way to become familiar with the mathematical formulas the underlie probability (and hence, all of statistics). 

    But let's be honest: when class is over, most people don't take the time to calculate those probabilities—at least, not by hand. Some people even resort to "just making it up." Needless to say, we at Minitab are firmly opposed to just making it up.

    The good news is that determining the real odds of something happening doesn't have to be hard work! If you don't want to calculate the probabilities by hand, just let a statistical software package such as Minitab do it for you. 

    Computing Binomial Probabilities

    Let's look at how to compute binomial probabilities. The process we'll go through is similar for any of the 24 distributions Minitab includes.

    We use the binomial distribution to characterize a process with two outcomes—for example, if a part passes or fails inspection, if a candidate wins or loses an election, or if a coin lands on heads or tails. This distribution is used frequently in quality control, opinion surveys, medical research, and insurance.coin flip

    Suppose I want to know the probability of getting a certain number of heads in 10 tosses of a fair coin. I need to calculate the odds for a binomial distribution with 10 trials (n=10) and probability of success p=0.5.

    To compute the probability of exactly 8 successes, select Calc > Probability Distributions > Binomial...

    binomial distribution

    Choose “probability” in the dialog, then enter the number of trials (10) and the probability of success (0.5) for “event probability." If we wanted to calculate the odds for more than one number of events, we could enter them in a worksheet column. But since for now we just want the probability of getting exactly 8 heads in 10 tosses, choose the "Input Constant" option, enter 8, and press OK. 

    binomial probability

    The following output appears in the session window. It tells us that if we toss a fair coin with an 50% probability of landing on heads, the odds of getting exactly 8 heads out of 10 tosses are just 4%.

    binomial probability out

    What if we wanted to know the cumulative probability of getting 8 heads in 10 tosses? Cumulative probability is the odds of one, two, or more events taking place. The word to remember is "or," because that's what cumulative probability tells you. What are the chances that when you toss this coin 10 times, you'll get 8 or fewer heads? That's cumulative probability.

    To compute cumulative probabilities, select “cumulative probability” in the binomial distribution dialog.

    binomial cumulative probability dialog

    The probability of 8 or fewer successes, is P(X ≤ 8) = 0.989258, or 98%:

    binomial cumulative probability output

    Creating a Table of Probabilities

    We can also use Minitab to calculate a full table of probabilities. In the worksheet, enter all of the values of the number of successes in a column. For example, for a series of 10 tosses, you would enter 1, 2, 3, 4, 5, 6, 7, 8, 9, 10. Next we'll select Calc > Probability Distributions > Binomial... again, but this time choose “Input column” and select C1 instead of using the "Input constant." Specify a different column for storage and press OK.

    binomial distribution probability table dialog

    The probabilities appear in column C2:

    binomial distribution probability table output

    Visualizing the Probabilities

    Suppose you want to see the distribution of these probabilities in a graph? Select Graph > Bar Charts..., then use the dialog box choose View Single. 

    bar chart selection dialog box

    Just complete the dialog as shown:

    bar chart creation dialog

    When you press OK, Minitab produces this bar chart: 

    bar chart of binomial probabilities

    If you need to know the precise value for a given number of events, just hover over that column and Minitab displays the details:

    edit graph dialog

    As you can see, using Minitab to check and graph the probabilities of different events is not difficult. I hope knowing this increases the odds that the next time you wonder about the likelihood of an event, you'll be able to find it quickly and accurately!

    A Field Guide to Statistical Distributions

    $
    0
    0

    by Matthew Barsalou, guest blogger. 

    The old saying “if it walks like a duck, quacks like a duck and looks like a duck, then it must be a duck” may be appropriate in bird watching; however, the same idea can’t be applied when observing a statistical distribution. The dedicated ornithologist is often armed with binoculars and a field guide to the local birds and this should be sufficient. A statologist (I just made the word up, feel free to use it) on the other hand, is ill-equipped for the visual identification of his or her targets.

    Normal, Student's t, Chi-Square, and F Distributions

    Notice the upper two distributions in figure 1. The normal distribution and student’s t distribution may appear similar. However, the standard normal distribution is calculated using n and student’s t distribution is calculated using n-1. This may appear to be a minor difference, but when n is small, student’s t distribution displays much more peakedness. Student’s t distribution approaches the normal distribution as the sample size increases, but it never truly matches the shape of the normal distribution.

    Observe the Chi-square and F distribution in the lower half of figure 1. The shapes of the distributions can vary and even the most astute observer will not be able to differentiate between them by eye. Many distributions can be sneaky like that. It is a part of their nature that we must accept as we can’t change it.

    Distribution Field Guide Figure 1Figure 1

    Binomial, Hypergeometric, Poisson, and Laplace Distributions

    Notice the distributions illustrated in figure 2. A bird watcher may suddenly encounter four birds sitting in a tree; a quick check of a reference book may help to determine that they are all of a different species. The same can’t always be said for statistical distributions. Observe the binomial distribution, hypergeometric distribution and Poisson distribution. We can’t even be sure the three are not the same distribution. If they are together with a Laplace distribution, an observer may conclude “one of these does not appear to be the same as the others.” But they are all different, which our eyes alone may fail to tell us.

    Distribution Field Guide Figure 2Figure 2

    Weibull, Cauchy, Loglogistic, and Logistic Distributions

    Suppose we observe the four distributions in figure 3.What are they? Could you tell if they were not labeled? We must identify them correctly before we can do anything with them. One is a Weibull distribution, but all four could conceivably be various Weibull distributions. The shape of the Weibull distribution varies based upon the shape parameter (κ) and scale parameter (λ).The Weibull distribution is a useful, but potentially devious distribution that can be much like the double-barred finch, which may be mistaken for an owl upon first glance.

    Distribution Field Guide Figure 3Figure 3

    Attempting to visually identify a statistical distribution can be very risky. Many distributions such as the Chi-Square and F distribution change shape drastically based on the number of degrees of freedom. Figure 4 shows various shapes for the Chi-Square, F distribution and the Weibull distribution. Figure 4 also compares a standard normal distribution with a standard deviation of one to a t distribution with 27 degrees of freedom; notices how the shapes overlap to the point where it is no longer possible to tell the two distributions apart.

    Although there is no definitive Field Guide to Statistical Distributions to guide us, there are formulas available to correctly identify statistical distributions. We can also use Minitab Statistical Software to identify our distribution.

    Distribution Field Guide Figure 4Figure 4

    Go to Stat > Quality Tools > Individual Distribution Identification... and enter the column containing the data and the subgroup size. The results can be observed in either the session window (figure 5) or the graphical outputs shown in figures 6 through 9.

    In this case, we can conclude we are observing a 3-parameter Weibull distribution based on the p value of 0.364.

    Distribution Field Guide Figure 5

    Figure 5

     

    Distribution Field Guide Figure 6Figure 6

    Distribution Field Guide Figure 7Figure 7

    Distribution Field Guide Figure 8Figure 8

    Distribution Field Guide Figure Figure 9

     

     

     

    About the Guest Blogger

    Matthew Barsalou is a statistical problem resolution Master Black Belt at BorgWarner Turbo Systems Engineering GmbH. He is a Smarter Solutions certified Lean Six Sigma Master Black Belt, ASQ-certified Six Sigma Black Belt, quality engineer, and quality technician, and a TÜV-certified quality manager, quality management representative, and auditor. He has a bachelor of science in industrial sciences, a master of liberal studies with emphasis in international business, and has a master of science in business administration and engineering from the Wilhelm Büchner Hochschule in Darmstadt, Germany. He is author of the books Root Cause Analysis: A Step-By-Step Guide to Using the Right Tool at the Right TimeStatistics for Six Sigma Black Belts and The ASQ Pocket Guide to Statistics for Six Sigma Black Belts.

     


    3 Things a Histogram Can Tell You

    $
    0
    0

    Histograms are one of the most common graphs used to display numeric data. Anyone who takes a statistics course is likely to learn about the histogram, and for good reason: histograms are easy to understand and can instantly tell you a lot about your data.

    Here are three of the most important things you can learn by looking at a histogram. 

    Shape—Mirror, Mirror, On the Wall…

    If the left side of a histogram resembles a mirror image of the right side, then the data are said to be symmetric. In this case, the mean (or average) is a good approximation for the center of the data. And we can therefore safely utilize statistical tools that use the mean to analyze our data, such as t-tests.

    If the data are not symmetric, then the data are either left-skewed or right-skewed. If the data are skewed, then the mean may not provide a good estimate for the center of the data and represent where most of the data fall. In this case, you should consider using the median to evaluate the center of the data, rather than the mean.

    Did you know...

    If the data are left-skewed, then the mean is typically LESS THAN the median.    

    If the data are right-skewed, then the mean is typically GREATER THAN the median.

      Span—A Little or a Lot?

    Suppose you have a data set that contains the salaries of people who work at your organization. It would be interesting to know where the minimum and maximum values fall, and where you are relative to those values. Because histograms use bins to display data—where a bin represents a given range of values—you can’t see exactly what the specific values are for the minimum and maximum, like you can on an individual value plot. However, you can still observe an approximation for the range and see how spread out the data are. And you can answer questions such as "Is there a little bit of variability in my organization's salaries, or a lot?"

    Outliers (and the ozone layer)

    Outliers can be described as extremely low or high values that do not fall near any other data points. Sometimes outliers represent unusual cases. Other times they represent data entry errors, or perhaps data that does not belong with the other data of interest. Whatever the case may be, outliers can easily be identified using a histogram and should be investigated as they can shed interesting information about your data. 

    Rewind to the mid-1980s when scientists reported depleting ozone levels above Antarctica. The Goddard Space Center had studied atmospheric ozone levels, but surprisingly didn’t discover the issue. Why? The analysis they used automatically eliminated any Dobson readings below 180 units because ozone levels that low were thought to be impossible.

     

    10 Tips to Increase your Minitab Efficiency

    $
    0
    0

    by Rehman Khan, guest blogger

    There are many articles giving Minitab tips already, so to be different I have done mine in the style of my books, which use example-based learning. All ten tips are shown using a single example.

    If you don’t already know these 10 tips you will get much more benefit if you work along with the example. You don’t need to download any files to work along—although, if you don’t have Minitab already, you may want to download the free 30-day trial

    First I will list my 10 tips, and then as we go through the examples I will highlight where they are going to be used. The 10 tips are

    1. Using the Auto-Fill function.
    2. Making patterned text data.
    3. Using Set Base when generating random data.
    4. Using the Edit Last Dialog Box function.
    5. Clearing a menu.
    6. Setting the order of a categorical axis on a graph.
    7. Updating a graph.
    8. Making a Similar Graph, which is especially useful when formatting has been changed.
    9. Using the Layout Tool.
    10. Using Conditional Formatting.

    We are going to generate 3 columns of data which will have 30 rows in each column. Column 1 is called Shift, it relates to a production process where there is a Morning, Afternoon and Night shift. Columns C2 & C3 are two yield values, which are recorded for each shift. Remember this blog is about learning 10 useful Minitab tips rather than learning to examine the data for this process.

    First, start a new Minitab project and type in the column headings shown. Then type ‘Morning’, ‘Afternoon’ and then ‘Night’ in successive cells after shift.

    3 columns of data

    Tip 1: Using the Autofill function

    We could copy-and-paste the first three cells 9 times to make our 30 rows of data. But instead, we can highlight the three cells and then grab the Fill Handle and drag that down to Auto Fill our text. Try that now but don’t go too far down.

    autofill

    Tip 2: Making Patterned Text data

    Another way of getting Minitab to do all the laborious typing is to use the Make Patterned Data command. Select Calc > Make Patterned Data> Text Values... Complete the menu as shown and click OK.

    patterned data dialog 

    Now we will randomly generate two columns of yield data to simulate the performance of the shifts. However, random data is not as random as you might think.

    Tip 3: Using Set Base when generating random data.

    set baserows of data to generate

    The Set Base command fixes the starting point of Minitab’s Random number generator. So even though we are generating random data on different machines at different times, using the same starting point ensures that Minitab will give you the same random data that it gives me.  Select Calc > Set Base... Then enter 3 as the Base for the random number generator and click OK.

    We are going to use a Uniform distribution for Yield1. Go to Calc > Random Data > Uniform… then complete the menu as shown and then click OK. Because we used the Set Base command, all of our randomly generated data will be the same!

    Tip 4: Using the Edit Last Dialog Box function

    Edit last dialog - CTRL+E

    This is the must-know tip for Minitab. To quickly navigate to the last dialog box you had open, press the control key and then press ‘e’, written as ‘ctrl+e’. Alternatively, press the edit last dialog icon in the tool bar as shown. This should have re-opened the dialog box we used to generate 30 rows of random data from the uniform distribution.

    Tip 5: Clearing a dialog      

    To completely clear a dialog box, just press F3. This is very useful, since it will clear sub-dialogs as well. Press F3 to clear the menu now. Complete the menu as you did to create the Yield1 data, but this time store the data in column Yield2. Your data should look like the screenshot shown.

    randomly generated data

    bar graph

    For the next part of the demonstration we need to produce a graph. Go to Graph > Bar Chart...

    From the Bar Represents drop-down menu, select ‘A Function of a Variable.’ Ensure that One Y Simple is selected and then click OK.  Complete the dialog box as the screenshot shows. Then click OK to produce the chart.

    Tip 6: Setting the order of a categorical axis on a graph

    Notice that the chart’s X-axis labels are in alphabetical order. This is the default for Minitab but we sometimes need to change the order to be user-friendly.

    12 Tip6 Value order 

    To change the order of the x-axis variables,go to the worksheet and 

    place the active cursor anywhere in column C1. Right-click and then select Column Properties > Value Order… There are a number of options available, but we will set the radio button for Value Order to ‘Order of occurrence in the worksheet’ as shown.

    value order for C1

         

     

      Tip 7: Updating a graph.

    We can now either recreate our graph—or we can update it. Look in the top-left corner of the graph. The yellow warning triangle and circular blue arrows mean that the graph is out of sync with the data in the worksheet. To update the graph, right-click on it and then select Update Graph Now. You also have the option of keeping the graph automatically updated. Note: some graphs cannot be updated, they must be recreated when data are changed.

    update graph

    formatted graph

    Tip 8: Making a Similar Graph, which is especially useful when formatting has been changed.

    17 Tip8 Make Similar Graph

    The next graph shown is the same as the one you should have open, but I have changed the format to meet a fictitious company standard. If I need to make similar graphs and not repeat the formatting adjustments every time, there is a shortcut for doing this. First, ensure the graph is selected by left clicking on it. Then go to Editor > Make Similar Graph…  

         

     

     

    The dialog allows basic changes to the graph. Change ‘Yield1’ to ‘Yield2’ in the new variable column. Then Click OK to produce the similar graph for column Yield2.

    Similar Graph

    arrange graphs dialog

    Tip 9: Using the Layout Tool

    We already have two graphs. To demonstrate our next tip, let’s also make two boxplots for Yield1 and Yield2, respectively..

    ordered graphs

    If we want to display these four graphs on the same plot we can use the Layout Tool to make a multi-graph plot using existing graphs. Select any graph and then go to Editor > Layout Tool…

    On the top-left of the layout tool we can change how many plots are shown, we will use a 2x2 layout but Minitab can go up to 9x9.

    The Layout Tool is an easy to use, but is best learned through a bit of experimentation. Try arranging and ordering the graphs in different ways. When you are done click on the Finish button.

    Note that you can change the formatting of the new plot can by just as you can adjust other graphs.

    21 Tip10 Conditional Formatting

    Tip 10: Using Conditional Formatting

    For the final tip, I am going to give you a brief introduction to the conditional formatting tools added in Minitab 17.

    conditional formatting

    If I wanted to quickly identify my best performers—which we’ll define as those with a Yield1 greater than 95—in the Project Window, I can use Conditional Formatting. Go to Data > Conditional Formatting > Highlight Cell > Greater Than...

    Complete the dialog as shown, then click OK. You’ll see the results in the Project Window. Conditional Formatting is very useful when sanitizing large data files.

    I hope you have enjoyed learning about my 10 favorite Minitab tricks, and that you find them helpful the next time you’re analyzing your own data!

    About the Guest Blogger…

    Rehman Khan is the author of Six Sigma Statistics using Minitab 17 and also Problem Solving and Data Analysis using Minitab.  Recently he has started his own Youtube channel called RMK Six Sigma. Rehman is a SigmaPro Master Black Belt and Charted Chemical Engineer. He works for FMC Chemicals Ltd in the UK as a Manufacturing Excellence Engineer.

     

    Chi-Square Analysis: Powerful, Versatile, Statistically Objective

    $
    0
    0

    To make objective decisions about the processes that are critical to your organization, you often need to examine categorical data. You may know how to use a t-test or ANOVA when you’re comparing measurement data (like weight, length, revenue, and so on), but do you know how to compare attribute or counts data? It easy to do with statistical software like Minitab. 

    failures per production line

    One person may look at this bar chart and decide that the production lines performed similarly. But another person may focus on the small difference between the bars and decide that one of the lines has outperformed the others. Without an appropriate statistical analysis, how can you know which person is right?

    When time, money, and quality depend on your answers, you can’t rely on subjective visual assessments alone. To answer questions like these with statistical objectivity, you can use a Chi-Square analysis.

    Which Analysis Is Right for Me?

    Minitab offers three Chi-Square tests. The appropriate analysis depends on the number of variables that you want to examine. And for all three options, the data can be formatted either as raw data or summarized counts.

    Chi-Square Goodness-of-Fit Test – 1 Variable

    Use Minitab’s Stat > Tables > Chi-Square Goodness-of-Fit Test (One Variable) when you have just one variable.

    The Chi-Square Goodness-of-Fit Test can test if the proportions for all groups are equal. It can also be used to test if the proportions for groups are equal to specific values. For example:

    • A bottle cap manufacturer operates three production lines and records the number of defective caps for each line. The manufacturer uses the Chi-Square Goodness-of-Fit Test to determine if the proportion of defectives is equal across all three lines.
    • A bottle cap manufacturer operates three production lines and records the number of defective caps and the total number produced for each line. One line runs at high speed and produces twice as many caps as the other two lines that run at a slower speed. The manufacturer uses the Chi-Square Goodness-of-Fit Test to determine if the number of defective units for each line is proportional to the volume of caps it produces.
    Chi-Square Test for Association – 2 Variables

    Use Minitab’s Stat > Tables > Chi-Square Test for Association when you have two variables.

    The Chi-Square Test for Association can tell you if there’s an association between two variables. In another words, it can test if two variables are independent or not. For example:

    • A paint manufacturer operates two production lines across three shifts and records the number of defective units per line per shift. The manufacturer uses the Chi-Square Goodness-of-Fit Test to determine if the percent defective is similar across all shifts and production lines. Or, are certain lines during certain shifts more prone to issues?

      Defectives per line per shift

       
    • A call center randomly samples 100 incoming calls each day of the week for each of its three locations, for a total of 1500 calls. They then record the number of abandoned calls per location per day. The call center uses a Chi-Square Test to determine if there are is any association between location and day of the week with respect to missed calls.

    call center data
     

    Cross Tabulation and Chi-Square – 2 or more variables

    Use Minitab’s Stat > Tables > Cross Tabulation and Chi-Square when you have two or more variables.

    If you simply want to test for associations between two variables, you can use either Cross Tabulation and Chi-Square or Chi-Square Test for Association. However, Cross Tabulation and Chi-Square also lets you control for the effect of additional variables. Here’s an example:

    • A tire manufacturer records the number of failed tires for four different tire sizes across two production lines and three shifts. The plant uses a Cross Tabulation and Chi-Square analysis to look for failure dependencies between the tire sizes and production lines, while controlling for any shift effect. Perhaps a particular production line for a certain tire size is more prone to failures, but only during the first shift.

    This analysis also offers advanced options. For example, if your categories are ordinal (good, better, best or small, medium, large) you can include a special test for concordance.

    Conducting a Chi-Square Analysis in Minitab

    Each of these analyses is easy to run in Minitab. For more examples that include step-by-step instructions, just navigate to the Chi-Square menu of your choice and then click Help > example.

    It can be tempting to make subjective assessments about a given set of data, their makeup, and possible interdependencies, but why risk an error in judgment when you can be sure with a Chi-Square test?

    Whether you’re interested in one variable, two variables, or more, a Chi-Square analysis can help you make a clear, statistically sound assessment.

    Using Designed Experiments (DOE) to Minimize Moisture Loss

    $
    0
    0

    cake!

    As a person who loves baking (and eating) cakes, I find it bothersome to go through all the effort of baking a cake when the end result is too dry for my taste. For that reason, I decided to use a designed experiment in Minitab to help me reduce the moisture loss in baked chocolate cakes, and find the optimal settings of my input factors to produce a moist baked chocolate cake. I’ll share the details of the design and the results in this post.

    Choosing Input Factors for the Designed Experiment

    Because I like to use premixed chocolate cake mixes, I decided to use two of my favorite cake mix brands for the experiment. For the purpose of this post, I’ll call the brands A and B. Thinking about what could impact the loss of moisture, it is likely that the baking time and the oven temperature will affect the results. Therefore, the factors or inputs that I decided to use for the experiment are:

    1. Cake mix brand: A or B (categorical data)
    2. Oven temperature: 350 or 380 degrees Fahrenheit (continuous data)
    3. Baking time:  38 or 46 minutes (continuous data)
    Measuring the Response

    Next, I needed a way to measure the moisture loss. For this experiment, I used an electronic food scale to weigh each cake (in the same baking pan) before and after baking, and then used those weights in conjunction with the formula below to calculate the percent of moisture lost for each cake:

    % Moisture Loss = 100 x initial weight – final weight
                                            initial weight

    Designing the Experiment

    For this experiment, I decided to construct a 23 full factorial design with center points to detect any possible curvature in the response surface. Since the cake mix brand is categorical and therefore has no center point between brand A and brand B, the number of center points will be doubled for that factor. Because of this, I’d have to bake 10 cakes which, even for me, is too many in a single day. Therefore, I decided to run the experiment over two days. Because differences between the days on which the data was collected could potentially introduce additional variation, I decided to add a block to the design to account for any potential variation due to the day. 

    To create my design in Minitab, I use Stat> DOE> Factorial > Create Factorial Design:

    select create factorial design

    Minitab 17 makes it easy to enter the details of the design.  First, I selected 3 as the number of factors:

    select three factors

    Next, I clicked on the Designs button above. In the Designs window, I can tell Minitab what type of design I’d like to use with my 3 factors:

    select type of design

    In the window above, I’ve selected a full 23 design, and also added 2 blocks (to account for variation between days), and 1 center point per block. After making the selections and clicking OK in the above window, I clicked on the Factors button in the main window to enter the details about each of my factors:

    factors

    Because center points are doubled for categorical factors, and because this design has two blocks, the final design will have a total of 4 center points.  After clicking OK in the window above, I ended up with the design shown below with 12 runs:

    design data

    Performing the Experiment and Analyzing the Data

    After spending an entire weekend baking cakes and calculating the moisture loss for each one, I entered the data into Minitab for the analysis. I also brought in a lot of cake to share with my colleagues at Minitab!

    data

    With the moisture loss for each of my 12 cakes recorded in column C8 in the experiment worksheet, I’m ready to analyze the results. 

    In Minitab, I used Stat> DOE> Factorial > Analyze Factorial Design... and then entered the Moisture Loss column in the Responses field:

    Analyze factorial DOE

    In the window above, I also clicked on Terms to make sure I’m only including the main effects and two-way interactions. After clicking OK in each window, Minitab produced a Pareto chart of the standardized effects that I could use to reduce my model:

    pareto of standardized effects

    I can see from the above graph that the main effects (A, B and C) all significantly impact the moisture of the cake, since the bars that represent those terms on the graph extend beyond the red vertical reference line.  All of the two-way interactions (AB, AC and BC) are not significant.

    I can also see the same information in the ANOVA table in Minitab’s session window:

    ANOVA results

    In the above ANOVA table, we can see that the cake mix brand, oven temp, and baking time are all significant since their p-values are lower than my alpha of 0.05. 

    We can also see that all of the 2-way interactions have p-values higher than 0.05, so I’ll conclude that those interactions are not significant and should be removed from the model.

    Interestingly, the p-value for the blocksis significant (with a p-value of 0.01).  This indicates that there was indeed a difference between the two days in which the data was collected which impacted the results. I'm glad I accounted for that additional variation by including a block in my design!

    Analyzing the Reduced Model

    To analyze my reduced model, I can go back to Stat> DOE> Factorial > Analyze Factorial Design. This time when I click the Terms button I’ll keep only the main effects, and remove the two-way interactions. Minitab displays the following ANOVA table for the reduced model:

    ANOVA for reduced model

    The table shows that all the terms I’ve included (mix brand, oven temp, and baking time) are significant since all the p-values for these terms are lower than 0.05. We can also see that the test for curvature based on the center points is not significant (p-value = 0.587), so we can conclude that the relationship between the three factors and moisture loss is linear.

    The r-squared, r-squared adjusted, and r-squared predicted are all quite high, so this model seems to be a very good fit to the data.

    Checking the Residuals

    Now I can take a look at the residual plots to make sure all the model assumptions for my model have been met:

    residual plots

    The residuals in the graph above appear to be normally distributed. The residuals versus fits graph appears to show the points are randomly scattered above and below 0 (which indicates constant variance), and the residuals versus order graph doesn’t suggest any patterns that could be due to the order in which the data was collected. 

    Now that I'm confident the assumptions for the model have been met, I’ll use this model to determine the optimal settings of my factors so that going forward all the cakes I make will be moist and fabulous! 

    Optimizing the Response

    I can use Minitab’s Response Optimizer and my model to tell me exactly what combination of cake mix brand, oven temperature, and baking time I’ll want to use to get the moistest cake. I select Stat> DOE> Factorial> Response Optimizer:

    response optimizer dialog 

    In the above window, I can tell Minitab what my goal is. In this case, I want to know what input settings to use so that the moisture loss will be minimized. Therefore, I choose Minimize above and then click OK:

    response optimizer

    In the above graph, the optimal settings for my factors are marked in red near the top. Using the model that I’ve fit to my data, Minitab is telling me that I can use Brand B with an oven temperature of 350 and a baking time of 38 minutes to minimize the moisture loss. Using those values for the inputs, I can expect the moisture loss will be approximately 3.3034, which is quite low compared to the moisture loss for the cakes collected as part of the experiment.

    Success!  Now I can use these optimal settings, and I’ll never waste my time baking a dry cake again.

    If you’ve enjoyed this post about DOE, you may also like to read some of our other DOE blog posts.

    Three Common P-Value Mistakes You'll Never Have to Make

    $
    0
    0

    Statistics can be challenging, especially if you're not analyzing data and interpreting the results every day. Statistical software makes things easier by handling the arduous mathematical work involved in statistics. But ultimately, we're responsible for correctly interpreting and communicating what the results of our analyses show.

    The p-value is probably the most frequently cited statistic. We use p-values to interpret the results of regression analysis, hypothesis tests, and many other methods. Every introductory statistics student and every Lean Six Sigma Green Belt learns about p-values. 

    Yet this common statistic is misinterpreted so often that at least one scientific journal has abandoned its use.

    What Does a P-value Tell You? 

    Typically, a P value is defined as "the probability of observing an effect at least as extreme as the one in your sample data—if the null hypothesis is true." Thus, the only question a p-value can answer is this one:

    How likely is it that I would get the data I have, assuming the null hypothesis is true?

    If your p-value is less than your selected alpha level (typically 0.05), you reject the null hypothesis in favor of the alternative hypothesis. If the p-value is above your alpha value, you fail to reject the null hypothesis. It's important to note that the null hypothesis is never accepted; we can only reject or fail to reject it. 

    The P-Value in a 2-Sample t-Test

    Consider a typical hypothesis test—say, a 2-sample t-test of the mean weight of boxes of cereal filled at different facilities. We collect and weigh 50 boxes from each facility to confirm that the mean weight for each line's boxes is the listed package weight of 14 oz. 

    Our null hypothesis is that the two means are equal. Our alternative hypothesis is that they are not equal. 

    To run this test in Minitab, we enter our data in a worksheet and select Stat > Basic Statistics > 2-Sample T-test. If you'd like to follow along, you can download the data and, if you don't already have it, get the 30-day trial of Minitab. In the t-test dialog box, select Both samples are in one column from the drop-down menu, and choose "Weight" for Samples, and "Facility" for Sample IDs.

    t test for the mean

    Minitab gives us the following output, and I've highlighted the p-value for the hypothesis test:

    t-test output

    So we have a p-value of 0.029, which is less than our selected alpha value of 0.05. Therefore, we reject the null hypothesis that the means of Line A and Line B are equal. Note also that while the evidence indicates the means are different, that difference is estimated at 0.338 oz—a pretty small amount of cereal. 

    So far, so good. But this is the point at which trouble often starts.

    Three Frequent Misstatements about P-Values

    The p-value of 0.029 means we reject the null hypothesis that the means are equal. But that doesn't mean any of the following statements are accurate:

    1. "There is 2.9% probability the means are the same, and 97.1% probability they are different." 
      We don't know that at all. The p-value only says that if the null hypothesis is true, the sample data collected would exhibit a difference this large or larger only 2.9% of the time. Remember that the p-value doesn't tell you anything directly about what you've seen. Instead, it tells you the odds of seeing it. 

    2. "The p-value is low, which indicates there's an important difference in the means." 
      Based on the 0.029 p-value shown above, we can conclude that a statistically significant difference between the means exists. But the estimated size of that difference is less than a half-ounce, and won't matter to customers. A p-value may indicate a difference exists, but it tells you nothing about its practical impact.

    3. "The low p-value shows the alternative hypothesis is true."
      A low p-value provides statistical evidence to reject the null hypothesis—but that doesn't prove the truth of the alternative hypothesis. If your alpha level is 0.05, there's a 5% chance you will incorrectly reject the null hypothesis. Or to put it another way, if a jury fails to convict a defendant, it doesn't prove the defendant is innocent: it only means the prosecution failed to prove the defendant's guilt beyond a reasonable doubt. 

    These misinterpretations happen frequently enough to be a concern, but that doesn't mean that we shouldn't use p-values to help interpret data. The p-value remains a very useful tool, as long as we're interpreting and communicating its significance accurately.

    P-Value Results in Plain Language

    It's one thing to keep all of this straight if you're doing data analysis and statistics all the time. It's another thing if you're only analyze data occasionally, and need to do many other things in between—like most of us. "Use it or lose it" is certainly true about statistical knowledge, which could well be another factor that contributes to misinterpreted p-values. 

    If you're leery of that happening to you, a good way to avoid that possibility is to use the Assistant in Minitab to perform your analyses. If you haven't used it yet, the Assistant menu guides you through your analysis from start to finish. The dialog boxes and output are all in plain language, so it's easy to figure out what you need to do and what the results mean, even if it's been a while since your last analysis. (But even expert statisticians tell us they like using the Assistant because the output is so clear and easy to understand, regardless of an audience's statistical background.) 

    So let's redo the analysis above using the Assistant, to see what that output looks like and how it can help you avoid misinterpreting your results—or having them be misunderstood by others!

    Start by selecting Assistant > Hypothesis Test... from the Minitab menu. Note that a window pops up to explain exactly what a hypothesis test does. 

    assistant hypothesis test

    The Assistant asks what we're trying to do, and gives us three options to choose from.

    hypothesis test chooser

    We know we want to compare a sample from Line A with a sample from Line B, but what if we can't remember which of the 5 available tests is the appropriate one in this situation? We can get guidance by clicking "Help Me Choose."

    help me choose the right hypothesis test

    The choices on the diagram direct us to the appropriate test. In this case, we choose continuous data instead of attribute (and even if we'd forgotten the difference, clicking on the diamond would explain it). We're comparing two means instead of two standard deviations, and we're measuring two different sets of items since our boxes came from different production lines. 

    Now we know what test to use, but suppose you want to make sure you don't miss anything that's important about the test, like requirements that must be met? Click the "more..." link and you'll get those details. 

    more info about the 2-Sampe t-Test

    Now we can proceed to the Assistant's dialog box. Again, statistical jargon is minimized and everything is put in straightforward language. We just need to answer a few questions, as shown. Note that the Assistant even lets us tell it how big a difference needs to be for us to consider it practically important. In this case, we'll enter 2 ounces.

    Assistant 2-sample t-Test dialog

    When we press OK, the Assistant performs the t-test and delivers three reports. The first of these is a summary report, which includes summary statistics, confidence intervals, histograms of both samples, and more. And interpreting the results couldn't be more straightforward than what we see in the top left quadrant of the diagram. In response to the question, "Do the means differ?" we can see that p-value of 0.029 marked on the bar, very far toward the "Yes" end of the scale. 

    2-Sample t-Test summary report

    Next is the Diagnostic Report, which provides additional information about the test. 

    2-Sample t-Test diagnostic report

    In addition to letting us check for outliers, the diagnostic report shows us the size of the observed difference, as well as the chances that our test could detect a practically significant difference of 2 oz. 

    The final piece of output the Assistant provides is the report card, which flags any problems or concerns about the test that we would need to be aware of. In this case, all of the boxes are green and checked (instead of red and x'ed). 

    2-Sample t-Test report card

    When you're not doing statistics all the time, the Assistant makes it a breeze to find the right analysis for your situation and to make sure you interpret your results the right way. Using it is a great way to make sure you're not attaching too much, or too little, importance on the results of your analyses.

     

    Viewing all 828 articles
    Browse latest View live


    <script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>