New View of Statistics: T Test & ANOVA-1 (2024)

A New View of Statistics

© 2000 Will G Hopkins

Go to: Next · Previous · Contents· Search · Home

Generalizing to a Population:
SIMPLE MODELS AND TESTS continued

New View of Statistics: T Test & ANOVA-1 (1)T Test and One-Way ANOVA

model: numeric<=nominal
example: height<=sex
New View of Statistics: T Test & ANOVA-1 (2)In other words, if you know someone's sex, what does that tell you about their height? Or, how well do the height data fall into two groups when you label the values by sex? The test statistic for the test of whether sex has an effect on height is called Student's t, or just t. Hence the name of this model, the t test.

When there are three or more levels for the nominal variable, a simple approach is to run a series of t tests between all the pairs of levels. For example, we might be interested in the heights of athletes in three sports, so we could run t test for each pair of sports. (Note that this approach is not the same as a paired t test. That comes later.) A more powerful approach is to analyze all the data in one go. The model is the same, but it is now called a one-way analysis of variance (ANOVA), and the test statistic is the F ratio. So t tests are just a special case of ANOVA: if you analyze the means of two groups by ANOVA, you get the same results as doing it with a t test.

The term analysis of variance is a source of confusion for newbies. In spite of its name, ANOVA is concerned with differences between means of groups, not differences between variances. The name analysis of variance comes from the way the procedure uses variances to decide whether the means are different. A better acronym for this model would be ANOVASMAD (analysis of variance to see if means are different)! The way it works is simple: the program looks to see what the variation (variance) is within the groups, then works out how that variation would translate into variation (i.e. differences) between the groups, taking into account how many subjects there are in the groups. If the observed differences are a lot bigger than what you'd expect by chance, you have statistical significance. In our example, there are only two groups, so variation between groups is just the difference between the means.

I won't bother with trying to represent this model as an equation like Y = mX + c. Suffice to say that it can be done, simply by making an X variable representing sex that has the value 0 for females and 1 for males, say (or vice versa). So it is also a "linear" model, even though we don't normally think about it as a straight line. The parameters in the model are simply the mean for the females and the mean for the males.

The spreadsheet for analysis of controlled trials includes a comparison of the means (and standard deviations) of two groups at baseline. You can use it for any tests of two independent groups, as in the above example.. Ignore all the stuff related to comparisons of changes in the mean in the two groups.

Comparisons of Means
With a t test, the thing we're most interested in is, of course, a comparison of the two means. You should think about the best way to express the difference in the means for your data: raw units, percent difference, or effect size. And don't forget to look at and discuss the magnitude of the difference and the magnitude of its confidence limits.

With three or more levels for the nominal variable, we can start asking interesting questions about the differences between pairs or combinations of means. Such comparisons of means are known as estimates or contrasts. New View of Statistics: T Test & ANOVA-1 (3)For example, suppose we are exploring the relationship between training hours per week (the dependent variable) and sport (the nominal independent variable). Suppose sport has three levels: runners, cyclists, and swimmers, as shown. We can ask the question, are there differences overall between the sports? The answer would be given by the p value for sport in the model. And what about the difference between cycling and running? Yes, we can dial up the difference and look at its p value or confidence interval. We do that by subtracting the value for the parameter (the mean) for cycling from that for running, using the appropriate syntax in the stats program. We could even ask how different swimming was from the average of running and cycling, and so on. There's also a special kind of contrast (polynomials) you can apply if the levels are a numbered sequence and you want to describe a curve drawn through the values for each level.

If you're expressing a difference between means as an effect size, the standard deviation to use in the calculation is the root mean square error (RMSE) in the ANOVA. An ANOVA is based on the assumption that the standard deviation in the same in all the groups, and the RMSE represents the estimate of that standard deviation. You can think of the RMSE as the average standard deviation for all of the groups.

With lots of contrasts, the chance of any one of them being spuriously statistically significant--in other words, the overall chance of a Type I error--goes up. So stats programs usually have built-in ways of controlling the overall Type I error rate in an ANOVA. Basically they adjust the p value down for declaring statistical significance, although you don't see it like that on the printout. These methods have statisticians' names: Tukey, Duncan, Bonferroni... They're also known as post-hoc tests or simply post hocs. I don't use them, because I now use confidence limits and clinical significance rather than statistical significance, so I don't test anything.

One approach to controlling the Type I error rate with multiple contrasts is simply not to perform the contrasts unless the overall effect is significant. In other words, you don't ask where the differences are between groups unless there is an overall difference between groups. Sounds reasonable, but wait a moment! If there is no overall statistically significant difference between groups, surely none of the contrasts will turn up significant? Yes, it can happen! There's jitter in the p values, and there's nothing to say that the p value for the overall effect is any more valid than the p value for individual contrasts. So if you've set up your study with a particular contrast in mind--a pre-planned contrast--go ahead and do that contrast, regardless of the p value for the overall effect. Performing the pre-planned contrast does not have to be contingent upon obtaining significance for the overall effect. Those of us who prefer confidence intervals to p values can understand why: the estimate of the difference between groups has a confidence interval that may or may not overlap zero, and the confidence interval for the overall effect (expressed in some measure of goodness of fit) may or may not overlap zero. There is no need to reconcile the two.

Goodness of Fit
What statistic do we use to talk about how well the ANOVA model fits the data? It's not used that frequently, but you can extract an R2 just like you do for a straight line. The R2 represents how well all the levels of the grouping (nominal) variable fit the data. More about goodness of fit soon.Go to: Next · Previous · Contents· Search · Home editor
Last updated 2 Nov 03

New View of Statistics: T Test & ANOVA-1 (2024)

FAQs

When to use one-way ANOVA vs t-test? ›

The Student's t test is used to compare the means between two groups, whereas ANOVA is used to compare the means among three or more groups. In ANOVA, first gets a common P value. A significant P value of the ANOVA test indicates for at least one pair, between which the mean difference was statistically significant.

What is the relationship between t-test and ANOVA? ›

ANOVA compares means among three or more groups, whereas t-tests solely compare means between two groups. ANOVA encompasses an analysis of between-group and within-group variation, whereas t-tests focus solely on within-group variation.

Why should you use ANOVA instead of several t-tests? ›

So, if two t-tests are being conducted, there is a 10% chance of conducting a Type I error. Using ANOVA in this scenario (that is comparing means of three or more groups) restricts the chance of Type I error to 5% and therefore results are more statistically significant.

What are the assumptions of t-test and ANOVA? ›

Assumptions. Both the t-test and the ANOVA have the same assumptions: normality and hom*ogeneity of variance. The normality assumptions can be assessed with a Shapiro Wilks test or by a Q-Q scatterplot. The hom*ogeneity of variance test can be assessed with the Levene's test.

When not to use ANOVA? ›

ANOVA requires the dependent variable to be continuous (interval/ratio), and the independent variable to be categorical (nominal/ordinal). If your variables do not meet these requirements, then ANOVA may not be the best choice.

What is an advantage of one-way ANOVA over multiple t-tests? ›

As such, three t-tests would be 15% (actually, 14.3%) and so on. These are unacceptable errors. An ANOVA controls for these errors so that the Type I error remains at 5% and you can be more confident that any statistically significant result you find is not just running lots of tests.

What is the major advantage to using ANOVA over t-test? ›

With a t-test, one would have to do multiple one-to-one comparisons. By contrast, with ANOVA, one can run a single test, with multiple different recipes in comparison to each other.

What is the main advantage that ANOVA has compared to t-test? ›

The t-test is a method that determines whether two populations are statistically different from each other, whereas ANOVA determines whether three or more populations are statistically different from each other.

When to use t-test vs ANOVA vs chi square? ›

While t-tests and ANOVA primarily deal with continuous dependent variables, Chi-Square tests come into play when there is a categorical dependent variable, often in the context of logistic regression.

When to use a t-test? ›

A t-test may be used to evaluate whether a single group differs from a known value (a one-sample t-test), whether two groups differ from each other (an independent two-sample t-test), or whether there is a significant difference in paired measurements (a paired, or dependent samples t-test).

What does ANOVA tell you? ›

ANOVA, which stands for Analysis of Variance, is a statistical test used to analyze the difference between the means of more than two groups.

When ANOVA is used and why it is a better way than performing multiple t-tests What is the purpose of doing a post hoc test? ›

Post hoc tests attempt to control the experimentwise error rate (usually alpha = 0.05) in the same manner that the one-way ANOVA is used instead of multiple t-tests. Post hoc tests are termed a posteriori tests; that is, performed after the event (the event in this case being a study).

What are the 3 statistical conditions or assumptions required for ANOVA? ›

There are three primary assumptions in ANOVA: The responses for each factor level have a normal population distribution. These distributions have the same variance. The data are independent.

What are the limitations of the t-test? ›

While the t-test is a versatile tool, it does have some limitations. It assumes normality and hom*ogeneity of variances, and violations of these assumptions can affect the accuracy of results. Additionally, the t-test is most effective with small sample sizes; the z-test may be more appropriate for larger samples.

When should you use a one-way ANOVA test? ›

One-way ANOVA is typically used when you have a single independent variable, or factor, and your goal is to investigate if variations, or different levels of that factor have a measurable effect on a dependent variable.

How do you know when to use a one-way or two-way ANOVA? ›

1. A one-way ANOVA is primarily designed to enable the equality testing between three or more means. A two-way ANOVA is designed to assess the interrelationship of two independent variables on a dependent variable.

How do you know when to use a t-test? ›

A t-test may be used to evaluate whether a single group differs from a known value (a one-sample t-test), whether two groups differ from each other (an independent two-sample t-test), or whether there is a significant difference in paired measurements (a paired, or dependent samples t-test).

Top Articles
'Kaos' review: Can Netflix’s Greek myth series go the distance?
KAOS Review: Miscast Jeff Goldblum Is Netflix Greek Mythology Dramedy's Fatal Flaw
Bubble Guppies Who's Gonna Play The Big Bad Wolf Dailymotion
Libiyi Sawsharpener
Windcrest Little League Baseball
Mohawkind Docagent
Space Engineers Projector Orientation
Cranberry sauce, canned, sweetened, 1 slice (1/2" thick, approx 8 slices per can) - Health Encyclopedia
South Bend Tribune Online
Tokioof
Pittsburgh Ultra Advanced Stain And Sealant Color Chart
Dutchess Cleaners Boardman Ohio
Craigslist Malone New York
Costco Gas Foster City
Weather Rotterdam - Detailed bulletin - Free 15-day Marine forecasts - METEO CONSULT MARINE
Lowe's Garden Fence Roll
Geometry Review Quiz 5 Answer Key
Rs3 Eldritch Crossbow
Optum Urgent Care - Nutley Photos
Mtr-18W120S150-Ul
Egizi Funeral Home Turnersville Nj
Home
Wkow Weather Radar
Apparent assassination attempt | Suspect never had Trump in sight, did not get off shot: Officials
Times Narcos Lied To You About What Really Happened - Grunge
Ultra Ball Pixelmon
Log in or sign up to view
Ipcam Telegram Group
Bridgestone Tire Dealer Near Me
Max 80 Orl
Rocketpult Infinite Fuel
Samsung 9C8
Space Marine 2 Error Code 4: Connection Lost [Solved]
Msnl Seeds
Mydocbill.com/Mr
Bbc Gahuzamiryango Live
About :: Town Of Saugerties
Main Street Station Coshocton Menu
Spn-523318
Metro Pcs Forest City Iowa
Linkbuilding uitbesteden
Lady Nagant Funko Pop
Noh Buddy
Hampton In And Suites Near Me
Dying Light Mother's Day Roof
Big Brother 23: Wiki, Vote, Cast, Release Date, Contestants, Winner, Elimination
Motorcycle For Sale In Deep East Texas By Owner
Image Mate Orange County
Hy-Vee, Inc. hiring Market Grille Express Assistant Department Manager in New Hope, MN | LinkedIn
Craigs List Sarasota
Ranking 134 college football teams after Week 1, from Georgia to Temple
Intuitive Astrology with Molly McCord
Latest Posts
Article information

Author: Horacio Brakus JD

Last Updated:

Views: 6527

Rating: 4 / 5 (51 voted)

Reviews: 90% of readers found this page helpful

Author information

Name: Horacio Brakus JD

Birthday: 1999-08-21

Address: Apt. 524 43384 Minnie Prairie, South Edda, MA 62804

Phone: +5931039998219

Job: Sales Strategist

Hobby: Sculling, Kitesurfing, Orienteering, Painting, Computer programming, Creative writing, Scuba diving

Introduction: My name is Horacio Brakus JD, I am a lively, splendid, jolly, vivacious, vast, cheerful, agreeable person who loves writing and wants to share my knowledge and understanding with you.