1、Designation: G 169 01Standard Guide forApplication of Basic Statistical Methods to WeatheringTests1This standard is issued under the fixed designation G 169; the number immediately following the designation indicates the year oforiginal adoption or, in the case of revision, the year of last revision
2、. A number in parentheses indicates the year of last reapproval. Asuperscript epsilon (e) indicates an editorial change since the last revision or reapproval.1. Scope1.1 This guide covers elementary statistical methods for theanalysis of data common to weathering experiments. Themethods are for deci
3、sion making, in which the experiments aredesigned to test a hypothesis on a single response variable. Themethods work for either natural or laboratory weathering.1.2 Only basic statistical methods are presented. There aremany additional methods which may or may not be applicableto weathering tests t
4、hat are not covered in this guide.1.3 This guide is not intended to be a manual on statistics,and therefore some general knowledge of basic and interme-diate statistics is necessary. The text books referenced at theend of this guide are useful for basic training.1.4 This guide does not provide a rig
5、orous treatment of thematerial. It is intended to be a reference tool for the applicationof practical statistical methods to real-world problems thatarise in the field of durability and weathering. The focus is onthe interpretation of results. Many books have been written onintroductory statistical
6、concepts and statistical formulas andtables. The reader is referred to these for more detailedinformation. Examples of the various methods are included.The examples show typical weathering data for illustrativepurposes, and are not intended to be representative of specificmaterials or exposures.2. R
7、eferenced Documents2.1 ASTM Standards:D 1898 Practice for Sampling of Plastics2E 41 Terminology Relating to Conditioning3G 113 Terminology Relating to Natural and ArtificialWeathering Tests of Nonmetallic Materials3G 141 Guide for Addressing Variability in Exposure Testingon Nonmetallic Materials32.
8、2 ISO Documents:ISO 3534/1 Vocabulary and Symbols Part 1: Probabilityand General Statistical Terms4ISO 3534/3 Vocabulary and Symbols Part 3: Design ofExperiments43. Terminology3.1 DefinitionsSee Terminology G 113 for terms relatingto weathering, Terminology E 41 for terms relating to condi-tioning a
9、nd handling, ISO 3534/1 for terminology relating tostatistics, and ISO 3534/3 for terms relating to design ofexperiments.3.2 Definitions of Terms Specific to This Standard:3.2.1 arithmetic mean; averagethe sum of values dividedby the number of values. ISO 3534/13.2.2 blocking variablea variable that
10、 is not under thecontrol of the experimenter, (for example, temperature andprecipitation in exterior exposure), and is dealt with byexposing all samples to the same effects3.2.2.1 DiscussionThe term “block” originated in agricul-tural experiments in which a field was divided into sections orblocks h
11、aving common conditions such as wind, proximity tounderground water, or thickness of the cultivatable layer.ISO 3534/33.2.3 correlationin weathering, the relative agreement ofresults from one test method to another, or of one test specimento another.3.2.4 medianthe midpoint of ranked sample values.
12、Insamples with an odd number of data, this is simply the middlevalue, otherwise it is the arithmetic average of the two middlevalues.3.2.5 nonparametric methoda statistical method that doesnot require a known or assumed sample distribution in order tosupport or reject a hypothesis.3.2.6 normalizatio
13、na mathematical transformation madeto data to create a common baseline.3.2.7 predictor variable (independent variable) a variablecontributing to change in a response variable, and essentiallyunder the control of the experimenter. ISO 3534/33.2.8 probability distribution (of a random variable)afuncti
14、on giving the probability that a random variable takes any1This guide is under the jurisdiction of ASTM Committee G03 on Weatheringand Durability and is the direct responsibility of Subcommittee G03.93 on Relatingto Statistics.Current edition approved Jan. 10, 2001. Published May 2001.2Discontinued
15、1998; see 1997 Annual Book of ASTM Standards, Vol 08.01.3Annual Book of ASTM Standards, Vol 14.04.4Available from American National Standards Institute, 11 W. 42nd St., 13thFloor, New York, NY 10036.1Copyright ASTM International, 100 Barr Harbor Drive, PO Box C700, West Conshohocken, PA 19428-2959,
16、United States.given value or belongs to a given set of values. ISO 3534/13.2.9 random variablea variable that may take any of thevalues of a specified set of values and with which is associateda probability distribution.3.2.9.1 DiscussionA random variable that may take onlyisolated values is said to
17、 be “discrete.” A random variablewhich may take any value within a finite or infinite interval issaid to be “continuous.” ISO 3534/13.2.10 replicatestest specimens with nominally identicalcomposition, form, and structure.3.2.11 response variable (dependent variable) a randomvariable whose value depe
18、nds on other variables (factors).Response variables within the context of this guide are usuallyproperty measurements (for example, tensile strength, gloss,color, and so forth). ISO 3534/34. Significance and Use4.1 The correct use of statistics as part of a weatheringprogram can greatly increase the
19、 usefulness of results. A basicunderstanding of statistics is required for the study of weath-ering performance data. Proper experimental design and sta-tistical analysis strongly enhances decision-making ability. Inweathering, there are many uncertainties brought about byexposure variability, metho
20、d precision and bias, measurementerror, and material variability. Statistical analysis is used tohelp decide which products are better, which test methods aremost appropriate to gauge end use performance, and howreliable the results are.4.2 Results from weathering exposures can show differ-ences bet
21、ween products or between repeated testing. Theseresults may show differences which are not statistically signifi-cant. The correct use of statistics on weathering data canincrease the probability that valid conclusions are derived.5. Test Program Development5.1 Hypothesis Formulation:5.1.1 All of th
22、e statistical methods in this guide are designedto test hypotheses. In order to apply the statistics, it isnecessary to formulate a hypothesis. Generally, the testing isdesigned to compare things, with the customary comparisonbeing:Do the predictor variables significantly affect theresponse variable
23、?Taking this comparison into consideration, it is possible toformulate a default hypothesis that the predictor variables donot have a significant effect on the response variable. Thisdefault hypothesis is usually called Ho, or the Null Hypothesis.5.1.2 The objective of the experimental design and st
24、atisti-cal analysis is to test this hypothesis within a desired level ofsignificance, usually an alpha level (a). The alpha level is theprobability below which we reject the null hypothesis. It can bethought of as the probability of rejecting the null hypothesiswhen it is really true (that is, the c
25、hance of making such anerror). Thus, a very small alpha level reduces the chance inmaking this kind of an error in judgment. Typical alpha levelsare 5 % (0.05) and 1 % (0.01). The x-axis value on a plot of thedistribution corresponding to the chosen alpha level is gener-ally called the critical valu
26、e (cv).5.1.3 The probability that a random variable X is greaterthan the critical value for a given distribution is written P(Xcv). This probability is often called the “p-value.” In thisnotation, the null hypothesis can be rejected ifP(Xcv) cv)Separate variances 3.116 4.9 0.036Pooled variances 3.00
27、0 6.0 0.024P(Xcv) indicates the probability that a Studentst-distributed random variable is greater than the cv, that is, thearea under the tail of the t-distribution to the right of Point t.Since this value in either case is below a pre-chosen alpha levelof 0.05, the result is significant. Note tha
28、t this result would notbe significant at an alpha level of 0.01.6.3 ANOVA:FIG. 1 Selecting a MethodTABLE 2 STUDENTS t-TEST EXAMPLEColor Change Formula1.000 A1.200 A1.100 A0.900 A1.100 A1.300 B1.400 B1.200 BG1690146.3.1 Analysis of Variance (ANOVA) performs comparisonslike the t-Test, but for an arbi
29、trary number of predictorvariables, each of which can have an arbitrary number oflevels. Furthermore, each predictor variable combination canhave any number of replicates. Like all the methods in thisguide, ANOVA works on a single response variable. Thepredictor variables must be discrete. See Table
30、 3.6.3.2 The ANOVA can be thought of in a practical sense asan extension of the t-Test to an arbitrary number of factors andlevels. It can also be thought of as a linear regression modelwhose predictor variables are restricted to a discrete set. Hereis the example cited in the t-Test, extended to in
31、clude anadditional formula, and another factor. The new factor is to testwhether the resulting formulation is affected by the technicianwho prepared it. There are two technicians and three formulasunder consideration.6.3.3 This example also illustrates that one need not haveidentical numbers of repl
32、icates for each sample. In thisexample, there are two replicates per factor combination forFormula A, but no replication appears for the other formulas.Analysis of VarianceResponse variable: COLOR CHANGESourceSum ofSquaresDegrees ofFreedom Mean square F Ratio P(Xcv)Formula 0.483 2 0.241 16.096 0.025
33、Technician 0.005 1 0.005 0.333 0.604Error 0.045 3 0.015 - -6.3.4 Assuming an alpha level of 0.05, the analysis indicatesthat the formula resulted in a significant difference in colorchange means, but the technician did not. This is evident fromthe probability values in the final column. Values below
34、 thealpha level allow rejection of the null hypothesis.6.4 Linear Regression:6.4.1 Linear regression is essentially an ANOVA in whichthe factors can take on continuous values. Since discretefactors can be set up as belonging to a subset of some largercontinuous set, linear regression is a more gener
35、al method. It isin fact the most general method considered in this guide. SeeTable 4.6.4.2 The most elementary form of linear regression is easyto visualize. It is the case in which we have one predictorvariable and one response variable. The easy way to think ofthe predictor variable is as an x-axi
36、s value of a two dimensionalplot. For each predictor variable level, we can plot thecorresponding measurement (response variable) as a value onthe ordinate axis. The idea is to see how well we can fit a lineto the points on the plot. See Table 5.6.4.3 For example, the following experiment looks at t
37、heeffect of an impact modifying ingredient level on impactstrength after one year of outdoor weathering in Arizona.6.4.4 The plot of ingredient level versus retained impactstrength shown with a linear fit and 95 % confidence bandslooks like: (See Fig. 2)6.4.5 This example illustrates the use of repl
38、icates at one ofthe levels. It is a good idea to test replicates at the levels thatare thought to be important or desirable. The analysis indicatesa good linear fit. We see this from the R2value (squaredmultiple R) of 0.976. The R2value is the fraction of thevariability of the response variable expl
39、ained by the regressionmodel, indicates the degree of fit to the model.6.4.6 The analysis of variance indicates a significant rela-tionship between modifier level and retained impact strength inthis test (the probability level is well below an alpha level of5 %).Linear Regression AnalysisResponse Va
40、riable: Impact Retention (%)Number of Observations: 7Multiple R: 0.988Squared Multiple R: 0.976SourceDegrees ofFreedom Sum of Squares F Ratio P(Xcv)Regression 1 0.0464 205.1 less than 0.0001Residual 5 0.0011 - -6.4.7 Regression can be easily generalized to more than onefactor, although the data gets
41、 difficult to visualize since eachfactor adds an axis to the plot (it is not so easy to viewmultidimensional data sets). It can also be adapted to nonlinearmodels. A common technique for achieving this is to transformdata so that it is linear. Another way is to use nonlinear leastsquares methods, wh
42、ich are beyond the scope of this guide.Regression can also be extended to cover mixed continuousand discrete factors. It should be noted that most spreadsheetand elementary data analysis applications can perform fairlysophisticated regression analysis.TABLE 3 ANOVA EXAMPLEColor Change Formula Techni
43、cian1.000 A Elmo1.100 A Elmo1.100 A Homer0.900 A Homer1.300 B Elmo1.400 B Judd1.200 B Homer0.700 C Elmo0.600 C HomerTABLE 4 REGRESSION EXAMPLEModifier Level Impact Retention After Exposure0.005 0.5350.01 0.60.02 0.6350.02 0.620.03 0.680.04 0.7540.05 0.79TABLE 5 PATHOLOGICAL LINEAR REGRESSION EXAMPLE
44、xv0.01 0.0299790.02 0.0543380.03 0.0885810.04 0.0824150.05 0.1266310.06 0.0734640.07 0.1232220.08 0.0970030.09 0.0997280.75 0.8059090.86 0.865667G1690156.4.8 Another use of regression is to compare two predictorrandom variables at a number of levels for each. For example,results from one exposure te
45、st can be plotted against the resultsfrom another exposure. If the points fall on a line, then onecould conclude that the tests are “in agreement.” This is calledcorrelation. The usual statistic in a linear correlation analysis isR2, which is a measure of deviation from the model (a straightline). T
46、he R2values near one indicate good agreement with themodel, while those near zero indicate poor agreement. Thistype of analysis is different from the approaches suggestedabove which were constructed to test whether one randomvariable depended somehow on others. It should be noted,however, that corre
47、lation can always be phrased in ANOVA-like terms. The correlation example included for the Spearmanrank correlation method illustrates this. The observations thenmake up a response random variable. Correlation on absoluteresults is not recommended in weathering testing. Instead,relative data (ranked
48、 data) often provide more meaningfulanalysis (see Spearmans rank correlation).6.4.9 Regression/correlation can lead to misleadingly highR2values when the x-axis values are not well-spaced. Considerthe following example, which contains a cluster of data thatdoes not exhibit a good linear fit, along w
49、ith a few outliers. Dueto the large spread in the x-axis values, the clustered dataappears almost as a single data point, resulting in a high R2value. (See Fig. 3).Linear Regression AnalysisNumber of Observations: 11Multiple R: 0.997Squared Multiple R: 0.994SourceDegrees ofFreedom Sum of Squares F Ratio P(Xcv)Regression 1 0.9235 1509 less than 0.0001Residual 9 0.0055 - -6.4.10 Even though the analysis indicates a good fit to alinear model, the cluster of data does not fit a linear model wellat all without the outliers. If the objective of this analysis werecorrelatio