ImageVerifierCode 换一换
格式:PDF , 页数:12 ,大小:138.90KB ,
资源ID:530383      下载积分:5000 积分
快捷下载
登录下载
邮箱/手机:
温馨提示:
如需开发票,请勿充值!快捷下载时,用户名和密码都是您填写的邮箱或者手机号,方便查询和重复下载(系统自动生成)。
如填写123,账号就是123,密码也是123。
特别说明:
请自助下载,系统不会自动发送文件的哦; 如果您已付费,想二次下载,请登录后访问:我的下载记录
支付方式: 支付宝扫码支付 微信扫码支付   
注意:如需开发票,请勿充值!
验证码:   换一换

加入VIP,免费下载
 

温馨提示:由于个人手机设置不同,如果发现不能下载,请复制以下地址【http://www.mydoc123.com/d-530383.html】到电脑端继续下载(重复下载不扣费)。

已注册用户请登录:
账号:
密码:
验证码:   换一换
  忘记密码?
三方登录: 微信登录  

下载须知

1: 本站所有资源如无特殊说明,都需要本地电脑安装OFFICE2007和PDF阅读器。
2: 试题试卷类文档,如果标题没有明确说明有答案则都视为没有答案,请知晓。
3: 文件的所有权益归上传用户所有。
4. 未经权益所有人同意不得将文件中的内容挪作商业或盈利用途。
5. 本站仅提供交流平台,并不能对任何下载内容负责。
6. 下载文件中如有侵权或不适当内容,请与我们联系,我们立即纠正。
7. 本站不保证下载资源的准确性、安全性和完整性, 同时也不承担用户因使用这些下载资源对自己和他人造成任何形式的伤害或损失。

版权提示 | 免责声明

本文(ASTM E2139-2005(2011) Standard Test Method for Same-Different Test《异同试验的标准试验方法》.pdf)为本站会员(postpastor181)主动上传,麦多课文库仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对上载内容本身不做任何修改或编辑。 若此文所含内容侵犯了您的版权或隐私,请立即通知麦多课文库(发送邮件至master@mydoc123.com或直接QQ联系客服),我们立即给予删除!

ASTM E2139-2005(2011) Standard Test Method for Same-Different Test《异同试验的标准试验方法》.pdf

1、Designation: E2139 05 (Reapproved 2011)Standard Test Method forSame-Different Test1This standard is issued under the fixed designation E2139; the number immediately following the designation indicates the year oforiginal adoption or, in the case of revision, the year of last revision. A number in pa

2、rentheses indicates the year of last reapproval. Asuperscript epsilon () indicates an editorial change since the last revision or reapproval.1. Scope1.1 This test method describes a procedure for comparingtwo products.1.2 This test method does not describe the Thurstonianmodeling approach to this te

3、st.1.3 This test method is sometimes referred to as the simple-difference test.1.4 A same-different test determines whether two productsare perceived to be the same or different overall.1.5 The procedure of the test described in this test methodconsists of presenting a single pair of samples to each

4、 assessor.The presentation of multiple pairs would require differentstatistical treatment and it is outside of the scope of this testmethod.1.6 This test method is not attribute-specific, unlike thedirectional difference test.1.7 This test method is not intended to determine themagnitude of the diff

5、erence; however, statistical methods maybe used to estimate the size of the difference.1.8 This test method may be chosen over the triangle orduo-trio tests where sensory fatigue or carry-over are aconcern, or where a simpler task is needed.1.9 This standard may involve hazardous materials, opera-ti

6、ons, and equipment. This standard does not purport toaddress all of the safety concerns, if any, associated with itsuse. It is the responsibility of the user of this standard toestablish appropriate safety and health practices and deter-mine the applicability of regulatory limitations prior to use.2

7、. Referenced Documents2.1 ASTM Standards:2E253 Terminology Relating to Sensory Evaluation of Ma-terials and ProductsE456 Terminology Relating to Quality and StatisticsE1871 Guide for Serving Protocol for Sensory Evaluationof Foods and Beverages2.2 ASTM Publications:2Manual 26 Sensory Testing Methods

8、, 2nd EditionSTP 758 Guidelines for the Selection and Training ofSensory Panel MembersSTP 913 Guidelines for Physical Requirements for SensoryEvaluation Laboratories2.3 ISO Standard:3ISO 5495 Sensory AnalysisMethodologyPaired Com-parison3. Terminology3.1 For definition of terms relating to sensory a

9、nalysis, seeTerminology E253, and for terms relating to statistics, seeTerminology E456.3.2 Definitions of Terms Specific to This Standard:3.2.1 a (alpha) riskprobability of concluding that a per-ceptible difference exists when, in reality, one does not (alsoknown as Type I Error or significance lev

10、el).3.2.2 b (beta) riskprobability of concluding that no per-ceptible difference exists when, in reality, one does (alsoknown as Type II Error).3.2.3 chi-square teststatistical test used to test hypotheseson frequency counts and proportions.3.2.4 D (delta)test sensitivity parameter established prior

11、to testing and used along with the selected values of a, b, andan estimated value of p1to determine the number of assessorsneeded in a study. Delta (D) is the minimum difference inproportions that the researcher wants to detect, where thedifference is D = p2 p1. D is not a standard measure ofsensory

12、 difference. The same value of D may correspond todifferent sensory differences for different values of p1(see 9.5for an example).3.2.5 Fishers Exact Test (FET)statistical test of theequality of two independent binomial proportions.3.2.6 p1proportion of assessors in the population whowould respond d

13、ifferent to the matched sample pair. Based onexperience with using the same-different test and possibly withthe same type of products, the user may have a prioriknowledge about the value of p1.3.2.7 p2proportion of assessors in the population whowould respond different to the unmatched sample pair.1

14、This test method is under the jurisdiction ofASTM Committee E18 on SensoryEvaluation and is the direct responsibility of Subcommittee E18.04 on Fundamen-tals of Sensory.Current edition approved Aug. 1, 2011. Published August 2011. Originallyapproved in 2005. Last previous edition approved in 2005 as

15、 E213905. DOI:10.1520/E2139-05R11.2For referenced ASTM standards, visit the ASTM website, www.astm.org, orcontact ASTM Customer Service at serviceastm.org. For Annual Book of ASTMStandards volume information, refer to the standards Document Summary page onthe ASTM website.3Available from American Na

16、tional Standards Institute (ANSI), 25 W. 43rd St.,4th Floor, New York, NY 10036, http:/www.ansi.org.1Copyright ASTM International, 100 Barr Harbor Drive, PO Box C700, West Conshohocken, PA 19428-2959, United States.3.2.8 power 1-b (beta) riskprobability of concluding thata perceptible difference exi

17、sts when, in reality, one of size Ddoes.3.2.9 productmaterial to be evaluated.3.2.10 sampleunit of product prepared, presented, andevaluated in the test.3.2.11 sensitivityterm used to summarize the performancecharacteristics of this test. The sensitivity of the test is definedby the four values sele

18、cted for a, b, p1, and D.4. Summary of Test Method4.1 Clearly define the test objective in writing.4.2 Choose the number of assessors based on the sensitivitydesired for the test. The sensitivity of the test is in part relatedto two competing risks: the risk of declaring a difference whenthere is no

19、ne (that is, a-risk), and the risk of not declaring adifference when there is one (that is, b-risk).Acceptable valuesof a and b vary depending on the test objective. The valuesshould be agreed upon by all parties affected by the results ofthe test.4.3 The two products of interest (A and B) are selec

20、ted.Assessors are presented with one of four possible pairs ofsamples: A/A, B/B, A/B, and B/A. The total number of samepairs (A/Aand B/B) usually equals the total number of differentpairs (A/B and B/A). The assessors task is to categorize thegiven pair of samples as same or different.4.4 The data ar

21、e summarized in a two-by-two table wherethe columns show the type of pair received (same or different)and the rows show the assessors response (same or different).A Fishers Exact Test (FET) is used to determine whether thesamples are perceptibly different. Other statistical methods thatapproximate t

22、he FET can sometimes be used.5. Significance and Use5.1 This overall difference test method is used when the testobjective is to determine whether a sensory difference exists ordoes not exist between two samples. It is also known as thesimple difference test.5.2 The test is appropriate in situations

23、 where samples haveextreme intensities, give rapid sensory fatigue, have longlingering flavors, or cannot be consumed in large quantities, ora combination thereof.5.3 The test is also appropriate for situations where thestimulus sites are limited to two (for example, two hands, eachside of the face,

24、 two ears).5.4 The test provides a measure of the bias where judgesperceive two same products to be different.5.5 The test has the advantage of being a simple andintuitive task.6. Apparatus6.1 Carry out the test under conditions that prevent contactbetween assessors until the evaluations have been c

25、ompleted,for example, booths that comply with STP 913.6.2 For food and beverage tests, sample preparation andserving sizes should comply with Practice E1871, or see Refs(1) or (2).47. Definition of Hypotheses7.1 This test can be characterized by a two-by-two table ofprobabilities according to the sa

26、mple pair that the assessors inthe population would receive and their responses, as follows:Assessor Would ReceiveMatched Pair(AA or BB)Unmatched Pair(AB or BA)AssessorsResponseSame: 1 p11p2Different: p1p2=(=p1+ D)Total: 1 1where p1and p2are the probabilities of responding differentfor those who wou

27、ld receive the matched pairs and theunmatched pairs, respectively.7.2 To determine whether the samples are perceptibly dif-ferent with a given sensitivity, the following one-sided statis-tical hypothesis is tested:Ho: p1= p2Ha: p10).Delta (D) will equal 0 and p1will equal p2if there is nodetectable

28、difference between the samples. This test addresseswhether or not D is greater than 0. Thus, the hypothesis isone-sided because it is not of interest in this test to considerthat responding different to the matched pair could be morelikely than responding different to the unmatched pair.8. Assessors

29、8.1 All assessors must be familiar with the mechanics of thesame-different test (the format, the task, and the procedure ofevaluation). Greater test sensitivity, if needed, may be achievedthrough selection of assessors who demonstrate above averageindividual sensitivity (see STP 758).8.2 In order to

30、 perform this test, assessors do not requirespecial sensory training on the samples in question. Forexample, they do not need to be able to recognize any specificattribute.8.3 The assessors must be sampled from a homogeneouspopulation that is well-defined. The population must be chosenon the basis o

31、f the test objective. Defining characteristics of thepopulation can be, for example, training level, gender, experi-ence with the product, and so forth.9. Number of Assessors9.1 Choose all the sensitivity parameters that are needed tochoose the number of assessors for the test. Choose the a-riskand

32、the b-risk. Based on experience, choose the expected valuefor p1. Choose D, p2 p1, the minimum difference in propor-tions that the researcher wants to detect. The most commonlyused values for a-risk, b-risk, p1and D are a = 0.05, b = 0.20,4The boldface numbers in parentheses refer to the list of ref

33、erences at the end ofthis standard.E2139 05 (2011)2p1= 0.3, and D = 0.3. These values can be adjusted on acase-by-case basis to reflect the sensitivity desired versus thenumber of assessors.9.2 Having defined the required sensitivity (a-risk, b-risk,p1, and D), determine the corresponding sample siz

34、e fromTable A1.1 (see Ref (9). This is done by first finding thesection of the table with a p1value corresponding to theproportion of assessors in the population who would responddifferent to the matched sample pair. Second, locate the totalsample size from the intersection of the desired a, p2(or D

35、),and b values. In the case of the most commonly used valueslisted in 9.1, TableA1.1 indicates that 84 assessors are needed.The sample size n is based on the number of same and differentsamples being equal The sample sizes listed are the totalsample size rounded up to the nearest number evenly divis

36、ibleby 4 since there are four possible combinations of the samples.To determine the number of same and different pairs to prepare,divide n by two.9.3 If the user has no prior experience with the same-different test and has no specific expectation for the value of p1,then two options are available. E

37、ither use p1= 0.3 and proceedas indicated in 9.2, or use the last section of Table A1.1. Thissection gives samples sizes that are the largest required, givena, b, and D, regardless of p1.9.4 Often in practice, the number of assessors is determinedby practical conditions (for example, duration of the

38、 experi-ment, number of available assessors, quantity of product, andso forth) However, increasing the number of assessors in-creases the likelihood of detecting small differences. Thus, oneshould expect to use larger numbers of assessors when trying todemonstrate that products are similar compared

39、to when one istrying to demonstrate that they are different.9.4.1 When the number of assessors is fixed, the power ofthe test (1-b) may be calculated by establishing a value for p1,defining the required sensitivity for a-risk and the D, locatingthe number of assessors nearest the fixed amount, and t

40、henfollowing up the column to the listed b-risk.9.5 If a researcher wants to be 90 % certain of detectingresponse proportions of p2= 60 % versus the expectedp1= 40 % with an a-risk of 5 %, then D = 0.60 0.40 = 0.20and b = 0.10 or 90 % power. The number of assessors neededin this case is 232 (Table A

41、1.1). If a researcher wants to be90 % certain of detecting response proportions of p2=70%versus the expected p1= 50 % with an a-risk of 5 %, then D =0.70 0.50 = 0.20 and b = 0.10 or 90 % power. The numberof assessors needed in this case is 224 (Table A1.1).10. Procedure10.1 Determine the number of a

42、ssessors needed for the testas well as the population that they should represent (forexample, assessors selected for a specific sensory sensitivity).10.2 It is critical to the validity of the test that assessorscannot identify the samples from the way in which they arepresented. One should avoid any

43、 subtle differences in tempera-ture or appearance, especially color, caused by factors such asthe time sequence of preparation. It may be possible to maskcolor differences using light filters, subdued illumination orcolored vessels. Prepare samples out of sight and in anidentical manner: same appara

44、tus, same vessels, same quanti-ties of product (see Practice E1871). The samples may beprepared in advance; however, this may not be possible for alltypes of products. It is essential that the samples cannot berecognized from the way they are presented.10.3 Prepare serving order worksheet and ballot

45、 in advanceof the test to ensure a balanced order of sample presentation ofthe two products, A and B. One of four possible pairs (A/A,B/B,A/B, and B/A) is assigned to each assessor. Make sure thisassignment is done randomly. Design the test so that thenumber of same pairs equals the number of differ

46、ent pairs. Thepresentation order of the different pairs should be balanced asmuch as possible. Serving order worksheets should alsoinclude the identification of the samples for each set.10.4 Prepare the response ballots in a way consistent withthe product you are evaluating. For example, in a taste

47、test,give the following instructions: (1) you will receive twosamples. They may be the same or different; (2) evaluate thesamples from left to right; and (3) determine whether they arethe same or different.10.4.1 The researcher can choose to add an instruction to theballot indicating whether the ass

48、essor may re-evaluate thesamples or not.10.4.2 The ballot should also identify the assessor and dateof test, as well as a ballot number that must be related to thesample set identification on the worksheet.10.4.3 A section soliciting comments may be includedfollowing the initial forced-choice questi

49、on.10.4.4 The example of a ballot is provided in Fig. X2.2.10.5 When possible, present both samples at the same time,along with the response ballot. In some instances, the samplesmay be presented sequentially if required by the type ofproduct or the way they need to be presented, or both.This maybe the case, for example, for the evaluation of a fragrance in aroom where the assessor must change rooms to evaluate thesecond sample.10.6 Collect all ballots and tabulate results for analysis.11. Analysis and Interpretation of Results11.1 The data from the test

copyright@ 2008-2019 麦多课文库(www.mydoc123.com)网站版权所有
备案/许可证编号:苏ICP备17064731号-1