Top Qs
Timeline
Chat
Perspective
Cronbach's alpha
Statistical measure of reliability From Wikipedia, the free encyclopedia
Remove ads
Cronbach's alpha (Cronbach's ) or coefficient alpha (coefficient ), is a reliability coefficient and a measure of the internal consistency of tests and measures.[1][2] It was devised by the American psychometrician Lee Cronbach. Today it enjoys such wide-spread usage that numerous studies warn against using Cronbach's alpha uncritically.[3][4][5][6][7]
Remove ads
History
Summarize
Perspective
In his initial 1951 publication, Lee Cronbach described the coefficient as Coefficient alpha[1] and included an additional derivation.[8] Coefficient alpha had been used implicitly in previous studies,[9][10][11][12] but his interpretation was thought to be more intuitively attractive relative to previous studies and it became quite popular.[13]
- In 1967, Melvin Novick and Charles Lewis proved that it was equal to reliability if the true scores[i] of the compared tests or measures vary by a constant, which is independent of the people measured. In this case, the tests or measurements were said to be "essentially tau-equivalent."[14]
- In 1978, Cronbach asserted that the reason the initial 1951 publication was widely cited was "mostly because [he] put a brand name on a common-place coefficient."[2]: 263 He explained that he had originally planned to name other types of reliability coefficients, such as those used in inter-rater reliability and test-retest reliability, after consecutive Greek letters (i.e., , , etc.), but later changed his mind.
- Later, in 2004, Cronbach and Richard Shavelson encouraged readers to use generalizability theory rather than . Cronbach opposed the use of the name "Cronbach's alpha" and explicitly denied the existence of studies that had published the general formula of KR-20 before Cronbach's 1951 publication of the same name.[7]
Remove ads
Prerequisites for using Cronbach's alpha
To use Cronbach's alpha as an accurate estimate of reliability, the following conditions must be met:[15][16]
- The "parts" (i.e. items, test parts, etc.) must be essentially tau-equivalent;
- Errors in the measurements are independent.
However, under the definition of CTT, the errors are defined to be independent.
This is often a source of confusion for users who might consider some aspect of the testing process to be an "error" (rater biases, examinee collusion, self-report faking). Anything that increases the covariance among the parts will contribute to greater true score variance. Under such circumstances, alpha is likely to over-estimate the reliability intended by the user.[16][3]
Remove ads
Formula and calculation
Summarize
Perspective
Reliability can be defined as one minus the error score variance divided by the observed score variance:
Cronbach's alpha is best understand as a direct estimate of this definitional formula with error score variance estimated as the sum of the variances of each "part" (e.g., items or testlets):[6]
where:
- represents the number of "parts" (items, test parts, etc.) in the measure;
- the term causes alpha to be an unbiased estimate of reliability when the parts are parallel or essentially tau equivalent;
- the variance associated with each part i; and
- the observed score variance (the variance associated with the total test scores).
The reason that the sum of the individual part variances estimates the error score variance is because
and the variance of a composite is equal to twice the sum of all covariances of the parts plus the sum of the variances of the parts: . Therefore estimates and estimates . It is much easier to compute alpha by summing the part variances (to estimate error score variance) than adding up all the unique part covariances (to estimate true score variance) .
Alternatively, alpha can be calculated through the following formula:[17]
where:
- represents the average variance
- represents the average inter-item covariance (or the average covariance of "parts").
Remove ads
Common misconceptions
Summarize
Perspective
Application of Cronbach's alpha is not always straightforward and can give rise to common misconceptions.
A high value of Cronbach's alpha indicates homogeneity between the items
Many textbooks refer to as an indicator of homogeneity[18] between items. This misconception stems from the inaccurate explanation of Cronbach (1951)[8] that high values show homogeneity between the items. Homogeneity is a term that is rarely used in modern literature, and related studies interpret the term as referring to uni-dimensionality. Several studies have provided proofs or counterexamples that high values do not indicate uni-dimensionality.[19][20][21][22][23] See counterexamples below.
in the uni-dimensional data above.
in the multidimensional data above.
The above data have , but are multidimensional.
The above data have , but are uni-dimensional.
Uni-dimensionality is a prerequisite for . One should check uni-dimensionality before calculating rather than calculating to check uni-dimensionality.
A high value of Cronbach's alpha indicates internal consistency
The term "internal consistency" is commonly used in the reliability literature, but its meaning is not clearly defined. The term is sometimes used to refer to a certain kind of reliability (e.g., internal consistency reliability), but it is unclear exactly which reliability coefficients are included here, in addition to . Cronbach (1951)[8] used the term in several senses without an explicit definition. Cortina (1993)[19] showed that is not an indicator of any of these.
Removing items using "alpha if item deleted" always increases reliability
Most psychometric software will produce a column labeled "alpha if item deleted" which is the coefficient alpha that would be obtained if an item were to be dropped. For good items, this value is lower than the current coefficient alpha for the whole scale. But for some weak or bad items, the "alpha if item deleted" value shows an increase over the current current coefficient alpha for the whole scale.
Removing an item using "alpha if item deleted" may result in 'alpha inflation,' where sample-level reliability is reported to be higher than population-level reliability.[24] It may also reduce population-level reliability.[25] The elimination of less-reliable items should be based not only on a statistical basis but also on a theoretical and logical basis. It is also recommended that the whole sample be divided into two and cross-validated.[24]
Remove ads
Notes
- The true score is the difference between the score observed during the test or measurement and the error in that observation. See classical test theory for further information.
References
External links
Wikiwand - on
Seamless Wikipedia browsing. On steroids.
Remove ads