# Odds ratio: Wikis

Note: Many of our articles have direct quotes from sources you can cite, within the Wikipedia article! This article doesn't yet, but we're working on it! See more info or our list of citable articles.

# Encyclopedia

The odds ratio [1][2][3] is a measure of effect size, describing the strength of association or non-independence between two binary data values. It is used as a descriptive statistic, and plays an important role in logistic regression. Unlike other measures of association for paired binary data such as the relative risk, the odds ratio treats the two variables being compared symmetrically, and can be estimated using some types of non-random samples.

## Definition

### Definition in terms of group-wise odds

The odds ratio is the ratio of the odds of an event occurring in one group to the odds of it occurring in another group, or to a sample-based estimate of that ratio. These groups might be men and women, an experimental group and a control group, or any other dichotomous classification. If the probabilities of the event in each of the groups are p1 (first group) and p2 (second group), then the odds ratio is:

${ p_1/(1-p_1) \over p_2/(1-p_2)}={ p_1/q_1 \over p_2/q_2}=\frac{\;p_1q_2\;}{\;p_2q_1\;},$

where qx = 1 − p x. An odds ratio of 1 indicates that the condition or event under study is equally likely to occur in both groups. An odds ratio greater than 1 indicates that the condition or event is more likely to occur in the first group. And an odds ratio less than 1 indicates that the condition or event is less likely to occur in the first group. The odds ratio must be greater than or equal to zero if it is defined. It is undefined if p2q1 equals zero.[1]

### Definition in terms of joint and conditional probabilities

The odds ratio can also be defined in terms of the joint probability distribution of two binary random variables. The joint distribution of binary random variables X and Y can be written

 Y = 1 Y = 0 X = 1 p11 p10 X = 0 p01 p00

where p11, p10, p01 and p00 are non-negative "cell probabilities" that sum to one. The odds for Y within the two subpopulations defined by X = 1 and X = 0 are defined in terms of the conditional probabilities given X:

 Y = 1 Y = 0 X = 1 p11 / (p11 + p10) p10 / (p11 + p10) X = 0 p01 / (p01 + p00) p00 / (p01 + p00)

Thus the odds ratio is

${ \dfrac{p_{11}/(p_{11}+p_{10})}{p_{10}/(p_{11}+p_{10})} \bigg / \dfrac{p_{01}/(p_{01}+p_{00})}{p_{00}/(p_{01}+p_{00})}} = \dfrac{p_{11}p_{00}}{p_{10}p_{01}}.$

The simple expression on the right, above, is easy to remember as the product of the probabilities of the "concordant cells" (X = Y) divided by the product of the probabilities of the "discordant cells" (X ≠ Y). However note that in some applications the labeling of categories as zero and one is arbitrary, so there is nothing special about concordant versus discordant values in these applications.

### Symmetry

If we had calculated the odds ratio based on the conditional probabilities given Y,

 Y = 1 Y = 0 X = 1 p11 / (p11 + p01) p10 / (p10 + p00) X = 0 p01 / (p11 + p01) p00 / (p10 + p00)

we would have gotten the same result

${ \dfrac{p_{11}/(p_{11}+p_{01})}{p_{01}/(p_{11}+p_{01})} \bigg / \dfrac{p_{10}/(p_{10}+p_{00})}{p_{00}/(p_{10}+p_{00})}} = \dfrac{p_{11}p_{00}}{p_{10}p_{01}}.$

Other measures of effect size for binary data such as the relative risk do not have this symmetry property.

### Relation to statistical independence

If X and Y are independent, their joint probabilities can be expressed in terms of their marginal probabilities px =  P(X = 1) and py =  P(Y = 1), as follows

 Y = 1 Y = 0 X = 1 pxpy px(1 − py) X = 0 (1 − px)py (1 − px)(1 − py)

In this case, the odds ratio equals one, and conversely the odds ratio can only equal one if the joint probabilities can be factored in this way. Thus the odds ratio equals one if and only if X and Y are independent.

### Recovering the cell probabilities from the odds ratio and marginal probabilities

The odds ratio is a function of the cell probabilities, and conversely, the cell probabilities can be recovered given knowledge of the odds ratio and the marginal probabilities P(X = 1) = p11 +  p10 and P(Y = 1) = p11 +  p01. If the odds ratio R differs from 1, then

$p_{11} = \frac{1 + (p_{1\cdot}+p_{\cdot 1})(R-1) - S}{2(R-1)}$

where p1• = p11 + p 10,  p•1 = p11 + p 01, and

$S = \sqrt{(1+(p_{1\cdot}+p_{\cdot 1})(R-1))^2 + 4R(1-R)p_{1\cdot}p_{\cdot 1}}.$

In the case where R = 1, we have independence, so p11 = p1•p•1.

Once we have p11, the other three cell probabilities can easily be recovered from the marginal probabilities.

## Example

Suppose that in a sample of 100 men, 90 have drunk wine in the previous week, while in a sample of 100 women only 20 have drunk wine in the same period. The odds of a man drinking wine are 90 to 10, or 9:1, while the odds of a woman drinking wine are only 20 to 80, or 1:4 = 0.25:1. The odds ratio is thus 9/0.25, or 36, showing that men are much more likely to drink wine than women. Using the above formula for the calculation yields the same result:

${ 0.9/0.1 \over 0.2/0.8}=\frac{\;0.9\times 0.8\;}{\;0.1\times 0.2\;} ={0.72 \over 0.02} = 36.$

The above example also shows how odds ratios are sometimes sensitive in stating relative positions: in this sample men are 90/20 = 4.5 times more likely to have drunk wine than women, but have 36 times the odds. The logarithm of the odds ratio, the difference of the logits of the probabilities, tempers this effect, and also makes the measure symmetric with respect to the ordering of groups. For example, using natural logarithms, an odds ratio of 36/1 maps to 3.584, and an odds ratio of 1/36 maps to −3.584.

## Statistical inference

Several approaches to statistical inference for odds ratios have been developed.

One approach to inference uses large sample approximations to the sampling distribution of the log odds ratio (the natural logarithm of the odds ratio). If we use the joint probability notation defined above, the population log odds ratio is

${\log\left(\frac{p_{11}p_{00}}{p_{01}p_{10}}\right) = \log(p_{11}) + \log(p_{00}\big) - \log(p_{10}) - \log(p_{01})}.\,$

If we observe data in the form of a contingency table

 Y = 1 Y = 0 X = 1 n11 n10 X = 0 n01 n00

then the probabilities in the joint distribution can be estimated as

 Y = 1 Y = 0 X = 1 $\hat{p}_{11}$ $\hat{p}_{10}$ X = 0 $\hat{p}_{01}$ $\hat{p}_{00}$

where p̂ = nij /  n, with n = n11 + n 10 + n01 + n 00 being the sum of all four cell counts. The sample log odds ratio is

${L=\log\left(\dfrac{\hat{p}_{11}\hat{p}_{00}}{\hat{p}_{10}\hat{p}_{01}}\right) = \log\left(\dfrac{n_{11}n_{00}}{n_{10}n_{01}}\right)}$.

The standard error for the log odds ratio is approximately

${{\rm SE} = \sqrt{\dfrac{1}{n_{11}} + \dfrac{1}{n_{10}} + \dfrac{1}{n_{01}} + \dfrac{1}{n_{00}}}}$.

This is an asymptotic approximation, and will not give a meaningful result if any of the cell counts are very small. If L is the sample log odds ratio, an approximate 95% confidence interval for the population log odds ratio is L ± 2SE. This can be mapped to exp(L − 2SE), exp(L + 2SE) to obtain a 95% confidence interval for the odds ratio. If we wish to test the hypothesis that the population odds ratio equals one, the two-sided p-value is 2P(Z< −|L|/SE), where P denotes a probability, and Z denotes a standard normal random variable.

An alternative approach to inference for odds ratios looks at the distribution of the data conditionally on the marginal frequencies of X and Y. An advantage of this approach is that the sampling distribution of the odds ratio can be expressed exactly.

## Role in logistic regression

Logistic regression is one way to generalize the odds ratio beyond two binary variables. Suppose we have a binary response variable Y and a binary predictor variable X, and in addition we have other predictor variables Z1, ..., Zp that may or may not be binary. If we use multiple logistic regression to regress Y on X, Z1, ..., Zp, then the estimated coefficient $\hat{\beta}_x$ for X is related to a conditional odds ratio. Specifically, at the population level

$\exp(\beta_x) = \frac{P(Y=1|X=1, Z_1, \ldots, Z_p)/P(Y=0|X=1, Z_1, \ldots, Z_p)}{P(Y=1|X=0, Z_1, \ldots, Z_p)/P(Y=0|X=0, Z_1, \ldots, Z_p)},$

so $\exp(\hat{\beta}_x)$ is an estimate of this conditional odds ratio. The interpretation of $\exp(\hat{\beta}_x)$ is as an estimate of the odds ratio between Y and X when the values of Z1, ..., Zp are held fixed.

## Insensitivity to the type of sampling

If the data form a "population sample", then the cell probabilities p̂ij are interpreted as the frequencies of each of the four groups in the population as defined by their X and Y values. In many settings it is impractical to obtain a population sample, so a selected sample is used. For example, we may choose to sample units with X = 1 with a given probability f, regardless of their frequency in the population (which would necessitate sampling units with X = 0 with probability 1 − f). In this situation, our data would follow the following joint probabilities:

 Y = 1 Y = 0 X = 1 fp11 / (p11 + p10) fp10(p11 + p10) X = 0 (1 − f)p01 / (p01 + p00) (1 − f)p00 / (p01 + p00)

The odds ratio p11p00 / p 01p10 for this distribution does not depend on the value of f. This shows that the odds ratio (and consequently the log odds ratio) is invariant to non-random sampling based on one of the variables being studied. Note however that the standard error of the log odds ratio does depend on the value of f. This fact is exploited in two important situations:

• Suppose it is inconvenient or impractical to obtain a population sample, but it is practical to obtain a convenience sample of units with different X values, such that within the X = 0 and X = 1 subsamples the Y values are representative of the population (i.e. they follow the correct conditional probabilities).
• Suppose the marginal distribution of one variable, say X, is very skewed. For example, if we are studying the relationship between high alcohol consumption and pancreatic cancer in the general population, the incidence of pancreatic cancer would be very low, so it would require a very large population sample to get a modest number of pancreatic cancer cases. However we could use data from hospitals to contact most or all of their pancreatic cancer patients, and then randomly sample an equal number of subjects without pancreatic cancer (this is called a "case-control study").

In both these settings, the odds ratio can be calculated from the selected sample, without biasing the results relative to what would have been obtained for a population sample.

## Use in quantitative research

Due to the widespread use of logistic regression, the odds ratio is widely used in many fields of medical and social science research. The odds ratio is commonly used in survey research, in epidemiology, and to express the results of some clinical trials, such as in case-control studies. It is often abbreviated "OR" in reports. When data from multiple surveys is combined, it will often be expressed as "pooled OR".

## Relation to relative risk

In clinical studies, as well as in some other settings, the parameter of greatest interest is often the relative risk rather than the odds ratio. The relative risk is best estimated using a population sample, but if the rare disease assumption holds, the odds ratio is a good approximation to the relative risk — the odds is p / (1 − p), so when p moves towards zero, 1 − p moves towards 1, meaning that the odds approaches the risk, and the odds ratio approaches the relative risk[4]. When the rare disease assumption does not hold, the odds ratio can overestimate the relative risk[5][6][7].

If the absolute risk in the control group is available, conversion between the two is calculated by:[5]:

$RR = \frac{OR}{1 - R_C + (R_C \times OR)}$

where:

• RR = relative risk
• OR = odds ratio
• RC = absolute risk in the control group, given as a fraction (for example: fill in 10% risk as 0.1)

## Alternative estimators of the odds ratio

The sample odds ratio n11n00 / n 10n01 is easy to calculate, and for moderate and large samples performs well as an estimator of the population odds ratio. When one or more of the cells in the contingency table can have a small value, the sample odds ratio can be biased and exhibit high variance. A number of alternative estimators of the odds ratio have been proposed to address this issue. One alternative estimator is the conditional maximum likelihood estimator, which conditions on the row and column margins when forming the likelihood to maximize (as in Fisher's exact test)[8]. Another alternative estimator is the Mantel-Haenszel estimator.

## Numerical examples

The following four contingency tables contain observed cell counts, along with the corresponding sample odds ratio (OR) and sample log odds ratio (LOR):

OR = 1, LOR = 0 OR = 1, LOR = 0 OR = 4, LOR = 1.39 OR = 0.25, LOR = −1.39
Y = 1 Y = 0 Y = 1 Y = 0 Y = 1 Y = 0 Y = 1 Y = 0
X = 1 10 10 100 100 20 10 10 20
X = 0 5 5 50 50 10 20 20 10

The following joint probability distributions contain the population cell probabilities, along with the corresponding population odds ratio (OR) and population log odds ratio (LOR):

OR = 1, LOR = 0 OR = 1, LOR = 0 OR = 16, LOR = 2.77 OR = 0.67, LOR = −0.41
Y = 1 Y = 0 Y = 1 Y = 0 Y = 1 Y = 0 Y = 1 Y = 0
X = 1 0.2 0.2 0.4 0.4 0.4 0.1 0.1 0.3
X = 0 0.3 0.3 0.1 0.1 0.1 0.4 0.2 0.4

## Worked example

Example 1: risk reduction Example 2: risk increase
Experimental group (E) Control group (C) Total (E) (C)
Events (E) EE = 15 CE = 100 115 EE = 75 CE = 100
Non-events (N) EN = 135 CN = 150 285 EN = 75 CN = 150
Total subjects (S) ES = EE + EN = 150 CS = CE + CN = 250 400 ES = 150 CS = 250
Event rate (ER) EER = EE / ES = 0.1, or 10% CER = CE / CS = 0.4, or 40% N/A EER = 0.5 (50%) CER = 0.4 (40%)
Equation Variable Abbr. Example 1 Example 2
EER − CER < 0: absolute risk reduction ARR (−)0.3, or (−)30% N/A
> 0: absolute risk increase ARI N/A 0.1, or 10%
(EER − CER) / CER < 0: relative risk reduction RRR (−)0.75, or (−)75% N/A
> 0: relative risk increase RRI N/A 0.25, or 25%
1 / (EER − CER) < 0: number needed to treat NNT (−)3.33 N/A
> 0: number needed to harm NNH N/A 10
EER / CER relative risk RR 0.25 1.25
(EE / EN) / (CE / CN) odds ratio OR 0.167 1.5
EE / (EE + CE) − EN / (EN + CN) attributable risk AR (−)0.34, or (−)34% 0.095, or 9.5%
(RR − 1) / RR attributable risk percent ARP N/A 20%
1 − RR (or 1 − OR) preventive fraction PF 0.75, or 75% N/A

## References

1. ^ Cornfield, J. "A Applications to cancer of the lung, breast, and cervix". Journal of the National Cancer Institute 11: 1269–1275.
2. ^ Mosteller, Frederick (1968). "Association and Estimation in Contingency Tables". Journal of the American Statistical Association 63 (321): 1–28.
3. ^ Edwards, A.W.F. (1963). "The measure of association in a 2x2 table". Journal of the Royal Statistical Society, Series A 126 (1): 109–114.
4. ^
5. ^ a b Zhang J, Yu KF (November 1998). "What's the relative risk? A method of correcting the odds ratio in cohort studies of common outcomes". JAMA 280 (19): 1690–1. PMID 9832001.
6. ^ Robbins AS, Chao SY, Fonseca VP (October 2002). "What's the relative risk? A method to directly estimate risk ratios in cohort studies of common outcomes". Ann Epidemiol 12 (7): 452–4. PMID 12377421.
7. ^ Nurminen, Markku. "To Use or Not to Use the Odds Ratio in Epidemiologic Analyses?". European Journal of Epidemiology 11 (4): 365–371.
8. ^ Rothman, Kenneth J.; Greenland, Sander; Lash, Timothy L. (2008). Modern Epidemiology. Lippincott Williams & Wilkins. ISBN 0781755646.

# Study guide

Up to date as of January 14, 2010

### From Wikiversity

The odds ratio [1][2][3] is a measure of effect size, describing the strength of association or non-independence between two binary data values. It is used as a descriptive statistic, and plays an important role in logistic regression.

## References

1. Cornfield, J. "A method of estimating comparative rates from clinical data. Applications to cancer of the lung, breast, and cervix". Journal of the National Cancer Institute 11: 1269–1275.
2. Mosteller, Frederick (1968). "Association and Estimation in Contingency Tables". Journal of the American Statistical Association 63: 1–28.
3. Edwards, A.W.F. (1963). "The measure of association in a 2x2 table". Journal of the Royal Statistical Society, Series A 126: 109–114.
 Run a search on Odds ratio at Wikipedia.