fbpx
Wikipedia

Categorical variable

In statistics, a categorical variable (also called qualitative variable) is a variable that can take on one of a limited, and usually fixed, number of possible values, assigning each individual or other unit of observation to a particular group or nominal category on the basis of some qualitative property.[1] In computer science and some branches of mathematics, categorical variables are referred to as enumerations or enumerated types. Commonly (though not in this article), each of the possible values of a categorical variable is referred to as a level. The probability distribution associated with a random categorical variable is called a categorical distribution.

Categorical data is the statistical data type consisting of categorical variables or of data that has been converted into that form, for example as grouped data. More specifically, categorical data may derive from observations made of qualitative data that are summarised as counts or cross tabulations, or from observations of quantitative data grouped within given intervals. Often, purely categorical data are summarised in the form of a contingency table. However, particularly when considering data analysis, it is common to use the term "categorical data" to apply to data sets that, while containing some categorical variables, may also contain non-categorical variables.

A categorical variable that can take on exactly two values is termed a binary variable or a dichotomous variable; an important special case is the Bernoulli variable. Categorical variables with more than two possible values are called polytomous variables; categorical variables are often assumed to be polytomous unless otherwise specified. Discretization is treating continuous data as if it were categorical. Dichotomization is treating continuous data or polytomous variables as if they were binary variables. Regression analysis often treats category membership with one or more quantitative dummy variables.

Examples of categorical variables edit

Examples of values that might be represented in a categorical variable:

  • The roll of a six-sided die: possible outcomes are 1, 2, 3, 4, 5, or 6.
  • Demographic information of a population: gender, disease status.
  • The blood type of a person: A, B, AB or O.
  • The political party that a voter might vote for, e. g. Green Party, Christian Democrat, Social Democrat, etc.
  • The type of a rock: igneous, sedimentary or metamorphic.
  • The identity of a particular word (e.g., in a language model): One of V possible choices, for a vocabulary of size V.

Notation edit

For ease in statistical processing, categorical variables may be assigned numeric indices, e.g. 1 through K for a K-way categorical variable (i.e. a variable that can express exactly K possible values). In general, however, the numbers are arbitrary, and have no significance beyond simply providing a convenient label for a particular value. In other words, the values in a categorical variable exist on a nominal scale: they each represent a logically separate concept, cannot necessarily be meaningfully ordered, and cannot be otherwise manipulated as numbers could be. Instead, valid operations are equivalence, set membership, and other set-related operations.

As a result, the central tendency of a set of categorical variables is given by its mode; neither the mean nor the median can be defined. As an example, given a set of people, we can consider the set of categorical variables corresponding to their last names. We can consider operations such as equivalence (whether two people have the same last name), set membership (whether a person has a name in a given list), counting (how many people have a given last name), or finding the mode (which name occurs most often). However, we cannot meaningfully compute the "sum" of Smith + Johnson, or ask whether Smith is "less than" or "greater than" Johnson. As a result, we cannot meaningfully ask what the "average name" (the mean) or the "middle-most name" (the median) is in a set of names.

This ignores the concept of alphabetical order, which is a property that is not inherent in the names themselves, but in the way we construct the labels. For example, if we write the names in Cyrillic and consider the Cyrillic ordering of letters, we might get a different result of evaluating "Smith < Johnson" than if we write the names in the standard Latin alphabet; and if we write the names in Chinese characters, we cannot meaningfully evaluate "Smith < Johnson" at all, because no consistent ordering is defined for such characters. However, if we do consider the names as written, e.g., in the Latin alphabet, and define an ordering corresponding to standard alphabetical order, then we have effectively converted them into ordinal variables defined on an ordinal scale.

Number of possible values edit

Categorical random variables are normally described statistically by a categorical distribution, which allows an arbitrary K-way categorical variable to be expressed with separate probabilities specified for each of the K possible outcomes. Such multiple-category categorical variables are often analyzed using a multinomial distribution, which counts the frequency of each possible combination of numbers of occurrences of the various categories. Regression analysis on categorical outcomes is accomplished through multinomial logistic regression, multinomial probit or a related type of discrete choice model.

Categorical variables that have only two possible outcomes (e.g., "yes" vs. "no" or "success" vs. "failure") are known as binary variables (or Bernoulli variables). Because of their importance, these variables are often considered a separate category, with a separate distribution (the Bernoulli distribution) and separate regression models (logistic regression, probit regression, etc.). As a result, the term "categorical variable" is often reserved for cases with 3 or more outcomes, sometimes termed a multi-way variable in opposition to a binary variable.

It is also possible to consider categorical variables where the number of categories is not fixed in advance. As an example, for a categorical variable describing a particular word, we might not know in advance the size of the vocabulary, and we would like to allow for the possibility of encountering words that we have not already seen. Standard statistical models, such as those involving the categorical distribution and multinomial logistic regression, assume that the number of categories is known in advance, and changing the number of categories on the fly is tricky. In such cases, more advanced techniques must be used. An example is the Dirichlet process, which falls in the realm of nonparametric statistics. In such a case, it is logically assumed that an infinite number of categories exist, but at any one time most of them (in fact, all but a finite number) have never been seen. All formulas are phrased in terms of the number of categories actually seen so far rather than the (infinite) total number of potential categories in existence, and methods are created for incremental updating of statistical distributions, including adding "new" categories.

Categorical variables and regression edit

Categorical variables represent a qualitative method of scoring data (i.e. represents categories or group membership). These can be included as independent variables in a regression analysis or as dependent variables in logistic regression or probit regression, but must be converted to quantitative data in order to be able to analyze the data. One does so through the use of coding systems. Analyses are conducted such that only g -1 (g being the number of groups) are coded. This minimizes redundancy while still representing the complete data set as no additional information would be gained from coding the total g groups: for example, when coding gender (where g = 2: male and female), if we only code females everyone left over would necessarily be males. In general, the group that one does not code for is the group of least interest.[2]

There are three main coding systems typically used in the analysis of categorical variables in regression: dummy coding, effects coding, and contrast coding. The regression equation takes the form of Y = bX + a, where b is the slope and gives the weight empirically assigned to an explanator, X is the explanatory variable, and a is the Y-intercept, and these values take on different meanings based on the coding system used. The choice of coding system does not affect the F or R2 statistics. However, one chooses a coding system based on the comparison of interest since the interpretation of b values will vary.[2]

Dummy coding edit

Dummy coding is used when there is a control or comparison group in mind. One is therefore analyzing the data of one group in relation to the comparison group: a represents the mean of the control group and b is the difference between the mean of the experimental group and the mean of the control group. It is suggested that three criteria be met for specifying a suitable control group: the group should be a well-established group (e.g. should not be an "other" category), there should be a logical reason for selecting this group as a comparison (e.g. the group is anticipated to score highest on the dependent variable), and finally, the group's sample size should be substantive and not small compared to the other groups.[3]

In dummy coding, the reference group is assigned a value of 0 for each code variable, the group of interest for comparison to the reference group is assigned a value of 1 for its specified code variable, while all other groups are assigned 0 for that particular code variable.[2]

The b values should be interpreted such that the experimental group is being compared against the control group. Therefore, yielding a negative b value would entail the experimental group have scored less than the control group on the dependent variable. To illustrate this, suppose that we are measuring optimism among several nationalities and we have decided that French people would serve as a useful control. If we are comparing them against Italians, and we observe a negative b value, this would suggest Italians obtain lower optimism scores on average.

The following table is an example of dummy coding with French as the control group and C1, C2, and C3 respectively being the codes for Italian, German, and Other (neither French nor Italian nor German):

Nationality C1 C2 C3
French 0 0 0
Italian 1 0 0
German 0 1 0
Other 0 0 1

Effects coding edit

In the effects coding system, data are analyzed through comparing one group to all other groups. Unlike dummy coding, there is no control group. Rather, the comparison is being made at the mean of all groups combined (a is now the grand mean). Therefore, one is not looking for data in relation to another group but rather, one is seeking data in relation to the grand mean.[2]

Effects coding can either be weighted or unweighted. Weighted effects coding is simply calculating a weighted grand mean, thus taking into account the sample size in each variable. This is most appropriate in situations where the sample is representative of the population in question. Unweighted effects coding is most appropriate in situations where differences in sample size are the result of incidental factors. The interpretation of b is different for each: in unweighted effects coding b is the difference between the mean of the experimental group and the grand mean, whereas in the weighted situation it is the mean of the experimental group minus the weighted grand mean.[2]

In effects coding, we code the group of interest with a 1, just as we would for dummy coding. The principal difference is that we code −1 for the group we are least interested in. Since we continue to use a g - 1 coding scheme, it is in fact the −1 coded group that will not produce data, hence the fact that we are least interested in that group. A code of 0 is assigned to all other groups.

The b values should be interpreted such that the experimental group is being compared against the mean of all groups combined (or weighted grand mean in the case of weighted effects coding). Therefore, yielding a negative b value would entail the coded group as having scored less than the mean of all groups on the dependent variable. Using our previous example of optimism scores among nationalities, if the group of interest is Italians, observing a negative b value suggest they obtain a lower optimism score.

The following table is an example of effects coding with Other as the group of least interest.

Nationality C1 C2 C3
French 0 0 1
Italian 1 0 0
German 0 1 0
Other −1 −1 −1

Contrast coding edit

The contrast coding system allows a researcher to directly ask specific questions. Rather than having the coding system dictate the comparison being made (i.e., against a control group as in dummy coding, or against all groups as in effects coding) one can design a unique comparison catering to one's specific research question. This tailored hypothesis is generally based on previous theory and/or research. The hypotheses proposed are generally as follows: first, there is the central hypothesis which postulates a large difference between two sets of groups; the second hypothesis suggests that within each set, the differences among the groups are small. Through its a priori focused hypotheses, contrast coding may yield an increase in power of the statistical test when compared with the less directed previous coding systems.[2]

Certain differences emerge when we compare our a priori coefficients between ANOVA and regression. Unlike when used in ANOVA, where it is at the researcher's discretion whether they choose coefficient values that are either orthogonal or non-orthogonal, in regression, it is essential that the coefficient values assigned in contrast coding be orthogonal. Furthermore, in regression, coefficient values must be either in fractional or decimal form. They cannot take on interval values.

The construction of contrast codes is restricted by three rules:

  1. The sum of the contrast coefficients per each code variable must equal zero.
  2. The difference between the sum of the positive coefficients and the sum of the negative coefficients should equal 1.
  3. Coded variables should be orthogonal.[2]

Violating rule 2 produces accurate R2 and F values, indicating that we would reach the same conclusions about whether or not there is a significant difference; however, we can no longer interpret the b values as a mean difference.

To illustrate the construction of contrast codes consider the following table. Coefficients were chosen to illustrate our a priori hypotheses: Hypothesis 1: French and Italian persons will score higher on optimism than Germans (French = +0.33, Italian = +0.33, German = −0.66). This is illustrated through assigning the same coefficient to the French and Italian categories and a different one to the Germans. The signs assigned indicate the direction of the relationship (hence giving Germans a negative sign is indicative of their lower hypothesized optimism scores). Hypothesis 2: French and Italians are expected to differ on their optimism scores (French = +0.50, Italian = −0.50, German = 0). Here, assigning a zero value to Germans demonstrates their non-inclusion in the analysis of this hypothesis. Again, the signs assigned are indicative of the proposed relationship.

Nationality C1 C2
French +0.33 +0.50
Italian +0.33 −0.50
German −0.66 0

Nonsense coding edit

Nonsense coding occurs when one uses arbitrary values in place of the designated "0"s "1"s and "-1"s seen in the previous coding systems. Although it produces correct mean values for the variables, the use of nonsense coding is not recommended as it will lead to uninterpretable statistical results.[2]

Embeddings edit

Embeddings are codings of categorical values into low-dimensional real-valued (sometimes complex-valued) vector spaces, usually in such a way that ‘similar’ values are assigned ‘similar’ vectors, or with respect to some other kind of criterion making the vectors useful for the respective application. A common special case are word embeddings, where the possible values of the categorical variable are the words in a language and words with similar meanings are to be assigned similar vectors.

Interactions edit

An interaction may arise when considering the relationship among three or more variables, and describes a situation in which the simultaneous influence of two variables on a third is not additive. Interactions may arise with categorical variables in two ways: either categorical by categorical variable interactions, or categorical by continuous variable interactions.

Categorical by categorical variable interactions edit

This type of interaction arises when we have two categorical variables. In order to probe this type of interaction, one would code using the system that addresses the researcher's hypothesis most appropriately. The product of the codes yields the interaction. One may then calculate the b value and determine whether the interaction is significant.[2]

Categorical by continuous variable interactions edit

Simple slopes analysis is a common post hoc test used in regression which is similar to the simple effects analysis in ANOVA, used to analyze interactions. In this test, we are examining the simple slopes of one independent variable at specific values of the other independent variable. Such a test is not limited to use with continuous variables, but may also be employed when the independent variable is categorical. We cannot simply choose values to probe the interaction as we would in the continuous variable case because of the nominal nature of the data (i.e., in the continuous case, one could analyze the data at high, moderate, and low levels assigning 1 standard deviation above the mean, at the mean, and at one standard deviation below the mean respectively). In our categorical case we would use a simple regression equation for each group to investigate the simple slopes. It is common practice to standardize or center variables to make the data more interpretable in simple slopes analysis; however, categorical variables should never be standardized or centered. This test can be used with all coding systems.[2]

See also edit

References edit

  1. ^ Yates, Daniel S.; Moore, David S.; Starnes, Daren S. (2003). (2nd ed.). New York: Freeman. ISBN 978-0-7167-4773-4. Archived from the original on 2005-02-09. Retrieved 2014-09-28.
  2. ^ a b c d e f g h i j Cohen, J.; Cohen, P.; West, S. G.; Aiken, L. S. (2003). Applied multiple regression/correlation analysis for the behavioural sciences (3rd ed.). New York, NY: Routledge.
  3. ^ Hardy, Melissa (1993). Regression with dummy variables. Newbury Park, CA: Sage.

Further reading edit

  • Andersen, Erling B. 1980. Discrete Statistical Models with Social Science Applications. North Holland, 1980.
  • Bishop, Y. M. M.; Fienberg, S. E.; Holland, P. W. (1975). Discrete Multivariate Analysis: Theory and Practice. MIT Press. ISBN 978-0-262-02113-5. MR 0381130.
  • Christensen, Ronald (1997). Log-linear models and logistic regression. Springer Texts in Statistics (Second ed.). New York: Springer-Verlag. pp. xvi+483. ISBN 0-387-98247-7. MR 1633357.
  • Friendly, Michael. Visualizing categorical data. SAS Institute, 2000.
  • Lauritzen, Steffen L. (2002) [1979]. Lectures on Contingency Tables (PDF) (updated electronic version of the (University of Aalborg) 3rd (1989) ed.).
  • NIST/SEMATEK (2008) Handbook of Statistical Methods

categorical, variable, statistics, categorical, variable, also, called, qualitative, variable, variable, that, take, limited, usually, fixed, number, possible, values, assigning, each, individual, other, unit, observation, particular, group, nominal, category,. In statistics a categorical variable also called qualitative variable is a variable that can take on one of a limited and usually fixed number of possible values assigning each individual or other unit of observation to a particular group or nominal category on the basis of some qualitative property 1 In computer science and some branches of mathematics categorical variables are referred to as enumerations or enumerated types Commonly though not in this article each of the possible values of a categorical variable is referred to as a level The probability distribution associated with a random categorical variable is called a categorical distribution Categorical data is the statistical data type consisting of categorical variables or of data that has been converted into that form for example as grouped data More specifically categorical data may derive from observations made of qualitative data that are summarised as counts or cross tabulations or from observations of quantitative data grouped within given intervals Often purely categorical data are summarised in the form of a contingency table However particularly when considering data analysis it is common to use the term categorical data to apply to data sets that while containing some categorical variables may also contain non categorical variables A categorical variable that can take on exactly two values is termed a binary variable or a dichotomous variable an important special case is the Bernoulli variable Categorical variables with more than two possible values are called polytomous variables categorical variables are often assumed to be polytomous unless otherwise specified Discretization is treating continuous data as if it were categorical Dichotomization is treating continuous data or polytomous variables as if they were binary variables Regression analysis often treats category membership with one or more quantitative dummy variables Contents 1 Examples of categorical variables 2 Notation 3 Number of possible values 4 Categorical variables and regression 4 1 Dummy coding 4 2 Effects coding 4 3 Contrast coding 4 4 Nonsense coding 4 5 Embeddings 4 6 Interactions 4 6 1 Categorical by categorical variable interactions 4 6 2 Categorical by continuous variable interactions 5 See also 6 References 7 Further readingExamples of categorical variables editExamples of values that might be represented in a categorical variable The roll of a six sided die possible outcomes are 1 2 3 4 5 or 6 Demographic information of a population gender disease status The blood type of a person A B AB or O The political party that a voter might vote for e g Green Party Christian Democrat Social Democrat etc The type of a rock igneous sedimentary or metamorphic The identity of a particular word e g in a language model One of V possible choices for a vocabulary of size V Notation editFor ease in statistical processing categorical variables may be assigned numeric indices e g 1 through K for a K way categorical variable i e a variable that can express exactly K possible values In general however the numbers are arbitrary and have no significance beyond simply providing a convenient label for a particular value In other words the values in a categorical variable exist on a nominal scale they each represent a logically separate concept cannot necessarily be meaningfully ordered and cannot be otherwise manipulated as numbers could be Instead valid operations are equivalence set membership and other set related operations As a result the central tendency of a set of categorical variables is given by its mode neither the mean nor the median can be defined As an example given a set of people we can consider the set of categorical variables corresponding to their last names We can consider operations such as equivalence whether two people have the same last name set membership whether a person has a name in a given list counting how many people have a given last name or finding the mode which name occurs most often However we cannot meaningfully compute the sum of Smith Johnson or ask whether Smith is less than or greater than Johnson As a result we cannot meaningfully ask what the average name the mean or the middle most name the median is in a set of names This ignores the concept of alphabetical order which is a property that is not inherent in the names themselves but in the way we construct the labels For example if we write the names in Cyrillic and consider the Cyrillic ordering of letters we might get a different result of evaluating Smith lt Johnson than if we write the names in the standard Latin alphabet and if we write the names in Chinese characters we cannot meaningfully evaluate Smith lt Johnson at all because no consistent ordering is defined for such characters However if we do consider the names as written e g in the Latin alphabet and define an ordering corresponding to standard alphabetical order then we have effectively converted them into ordinal variables defined on an ordinal scale Number of possible values editCategorical random variables are normally described statistically by a categorical distribution which allows an arbitrary K way categorical variable to be expressed with separate probabilities specified for each of the K possible outcomes Such multiple category categorical variables are often analyzed using a multinomial distribution which counts the frequency of each possible combination of numbers of occurrences of the various categories Regression analysis on categorical outcomes is accomplished through multinomial logistic regression multinomial probit or a related type of discrete choice model Categorical variables that have only two possible outcomes e g yes vs no or success vs failure are known as binary variables or Bernoulli variables Because of their importance these variables are often considered a separate category with a separate distribution the Bernoulli distribution and separate regression models logistic regression probit regression etc As a result the term categorical variable is often reserved for cases with 3 or more outcomes sometimes termed a multi way variable in opposition to a binary variable It is also possible to consider categorical variables where the number of categories is not fixed in advance As an example for a categorical variable describing a particular word we might not know in advance the size of the vocabulary and we would like to allow for the possibility of encountering words that we have not already seen Standard statistical models such as those involving the categorical distribution and multinomial logistic regression assume that the number of categories is known in advance and changing the number of categories on the fly is tricky In such cases more advanced techniques must be used An example is the Dirichlet process which falls in the realm of nonparametric statistics In such a case it is logically assumed that an infinite number of categories exist but at any one time most of them in fact all but a finite number have never been seen All formulas are phrased in terms of the number of categories actually seen so far rather than the infinite total number of potential categories in existence and methods are created for incremental updating of statistical distributions including adding new categories Categorical variables and regression editCategorical variables represent a qualitative method of scoring data i e represents categories or group membership These can be included as independent variables in a regression analysis or as dependent variables in logistic regression or probit regression but must be converted to quantitative data in order to be able to analyze the data One does so through the use of coding systems Analyses are conducted such that only g 1 g being the number of groups are coded This minimizes redundancy while still representing the complete data set as no additional information would be gained from coding the total g groups for example when coding gender where g 2 male and female if we only code females everyone left over would necessarily be males In general the group that one does not code for is the group of least interest 2 There are three main coding systems typically used in the analysis of categorical variables in regression dummy coding effects coding and contrast coding The regression equation takes the form of Y bX a where b is the slope and gives the weight empirically assigned to an explanator X is the explanatory variable and a is the Y intercept and these values take on different meanings based on the coding system used The choice of coding system does not affect the F or R2 statistics However one chooses a coding system based on the comparison of interest since the interpretation of b values will vary 2 Dummy coding edit Dummy coding is used when there is a control or comparison group in mind One is therefore analyzing the data of one group in relation to the comparison group a represents the mean of the control group and b is the difference between the mean of the experimental group and the mean of the control group It is suggested that three criteria be met for specifying a suitable control group the group should be a well established group e g should not be an other category there should be a logical reason for selecting this group as a comparison e g the group is anticipated to score highest on the dependent variable and finally the group s sample size should be substantive and not small compared to the other groups 3 In dummy coding the reference group is assigned a value of 0 for each code variable the group of interest for comparison to the reference group is assigned a value of 1 for its specified code variable while all other groups are assigned 0 for that particular code variable 2 The b values should be interpreted such that the experimental group is being compared against the control group Therefore yielding a negative b value would entail the experimental group have scored less than the control group on the dependent variable To illustrate this suppose that we are measuring optimism among several nationalities and we have decided that French people would serve as a useful control If we are comparing them against Italians and we observe a negative b value this would suggest Italians obtain lower optimism scores on average The following table is an example of dummy coding with French as the control group and C1 C2 and C3 respectively being the codes for Italian German and Other neither French nor Italian nor German Nationality C1 C2 C3French 0 0 0Italian 1 0 0German 0 1 0Other 0 0 1Effects coding edit In the effects coding system data are analyzed through comparing one group to all other groups Unlike dummy coding there is no control group Rather the comparison is being made at the mean of all groups combined a is now the grand mean Therefore one is not looking for data in relation to another group but rather one is seeking data in relation to the grand mean 2 Effects coding can either be weighted or unweighted Weighted effects coding is simply calculating a weighted grand mean thus taking into account the sample size in each variable This is most appropriate in situations where the sample is representative of the population in question Unweighted effects coding is most appropriate in situations where differences in sample size are the result of incidental factors The interpretation of b is different for each in unweighted effects coding b is the difference between the mean of the experimental group and the grand mean whereas in the weighted situation it is the mean of the experimental group minus the weighted grand mean 2 In effects coding we code the group of interest with a 1 just as we would for dummy coding The principal difference is that we code 1 for the group we are least interested in Since we continue to use a g 1 coding scheme it is in fact the 1 coded group that will not produce data hence the fact that we are least interested in that group A code of 0 is assigned to all other groups The b values should be interpreted such that the experimental group is being compared against the mean of all groups combined or weighted grand mean in the case of weighted effects coding Therefore yielding a negative b value would entail the coded group as having scored less than the mean of all groups on the dependent variable Using our previous example of optimism scores among nationalities if the group of interest is Italians observing a negative b value suggest they obtain a lower optimism score The following table is an example of effects coding with Other as the group of least interest Nationality C1 C2 C3French 0 0 1Italian 1 0 0German 0 1 0Other 1 1 1Contrast coding edit The contrast coding system allows a researcher to directly ask specific questions Rather than having the coding system dictate the comparison being made i e against a control group as in dummy coding or against all groups as in effects coding one can design a unique comparison catering to one s specific research question This tailored hypothesis is generally based on previous theory and or research The hypotheses proposed are generally as follows first there is the central hypothesis which postulates a large difference between two sets of groups the second hypothesis suggests that within each set the differences among the groups are small Through its a priori focused hypotheses contrast coding may yield an increase in power of the statistical test when compared with the less directed previous coding systems 2 Certain differences emerge when we compare our a priori coefficients between ANOVA and regression Unlike when used in ANOVA where it is at the researcher s discretion whether they choose coefficient values that are either orthogonal or non orthogonal in regression it is essential that the coefficient values assigned in contrast coding be orthogonal Furthermore in regression coefficient values must be either in fractional or decimal form They cannot take on interval values The construction of contrast codes is restricted by three rules The sum of the contrast coefficients per each code variable must equal zero The difference between the sum of the positive coefficients and the sum of the negative coefficients should equal 1 Coded variables should be orthogonal 2 Violating rule 2 produces accurate R2 and F values indicating that we would reach the same conclusions about whether or not there is a significant difference however we can no longer interpret the b values as a mean difference To illustrate the construction of contrast codes consider the following table Coefficients were chosen to illustrate our a priori hypotheses Hypothesis 1 French and Italian persons will score higher on optimism than Germans French 0 33 Italian 0 33 German 0 66 This is illustrated through assigning the same coefficient to the French and Italian categories and a different one to the Germans The signs assigned indicate the direction of the relationship hence giving Germans a negative sign is indicative of their lower hypothesized optimism scores Hypothesis 2 French and Italians are expected to differ on their optimism scores French 0 50 Italian 0 50 German 0 Here assigning a zero value to Germans demonstrates their non inclusion in the analysis of this hypothesis Again the signs assigned are indicative of the proposed relationship Nationality C1 C2French 0 33 0 50Italian 0 33 0 50German 0 66 0Nonsense coding edit Nonsense coding occurs when one uses arbitrary values in place of the designated 0 s 1 s and 1 s seen in the previous coding systems Although it produces correct mean values for the variables the use of nonsense coding is not recommended as it will lead to uninterpretable statistical results 2 Embeddings edit Embeddings are codings of categorical values into low dimensional real valued sometimes complex valued vector spaces usually in such a way that similar values are assigned similar vectors or with respect to some other kind of criterion making the vectors useful for the respective application A common special case are word embeddings where the possible values of the categorical variable are the words in a language and words with similar meanings are to be assigned similar vectors Interactions edit An interaction may arise when considering the relationship among three or more variables and describes a situation in which the simultaneous influence of two variables on a third is not additive Interactions may arise with categorical variables in two ways either categorical by categorical variable interactions or categorical by continuous variable interactions Categorical by categorical variable interactions edit This type of interaction arises when we have two categorical variables In order to probe this type of interaction one would code using the system that addresses the researcher s hypothesis most appropriately The product of the codes yields the interaction One may then calculate the b value and determine whether the interaction is significant 2 Categorical by continuous variable interactions edit Simple slopes analysis is a common post hoc test used in regression which is similar to the simple effects analysis in ANOVA used to analyze interactions In this test we are examining the simple slopes of one independent variable at specific values of the other independent variable Such a test is not limited to use with continuous variables but may also be employed when the independent variable is categorical We cannot simply choose values to probe the interaction as we would in the continuous variable case because of the nominal nature of the data i e in the continuous case one could analyze the data at high moderate and low levels assigning 1 standard deviation above the mean at the mean and at one standard deviation below the mean respectively In our categorical case we would use a simple regression equation for each group to investigate the simple slopes It is common practice to standardize or center variables to make the data more interpretable in simple slopes analysis however categorical variables should never be standardized or centered This test can be used with all coding systems 2 See also editLevel of measurement List of analyses of categorical data Qualitative data Statistical data type One hot encodingReferences edit Yates Daniel S Moore David S Starnes Daren S 2003 The Practice of Statistics 2nd ed New York Freeman ISBN 978 0 7167 4773 4 Archived from the original on 2005 02 09 Retrieved 2014 09 28 a b c d e f g h i j Cohen J Cohen P West S G Aiken L S 2003 Applied multiple regression correlation analysis for the behavioural sciences 3rd ed New York NY Routledge Hardy Melissa 1993 Regression with dummy variables Newbury Park CA Sage Further reading editAndersen Erling B 1980 Discrete Statistical Models with Social Science Applications North Holland 1980 Bishop Y M M Fienberg S E Holland P W 1975 Discrete Multivariate Analysis Theory and Practice MIT Press ISBN 978 0 262 02113 5 MR 0381130 Christensen Ronald 1997 Log linear models and logistic regression Springer Texts in Statistics Second ed New York Springer Verlag pp xvi 483 ISBN 0 387 98247 7 MR 1633357 Friendly Michael Visualizing categorical data SAS Institute 2000 Lauritzen Steffen L 2002 1979 Lectures on Contingency Tables PDF updated electronic version of the University of Aalborg 3rd 1989 ed NIST SEMATEK 2008 Handbook of Statistical Methods Retrieved from https en wikipedia org w index php title Categorical variable amp oldid 1215520781, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.