Use LEFT and RIGHT arrow keys to navigate between flashcards;
Use UP and DOWN arrow keys to flip the card;
H to show hint;
A reads text to speech;
41 Cards in this Set
- Front
- Back
Nominal Data
|
Numbers serve ONLY as LABELS for classifying data, CATEGORICAL data
-MODE |
|
Ordinal Data
|
-Numbers have numerical significance but REVEAL no STRENGTH of DIFFERENCE, CATEGORICAL data
-MODE MEDIAN ex rank your fav soda |
|
Interval data
|
-Distance between numbers represent EQUAL VALUES
-MEAN MODE RANGE MEDIAN ex income ranges |
|
Ratio Data
|
-BEST KIND
-Has absolute zero point -difficult to collect -METRIC data -EVERYthing plus st dev equality of ratios ex income in exact $ |
|
Hypothesis Ho & H1
|
Ho- Null Hypothesis, Statement of no effect/difference
ex.age has no effect on water intake H1-Alternative, some difference/effect is expected ex younger consumers consume statistically more water than older |
|
Type I and Type II errors
|
Type I- False positive
ex.saying age has effect when it doesn't (WORST ERROR) type II-False negative ex.saying age doesn't make a difference when it does |
|
Dependent vs Independent Variables in crosstabs
|
Dep-What your trying to predict (ROWS) crosstabs
Indep-What you use to predict, demographics? (COLUMNS) crosstabs |
|
Crosstabs overall significance
|
-cross classifies one variable against another
-chi squared value gives overall significance (.02=98%) |
|
Degrees of freedom in Chi square
|
(row-1)(column-1)=df
|
|
Crosstabs adjusted residual value
|
-Adjusted residual must be >2 or < 2 to be significant
-if not high enough, increase sample size |
|
chi squared x^2 value
|
If X^2 is > or = to the critical value, we REJECT THE NULL, there is a RELATIONSHIP BETWEEN VARIABLE
-ORDINAL data or less |
|
T test basics
|
-tests for a significant difference between means
-If SD(variance) goes up, significance goes down -if sample size (N) goes down, significance goes down -t tests account for VARIABILITY and SAMPLE SIZE -INTERVAL |
|
One sample t test
|
-tests mean of a single variable against a given value
-if T value is > than critical value, its SIGNIFICANT (REJECT THE NULL, introduce product) |
|
Independent sample t test
|
-compares differences in the means of 2 independent groups(ROWS in SPSS)
-comparing average responses ACROSS PEOPLE |
|
Degrees of freedom calculation in T test
|
N1-N2-K=df
K=2 (ALWAYS) N1-group 1 sample size |
|
Paired sample T test
|
-compares mean differences in 2 variables
-comparing average responses to questions from SAME RESPONDANTS |
|
Lavene's test in independent sample t test
-Did we violate the constant variance assumption? |
-tests whether groups have similar variance
-assumption 1.interval data 2.constant variance 3.normal distribution |
|
Levenes test significance?
|
if p is > or = to .05 read top line
if p is < or = to .05 read bottom line (its SIGNIFICANT) P is SIG in levenes table in SPSS |
|
Anova basics
|
-tests for significant difference between the means of MORE THAN 2 GROUPS
(t test is only 2 groups) -If F is > than critical value we Reject the null(there is SIGNIFICANCE) |
|
F test Anova
|
F=(variance BETWEEN GROUPS)/(variance WITHIN GROUPS)
-as F increases we can reject the null |
|
Bonferroni test in Anova
|
-Multiple comparisons table in SPSS, single out 2 choices and compare significane
-Must be <.05 to be considered significant (95%) |
|
Degrees of Freedom in Anova
|
(# of groups-1)/(samp size - # of groups)
Top of table side of Table |
|
Regression basics
|
-Used to try and predict/explain the variance in MULTIPLE INDEPENDENT VARIABLES
-Need interval for dependent and interval/binary for independent |
|
3 reasons to fit a regression model
|
1.predict the dependent variable for a new set of data(ex predict sales)
2.Assess the explanatory power of the independent variable(s) (ex how is price effecting sales, elasticity) 3.Identify the subset of measure that is more effective for predicting the Dependent variable |
|
Regression and correlation relationship
|
-Correlation is how much two variables change TOGETHER
r=1 means the 2 variables are perfectly correlated (moving in same direction |
|
R^2 in regression
|
-gives the explained variance of the dependent variable by the independent variable
-AKA coefficient of determination ex R^2=.80 we can explain 80% of variation in food sales (DV) is explained by coupons (IV) (not adjusted) |
|
Simple regression
|
Involves ONLY 1 Independent variable
|
|
Multiple Regression
|
-involves 2 OR MORE independent variables
|
|
Stepwise Regression
|
Enters IV into the model one at a time but also tests IV's at each step for possible removal
-Fixes Multicollinearity |
|
Multicollinearity in regression
|
-problem when IVs are highly correlated AMONG THEMSELVES
too much of the explanatory power of 2 IV's may be shared VIF value of 4 or more is to high=problematic multicollinearity |
|
4 assumptions of regression to not violate
|
1. Linearity, NOT linear curve (rainbow)
2.Independent Errors, Not straight line, same direction (arrow) 3.Constant variance, narrow to wide (shotgun blast) 4.Normality of error, line should fit on line |
|
Key driver in regression
|
-B (beta weights) highest is most important
|
|
Regression equation
|
-Coefficient table in SPSS, bottom right box
y=Bo+B1(x1)+B2(x2)+error |
|
Multi dimensional scaling
|
AKA perceptual mapping
-non metric (RANKINGS) or metric (PREFERENCE RATINGS) |
|
Disparitys in multi dimensional scaling
|
disparity-differences in computer generated distances (reproduction) and the actual input provided (reality)
|
|
R squared in multi dimensional scaling
|
R squared-proportion of variance accounted for by the MDS procedure
>.6 is considered acceptable degenerate solution-in |
|
stress in multi dimensional scaling
|
stress-proportions of disparites to fit
0-1 where 1 equals a perfect fit |
|
Degenerate solution in multi dimensional scaling
|
invalid solution due to:
1.inconsistencies in data 2.too few compared objects compared to dimensions -2 dimensions need 9 object comparisons -looks like a circle or one big cluster |
|
Cluster analysis
|
used to identity natural segments (“clusters”) in a sample/population
-clusters that are homogeneous within and heterogeneous across. -metric or nonmetric |
|
agglomerative vs divisive cluster analysis
|
Agglomerativebegin with each element as a separate cluster and merge them into successively larger clusters (BOTTOM UP)
divisivebegin with the whole set and proceed to divide it into successively smaller clusters (TOP DOWN) |
|
Euclidian distance measure in cluster analysis
|
"ordinary" distance between two points that one would measure with a ruler
|