• Shuffle
    Toggle On
    Toggle Off
  • Alphabetize
    Toggle On
    Toggle Off
  • Front First
    Toggle On
    Toggle Off
  • Both Sides
    Toggle On
    Toggle Off
  • Read
    Toggle On
    Toggle Off
Reading...
Front

Card Range To Study

through

image

Play button

image

Play button

image

Progress

1/41

Click to flip

Use LEFT and RIGHT arrow keys to navigate between flashcards;

Use UP and DOWN arrow keys to flip the card;

H to show hint;

A reads text to speech;

41 Cards in this Set

  • Front
  • Back
Nominal Data
Numbers serve ONLY as LABELS for classifying data, CATEGORICAL data

-MODE
Ordinal Data
-Numbers have numerical significance but REVEAL no STRENGTH of DIFFERENCE, CATEGORICAL data

-MODE MEDIAN

ex rank your fav soda
Interval data
-Distance between numbers represent EQUAL VALUES

-MEAN MODE RANGE MEDIAN

ex income ranges
Ratio Data
-BEST KIND
-Has absolute zero point
-difficult to collect
-METRIC data

-EVERYthing plus st dev equality of ratios

ex income in exact $
Hypothesis Ho & H1
Ho- Null Hypothesis, Statement of no effect/difference
ex.age has no effect on water intake

H1-Alternative, some difference/effect is expected

ex younger consumers consume statistically more water than older
Type I and Type II errors
Type I- False positive
ex.saying age has effect when it doesn't (WORST ERROR)

type II-False negative
ex.saying age doesn't make a difference when it does
Dependent vs Independent Variables in crosstabs
Dep-What your trying to predict (ROWS) crosstabs

Indep-What you use to predict, demographics? (COLUMNS) crosstabs
Crosstabs overall significance
-cross classifies one variable against another

-chi squared value gives overall significance (.02=98%)
Degrees of freedom in Chi square
(row-1)(column-1)=df
Crosstabs adjusted residual value
-Adjusted residual must be >2 or < 2 to be significant

-if not high enough, increase sample size
chi squared x^2 value
If X^2 is > or = to the critical value, we REJECT THE NULL, there is a RELATIONSHIP BETWEEN VARIABLE

-ORDINAL data or less
T test basics
-tests for a significant difference between means

-If SD(variance) goes up, significance goes down
-if sample size (N) goes down, significance goes down

-t tests account for VARIABILITY and SAMPLE SIZE

-INTERVAL
One sample t test
-tests mean of a single variable against a given value

-if T value is > than critical value, its SIGNIFICANT (REJECT THE NULL, introduce product)
Independent sample t test
-compares differences in the means of 2 independent groups(ROWS in SPSS)

-comparing average responses ACROSS PEOPLE
Degrees of freedom calculation in T test
N1-N2-K=df

K=2 (ALWAYS)
N1-group 1 sample size
Paired sample T test
-compares mean differences in 2 variables

-comparing average responses to questions from SAME RESPONDANTS
Lavene's test in independent sample t test

-Did we violate the constant variance assumption?
-tests whether groups have similar variance

-assumption
1.interval data
2.constant variance
3.normal distribution
Levenes test significance?
if p is > or = to .05 read top line

if p is < or = to .05 read bottom line (its SIGNIFICANT)

P is SIG in levenes table in SPSS
Anova basics
-tests for significant difference between the means of MORE THAN 2 GROUPS
(t test is only 2 groups)

-If F is > than critical value we Reject the null(there is SIGNIFICANCE)
F test Anova
F=(variance BETWEEN GROUPS)/(variance WITHIN GROUPS)

-as F increases we can reject the null
Bonferroni test in Anova
-Multiple comparisons table in SPSS, single out 2 choices and compare significane

-Must be <.05 to be considered significant (95%)
Degrees of Freedom in Anova
(# of groups-1)/(samp size - # of groups)

Top of table

side of Table
Regression basics
-Used to try and predict/explain the variance in MULTIPLE INDEPENDENT VARIABLES

-Need interval for dependent and interval/binary for independent
3 reasons to fit a regression model
1.predict the dependent variable for a new set of data(ex predict sales)

2.Assess the explanatory power of the independent variable(s) (ex how is price effecting sales, elasticity)

3.Identify the subset of measure that is more effective for predicting the Dependent variable
Regression and correlation relationship
-Correlation is how much two variables change TOGETHER

r=1 means the 2 variables are perfectly correlated (moving in same direction
R^2 in regression
-gives the explained variance of the dependent variable by the independent variable

-AKA coefficient of determination

ex R^2=.80 we can explain 80% of variation in food sales (DV) is explained by coupons (IV) (not adjusted)
Simple regression
Involves ONLY 1 Independent variable
Multiple Regression
-involves 2 OR MORE independent variables
Stepwise Regression
Enters IV into the model one at a time but also tests IV's at each step for possible removal

-Fixes Multicollinearity
Multicollinearity in regression
-problem when IVs are highly correlated AMONG THEMSELVES

too much of the explanatory power of 2 IV's may be shared

VIF value of 4 or more is to high=problematic multicollinearity
4 assumptions of regression to not violate
1. Linearity, NOT linear curve (rainbow)

2.Independent Errors, Not straight line, same direction (arrow)

3.Constant variance, narrow to wide (shotgun blast)

4.Normality of error, line should fit on line
Key driver in regression
-B (beta weights) highest is most important
Regression equation
-Coefficient table in SPSS, bottom right box

y=Bo+B1(x1)+B2(x2)+error
Multi dimensional scaling
AKA perceptual mapping

-non metric (RANKINGS) or metric (PREFERENCE RATINGS)
Disparitys in multi dimensional scaling
disparity-differences in computer generated distances (reproduction) and the actual input provided (reality)
R squared in multi dimensional scaling
R squared-proportion of variance accounted for by the MDS procedure
>.6 is considered acceptable

degenerate solution-in
stress in multi dimensional scaling
stress-proportions of disparites to fit
0-1 where 1 equals a perfect fit
Degenerate solution in multi dimensional scaling
invalid solution due to:

1.inconsistencies in data

2.too few compared objects compared to dimensions
-2 dimensions need 9 object comparisons

-looks like a circle or one big cluster
Cluster analysis
used to identity natural segments (“clusters”) in a sample/population

-clusters that are homogeneous within and heterogeneous across.

-metric or nonmetric
agglomerative vs divisive cluster analysis
Agglomerativebegin with each element as a separate cluster and merge them into successively larger clusters (BOTTOM UP)

divisivebegin with the whole set and proceed to divide it into successively smaller clusters (TOP DOWN)
Euclidian distance measure in cluster analysis
"ordinary" distance between two points that one would measure with a ruler