• Shuffle
    Toggle On
    Toggle Off
  • Alphabetize
    Toggle On
    Toggle Off
  • Front First
    Toggle On
    Toggle Off
  • Both Sides
    Toggle On
    Toggle Off
  • Read
    Toggle On
    Toggle Off
Reading...
Front

Card Range To Study

through

image

Play button

image

Play button

image

Progress

1/42

Click to flip

Use LEFT and RIGHT arrow keys to navigate between flashcards;

Use UP and DOWN arrow keys to flip the card;

H to show hint;

A reads text to speech;

42 Cards in this Set

  • Front
  • Back
Objective of Descriptive stats
tabulation of data
central tendendy, variablilty "dispersion"
Frequency
number of cases in the class interval
Cumulative Frequency
number of cases in that class interval or lower
Proportion
number of cases in that class interval divided by the total number of cases (n)
-more useful to know and requires less info to interpret
Cumulative proportion
number of cases in that class interval or lower divided by the total number of cases (n)
Cumulative percent
cumulative porportion X 100
(Percent = proportion X100)
Steps to determine XXX frequency?
1. range +1
2. determine interval width : i (range/N - chose somthing evenly divisible)
3. use the lowest score in the distribution as the min value in the lowest interval. add i to each interval and work up
4. tally scores
5. Add tallies for interval frequency

review slides
Why make grouped freq. dis?
allow graphical displays
Frequency Histograms (bars touch)
x-axis = score values for grouped or ungrouped data
y-axis = frequency (number of cases)

*interval widths on the x-axis includes true limits and midpoint is marked

*include labels!

*nominal data (ie. religions) bars don't touch!
Frequency Polygons (dots and lines)
x-axis = midpoints plotted of each interval and connect the dots

Cumulative Frequency Polygons (dots and lines going up)
instead of graphing straight frequency, cumulative frequency is graphed (number of cases in that intercal or lower)

x-axis = true upper limit of each interval
y-axis = frenqency
Stem and Leaf Plots
Combines useful feature of histograms and frequency disributions

stems = class interval
leafs = specific scores within the intercal ordered from lowest to highest

*rotate 90 degrees and you have a histrogram

*preserves raw data (great!)
intraocular inspection
visual analysis
mean
sum of all scored divided by number of scores (n, sample size)

using a freq. dis. multiply each freq dis. add them up and divide by n

drawback: heavily influenced by extreme scores, instead use median
median
divides distribution into 2 equal halves

tie scores: if median lies between intervals with tied scores, take limits lower interval limit + number below the apparent median (if falls on a score count as .5) divided by the total amount of repeating scores
mode
most frequent score in the distribution
useful measure
*can have multiple modes (bi-modal/trimodal) if there are less frequent numbers between the two modes - judgement call needed at times or if two numbers (beside each other) are the same ie.7,8 both are the modes
when to use which measures of central tendency
bimodal distributions, use mode
one modal distribution, use median
no extremes, use mean
if symmetric distribution, mean=median=mode (essentially equal) but use mean because mean uses all info
+/- skews measures of distribution
-positively skewed (mode to the leeft), mean > median and no longer representative of central tendency
-negatively skewed (mode peaks to the right), mean < median and no longer representative of central tendency
measures of variability
provide a quantitive measure of the degree to which scores in a distribution are spread out or clustered together
range (whole numbers)
(highest-lowest+1)

(with exact measures just use highest-lowest score)

as a statistic:

you cannot observe in a sample a range that is greater than the population range
*sample range will average out to be lower than pop range, not a good characteristic (bias low)
variation, variance, and standard deviation notation
variation - SSx (has to be positive, if negative, you're wrong)
variance - pop: o2(sigma squared) sample: s2
standard - pop: (sigma) sample: s
variance
average variation
SSx/N
variation (if neg answer, wrong)
SUMxi2
standard d
square-root of the variance (ssx/N)
sample variability computations
variance: ssx/(n-1)

n-1 because we lost some of the variability because we used the sample mean to calculate the variation
comparison/transfer
see slides
z-scores
re-expresses raw scores in units of sd
-number represents units of standard deviations

*requires knowing the parameters of mew and sigma



properties:
-useful for interval scales
-mean always 0
-sd of a set of z-scores is 1
-variance of a set of z-scores is 1
-sum of squares (SSx) of a set of z-scores = N(pop) or n-i (sample)
statistic vs parameter
-both are a number that calculate data that quantifies a characteristic of the sample, but STATISTICS ARE FOR SAMPLES
AND PARAMETERS ARE FOR POPS
x-axis
abscissa
y-axis
yordinate
continuous data
use limits
SD scores
smaller = closer to the mean
Averaging several means
meanxn +mean2xn2 / n + n2
inferential stats
making inferences about a pop based on a sample set of numbers
sum of scores quantity scored
(x1+x2+x3)2
sum of squared raw scores
x12+x22+x32
range and number of possible scores
range: highest - lowest
number of possible scores: highest-lowest +1
transformations if a constant (k) is added/subtracted the...
MEAN of the transformed scores is changed by the value of the constant.!
(SD and variance dont change)
transformation: If every score in a distribution is
multiplied by a constant (k), the
the mean of the transformed scores becomes the product of the old mean and the constant.!

SD of the transformed scores is the product of the original standard deviation x the constant.
The variance of the
transformed scores is the product of the original variance x the square of the
constant (i.e., k2).!
semi interquartile range
q3-q1/2
the purpose of the finite sample correction for the standard deviation and
variance statistic
We use (n-1) to correct for the variation that is present in a sample that is, on
average, too low when we calculate variation around an estimated sample mean.
rules for transformations
divide by a constant, add to equal group A varibale

SD by same constant