• Shuffle
    Toggle On
    Toggle Off
  • Alphabetize
    Toggle On
    Toggle Off
  • Front First
    Toggle On
    Toggle Off
  • Both Sides
    Toggle On
    Toggle Off
  • Read
    Toggle On
    Toggle Off
Reading...
Front

Card Range To Study

through

image

Play button

image

Play button

image

Progress

1/76

Click to flip

Use LEFT and RIGHT arrow keys to navigate between flashcards;

Use UP and DOWN arrow keys to flip the card;

H to show hint;

A reads text to speech;

76 Cards in this Set

  • Front
  • Back
Standards of Measurement
- Mass (kg)
- Distance (m)
Data
Results of Measurements
Classifications of Data
1. Non - parametric
2. Parametric Data
Non - Parametric Data
- Does not meet the assumptions of normality.
1. NOMINAL SCALE (frequency data):
-- Mutually exclusive (one or the other) and exhaustive (option for everyone) categories
-- No qualitative differences between categories
-- Number of data points in each categories is the frequency
-- Ex: Sex (not gender), ethnicity
2. ORDINAL SCALE (rank order scale):
-- Qualitative order to the data, however, no equal differences between data points
-- Ex: tournament rankings, athletic team polls, Olympic medal
Parametric Data
- meet the assumptions of "normal" data
1. INTERVAL SCALE:
-- equal units, or intervals, of measure
-- no absolute zero point, so negative values are possible
-- Ex: temp (f or c), used to judge performances in sports.
2. RATIO SCALE:
-- Based on order
-- Points are equidistant and proportional (only if u dont have neg, which is rare)
-- Zero is the absence of value
-- Ex: time, height, weight, mass, distance
Statistics
- Mathematical technique for organizing, analyzing and presenting data for evaluation.
- In order for statistics to be acceptable, data must be:
1. Reliable
2. Valid
3. Objective
Can something be reliable, but not valid? Visa versa?
??
Reliable Data
Reproducible under similar conditions; consistency
Valid Data
- Appropriateness of the test in measuring what it is designed to measure
- Ex: 40 m dash measuring ATP/CP
Objective Data
Data is collected without bias
Variables
- Characteristics of a person, place, or thing that can assume more than one value
- Ex: Height, weight, 1.5 mile run
1. Variable Comparisons
2. Constant
3. Characteristics of Variables
4. Types of Variables
Variable Comparisons
- WITHIN subjects: An individual may score differently on the same variable over a period of time.
- BETWEEN subjects: Different individuals may score differently on the same variable.
Variable Constant
- Characteristics that can only assume one value
- Example: meters in a kilometer (1000), meters in a mile, cm in an inch, and pounds in a kg.
Characteristics of Variables
1. Continuous variable can assume any value (distance, temperature)
2. Discrete variables are limited to certain whole numbers, usually integers (number of ppl, HR)
Types of Variables
1. INDEPENDENT VARIABLE (IV): free to vary; it is permitted to exert influence over other variables and is controlled by the researcher.
- It is predicted to have an effect on the dependent variable
- "predictor variable"
2. DEPENDENT VARIABLE (DV):affected by the researchers manipulation of the IV and not "free" to vary; is the variable ebing measure for analysis
- aka "Outcome Measure"
3. INTERVENING VARIABLE: An extraneous variable that is not controlled for by the researcher and may have an unknown effect on the DV.
Theory
- Belief regarding a concept or series of related concepts
-- Not necessarily true of false
-- Produces hypotheses that can be tested via experimentation (sliding filament theory)
Hypothesis
- Prediction that can be tested to determine whether or not it is correct
-- More specific than a theory
-- After testing; the odds that a hypothesis is correct or incorrect can be stated
-- Based on ROL
-- Must be stated so that is can be statistically tested (research question)
Theories and Hypotheses
- A theory may be accepted as true if many hypotheses related to it are found to be true (by original and validation)
- Theories are usually revised many times over years, or even centuries
- Some theories are abandoned after experimentation
Null Hypothesis
(Ho)
- predicts NO difference between groups or treatments
- States that any differences that are observed are due to random errors, or chance
- Ho is the hypothesis tested by statistical analysis
- Ho and H1 are mutually exclusive, if one is true the other is false.
Research Hypothesis
(H1)
- predicts the relationships or differences btw or among groups of subjects or treatments
- H1 not tested statistically
Probability
- Statistical analysis report the probability that Ho is true
- Large probability: P > 0.05 that Ho is true, we accept Ho
- Small probability: P < 0.05 that Ho is true, we reject Ho and accept H1. (P< 0.05 means that there is a less tan 5% chance that Ho is true)
Validity Issues
1. Internal Validity
2. External Validity
Internal Validity
- Control within an experiment to determine that the results are due to the I.V. applied (i.e. Design of the study)
- A control group can be used to monitor if a "learning effect" occurs
- A familiarization trial
- Intervening variables: should be minimized
- Instrument error: failure of equipment to provide accurate data. Regular calibration is essential.
Investigation Error failure of the investigator to accurately record data.
External Validity
- The ability to generalize the results of a study to the population from which the sample was drawn.
- Needs to be random, if not is does no represent the population
Statistical Inference
- The process of generalizing from a sample to a population.
- External Validity
1. Subjects are assumed to represent the pop., so it can estimate the characteristics of the population
2. Error in predicting from a sample is inversely related to the size of the sample (large samples = smaller errors)
3. Parameter: characteristics of the entire population
4. Statistic: characteristic of a sample that is used to estimate the value of the population parameter
Misuse of Statistics
- Misleading...
1. Questionable definition of terms (most, up to, trend)
2. Non-random and/or small samples
3.Biased researchers who could profit from results (supplement companies paying for results)
4. Misuse out of their intended context (only reporting a small portion)
5. Average is misleading. You have more than half of the scores above the average, when some scores are extremely low
Rank Order Distribution
- Bigger amount of people, then you cant really use rank order really
Normal Curve
- Karl Gauss
- Referred to as Gaussian curve
- The normal distribution forms the basis of all statistical inference
- Characterized by symmetrical distribution of data about the center of the curve, not all symmetric curves are normal.
- The mean (avg) median (50th percentile) and mode (score with highest frequency) are all in the center of the curve
- Frequency of scores declines in a Predictable manner as the scores deviate farther from the center of the curve
Bell-shaped curve
- with more people you can create a normal curve
- Not all bell-shaped curves are normal, but all normal curves are bell shaped.
Bimodial Curves
- A curve with two modes, or peaks
- Major and minor modes may be identified
Skewed Curves
- Formed when a disproportionate number of subjects score toward one end of the scale
Positive Skew
- Skewed curves
- The mode of the curve is pushed to the left (lower end) and the right tail is longer and points in a positive direction
Mesokurtic
- Skewed curve
- Middle curve
- Bell shaped
Leptokurtic curve
- Skewed curve
- limited range wit most scores near the means
Platykurtic Curve
- Skewed curve
- wide range of scores with low frequencies in the midrange of the curve
Percent
- By the hundreds
- cent is latin root word for one hundred
Percentage
- part of a whole expressed in hundredths
- provides the score's relation to the total possible score
Percentile
- %ile
- The fraction, in hundredths, of the ordered scores that are < or equal to a give raw score.
- 75th percentile means that 75 % of the scores are less than or equal to that score.
- 34% of data is between the 2 lines (the mean and the first SD)
Standard Scores
- A score derived from raw data and that has a known basis for comparison such as central tendency (mean) and variability (range)
- Allow the evaluation of raw scores and the comparison of two sets of data tat have different units of measurement.
Raw Score
- data from measurements
- Dont provide as much information as the %iles.
Caution when using Percentiles
- When the error in a measure is relatively large, it would be inappropriate to report an exact %ile rank, as the score is only an approximation
- Quartiles may be used
- Ceiling Effect: there will be a lower %ile increase at the higher end of the scale for the same raw score inc in other parts of the scale.
-- Caused by the plateau of the normal curve
-- Consider this when providing grades, incentives and evaluations
Calculating Percentiles
1. Rank Order distribution
2. Simple Frequency Distributions
3. Grouped f distribution
Rank Order Distribution
- What % of scores falls at or below the level being sought
- %ile = (N scores < or equal x) / N * 100
Simple Frequency Distributions
- Using a cumulative f column
Measures of Central Tendency
- Values that describe the middle, or central, characteristics if a set of data
Mode
- The score that occurs most frequently in a distribution
- Pros: Ease of calculations and use with nominal data
- Cons: disregards extreme scores and is a terminal statistic
- USE IF:
-- Rough estimates of C.T. is needed
-- only measure of CT for nominal data
-- Not much info
Terminal Statistic
- Statistic that can not be used for further calculations
Median
- Score at the 50th percentile, or the middle score; divides the data set in half
- Based on the # of scores, not their value, therefore more appropriate for ordinal or highly skewed data
- An outlier does not skew the median
- Does not take the "value" of scores into account (therefore some information is lost)
- USE IF
-- Ordinal data
-- The curve is badly skewed
Calculating the Median
- In a set of rank ordered data where N is odd the median is the middle score
- When N is even the median falls between two scores and is either
-- The higher of the 2 scores (ordinal data)
-- The average of the 2 scores (interval or ratio data)
Mean
- Arthmetic average
- Most common measure of central tendency
- CONSIDERS BOTH THE # OF SCORES AND THEIR VALUES
- Gives weight to each score according to its value or distance from the other scores
- Most sensitive measure of central tendency
- Appropriate for ratio and interval data however not for ordinal data.
- Most sensitive measure of C.T., because of outliers
- USE IF
-- Interval or ratio data
-- Normal distribution
-- Both order and relitive value are important
--Further calculations are to be made (really this one)
Outliers
- Scores that are considerably higher or lower than the other scores
-- These can pull the mean toward the extremes
-- Remember the positively skewed curve
Calculating the Mean
- The sum of the variable scores (X), divided by the total number of scores (N)
Positively Skewed Curve
- Curve towards the left side
- Order: Mode, Median, Mean
Negatively Skewed Curve
- Curve towards the right side
- Order: Mean, Median, Mean
Variability
- A measure of the spread, or dispersion of a set of data
- Two or more, sets of data can be compared by means, but it may be more important to compare variability of the data.
Types of Variability Measures
1. Range
2. Interquartile Range
3. Variance
4. Standard Deviation
Range
- (R)
- High to low score
- a quick, rough estimate of variability
- Unstable when outliers are in the data set
Interquartile Range
- (IQR)
- Difference between the raw scores at the 75th and 25th percentiles
- IQR = Q3 - Q1
-- Used with ordinal data, or highly skewed ratio or interval data
-- Only considers 50% of the data, or the middle half
-- Like the Median, it is not affected by outliers.
Variance
- (V)
- The average of the squared deviations from the mean; considers each score and its distance from the mean
- It is not right score so not used in a lot of research
Deviation
- (d)
- distance of each raw score from the mean
- Considers the value of each data point
- The sum of all deviations must be zero
Standard Deviation
- SD
- The square root of the Variance, or square root of the avg of the squared deviations from the mean
-- The variance is out of line with the original raw data, so the square root of the variance is computed
--- The resulting value is now in alignment with the original raw data.
- Many advanced statistics use X for central tendency and SD for dispersion
Standard Deviation of a Population
1. Definition Method
2. The Raw score method
Standard Deviation of a Sample
1. Samples rarely contain the extreme values of a population, thus, the variance of a sample is almost never as large as the population variance.
2. We use a correction factor so that the population estimate is not biased by a small sample
3. Degrees of Freedom (Df) : the # of values in a data set that are free to vary. When we estimate a value of a population from a sample (Parameter) we are putting limits on the data
- Normal: the N is large, and the distribution of the data is "normal," there are 6 standard deviations within the range of data points; 3 below and tree above x.
Standardized Scores
- Used to compare similar or dissimilar performances.
- Percentiles are an example of such standardized scales
Normal Curve and Standardized Scores
- Many variables will assume a normal distribution if enough cases are observed, also nearly all physiological phenomenon
- Data must be from an interval or ratio scle
Z-scores
- A raw score expressed in standard deviation units
- The percentage of the area under the normal curve between the x and any z score is constant and can be calculated.
- 34.14% of the scores in a normal distribution lie between the mean and one z score +/-, therefore 68.28% of a popualtion lies between +/- 1 z score
Confidence Interval
- 1.96 +/- is 95% exactly around the mean.
- 2.5 +/- 98%
- 49.87% if the scores lie between the x and 3 SD or 99.74% lie +/- 3 SD from the x
Predicting Population Parameters
- population is too big, therefore researcher assumes the sample represents the pop.
-- must be random and as large as possible (keep in mind resources)
- Sample size limits: time, willingness of subjects, funds, equipment, facilities, etc.
-- skinfolds are fast, inexpensive, field technique and valid
-- UWW is $$, and time consuming.
- Snowball samplling
Estimating Sample Error
- Amount of error in the estimate of a parameter that is based on a sample statistic - pop estimate
Standard Error of the Mean
- A numeric value that indicates the amount of error that may occur when a random ample mean is used as a predictor of the the population mean
- The larger the sample, the smaller the error in predicting the pop mean.
- If all the sample means (x) were then arranged in a frequency curve, the curve would approx a normal curve
The Central Limit Theory
- The means from a series of samples randomly selected from a pop. will be normally distributed even if the pop. They were chosen from is not normal.
The Grand Mean
The mean of the means
Pearsons product moment correlation coefficient (r)
- Always between -1 and +1
- if r =0 no relationship
- Sign of the coefficient does NOT indicate the strength or usefulness of the relationship coefficient
- Sometimes the true line of best fit is curvilinear, not straight, and not correctly measured by Pearson's r!
- Correlations have nothing to do with cause and effect.
Correlation
- it is a inferential statistic, it simply states a relationship between variables
R values
- 7< r < .85
- The significance of r, the overall usefullness of the r value is based on size and significance is based on the size and significance
Common Variance
(r squared)
- The amount of the variance in Y that can be explained by the variance in x
- AKA the coefficient of determination