Use LEFT and RIGHT arrow keys to navigate between flashcards;
Use UP and DOWN arrow keys to flip the card;
H to show hint;
A reads text to speech;
37 Cards in this Set
- Front
- Back
Name of Weber 2003 paper
|
Theoretically speaking. MISQ
|
|
Name of Whetten 1989 paper
|
What constitutes a theoretical contribution? AMR
|
|
Name of Weick 1995 paper
|
What theory is not, theorizing is. ASQ
|
|
Name of Sutton & Straw paper
|
What theory is not. ASQ
|
|
Name of Orlikowski & Baroudi 1991 paper
|
Studying information technology in organizations: Research approaches and assumptions. ISR,
|
|
Name of Sutton & Rafaeli papers 1988, 1992
|
(1988). Untangling the relationship between displayed emotions and organizational sales: The case of convenience stores. AMJ
(1992). Journey 4: How we untangled the relationship between displayed emotion and organizational sales: |
|
Name of Eisenhardt 1989 papers
|
Building theories from case study research. AMR
Making fast strategic decisions in high-velocity environments. AMJ |
|
Name of Landauer & Prabhu book
|
Handbook of Human-Computer Interaction
|
|
Name of Gould, Boies, & Lewis 1991 paper
|
Making usable, useful, productivity-enhancing computer applications. CACM
|
|
Name of Siegel & Dray 2005 paper
|
Avoiding the next schism: Ethnography and usability. Interactions
|
|
Name of Gopal & Prasad 2000 paper
|
Understanding GDSS in symbolic context: Shifting the focus from technology to interaction. MISQ
|
|
4 lessons from Gopal & Prasad 2000 paper
|
1. uses interpretive methods to investigate contradictions in traditional GDSS
2. Three interviewers followed teachers in a GDSS session to collect data 3. authors report that emotional plesantness varies among GDSS users, may depend on social contecxt 4. authors conclude that interpretive/symbolic method is useful to future GDSS study |
|
Method used in Hawthorne study
|
started as empirical and switched to ethnography
|
|
Name two examples of an ethnography study
|
Gopal & Prasad, GDSS, 2000 paper
Hawthorne study |
|
3 lessons learned from Siegle & Dray 2005
|
1. usability is traditionally studied using lab experiments
2. ethnography recently attempts to study utility in field studies 3. combining both methods is better than either separate |
|
3 lessons learned from Gould, Boies & Lewis 1991
|
1. traditional usability design includes the following:
+ early focus on users + integrated design with UI, help, training, docs + early and continued user testing 2. traditional design not used as often in practice as should be 3. calls for new technique, better design software, benchmarks |
|
Lessons learned from LAndauer 1997 book on HCI
|
1. behavioral research is important in evaluating new software research
2. teams of behavioral researchers should conduct research, including marketing researchers 3. statistically significance is not important in evaluating new software, suggest odds-ratio instead. |
|
Lessons learned from Mann Gulch Disaster study
|
1. example of case study that is unable to be replicated in the lab
2. firefighter managers said the findings were brilliant |
|
Lessons learned from Eisenhardt's building theory from case study paper
|
1. Select research question
2. select theoretically useful cases 3. Craft instruments and protocols 4. collect data in field 5. analyze data 6. shape hypotheses 7. enfold literature, both supporting and nonsupporting 8. Reaching closure |
|
Lessons learned from Eisenhardt's making fast decisions paper
|
1. example of case study
2. findings include: + fast decision makers use more information than do slow decision makers + fast decision makers generate more alternatives than slow decision makers + using knowledgeable consultants made decision faster + the greater use of active conflict resolution, the faster the strategic decision process 3. generally, fast decision making following these guidelines leads to superior performance |
|
6 lessons from Alex Durcikova discussion
|
1. Be sensitive to subjects needs & management's shouldn't look bad.
2. Use the terms that th subjects company uses. 3. Administer survey in a consistent way 4. specify the domain of the construct 5. create 5-10 survey questions to investigate the construct 6. put similiar survey questions together with a simple paragraph introduction |
|
lessons from Sutton & Ralaeli 1988, 1992 papers
|
1. data may not support your theory
2. qualitative research can be useful 3. demonstrates difficulty in getting published |
|
lessons from Weber 2003 paper
|
1. Choose a phenomena
2. Articulate the phenomena 3. Articulate the constructs of a theory 4. Articulate the relationships between constructs in your theory 5. Articulate the boundaries or limits to where and under what conditions your theory applies. |
|
lessons from Whetten 1989 paper
|
Contribution is demonstrated by adding value:
1. Adding or deleting an existing construct to an existing theory is not enough contribution 2. Adding or deleting constructs with an explanation that changes previous relationships or understanding of relationships is valuable contribution 3. How does modifying the constructs significantly alter our understanding of the phenomena 4. surprising results, with explanation that increases our understanding, is valuable contribution. 5. "New applications should improve the tool, not merely reaffirm its utility." |
|
lessons from Sutton & Straw 1995 paper
|
Theory is not just the following:
1. Referencing previous theory to explain only what was done previously is not theory 2. Data is not theory 3. List of variables and constructs are not theory 4. Diagrams are not theory 5. Hypotheses are not theory Good theory includes all the above, but answer the Why questions and establishes an explaination for cause and effect. |
|
lesson from the Weick 1995 paper
|
"Most products that are labeled theories actually approximate theory."
The process of theorizing is valuable and should not be discounted. |
|
List Creswell's four research paradigms
|
Pragmatism
Advocacy Constructivism(Interpretivism) Positivism |
|
What is pragmatism
|
research method concerned with applications and solutions
|
|
what is advocacy
|
research method concerned with change and political agenda
|
|
what is constructivism
|
research where knowledge depends on one's interpretation of the world
|
|
what is positivism
|
research using traditional scientific method
|
|
define construct validity
|
answers if you are measuring what you claim to be measuring. e.g. you claim to measure anger but are really measuring aggression. A construct should be composed of multiple variable rather than just one variable.
|
|
define External validity
|
answers how generalizable the results are to people, places, times, or things outside the subject samples used in the study.
|
|
define Internal validity
|
the extent to which the treatment is the cause of the observed changes and not attributable to other possible causes
|
|
define Content validity
|
the degree to which items in an instrument reflect the content universe to which the instrument will be generalized.
|
|
define reliability
|
an indication of measurement accuracy, that is, the extent to which instrumentation produces consistent or error-free results.
|
|
3 lessons learned from Orlikowski & Baroudi 1991 paper
|
1. positivist approach is the predominant method in MIS
2. explored epistimology (how we come to know things) 3. Much can be gained if a plurality of research perspectives is effectively employed to investigate information systems phenomena |