What percentage of clinicians say they understand the research?
What percent of articles are considered accessible by those with basic stats training?
What were the main points about the ELITE1 and ELITE 2 trial?
ELITE1 showed losartan had 50% less risk of mortality, but mortality was a secondary endpoint.
In ELITE2, the primary endpoint was mortality, and it showed that mortality was unchanged with losartan.
Thus don't use secondary endpoints to make judgements.
What are some stats on highly positive results that end up being contraindicated?
- analyzed 115 articles published from 1990-2003 in major journals and speciality journals that had over 1000 citations.
- -49 reported evaluations of health care interventions
- 45 claimed the interventions were effective
- by 2004, 5/6 non-randomized studies, and 9/39 randomized trials were contraindicated or exaggerated
How much of research may not be randomized?
What was a problem with the RALES study?
- spironolactone vs placebo
- very stringent inclusion/exclusion criteria
- Spironolactone was protective in the study
In real life, the patients had a higher risk of hyperkalemia
- Pts not monitored ascarefully, lower compliance, interaction with other drugs (Ace-inhibitors)
- 24% got hyperkalemia, compared to 1.7%
What is the main recommendation of the RALES study?
spironolactone is protective in heart failure, but only in a select few patients who meet inclusion criteria.
What were the main limitations of the study that looked at the risks for new users of NSAIDs?
- Confounding [non randomized, lack of detailed clinical info]
- Availability and use of drugs [ cox-2 have a restricted status, non-selective are available OTC]
- Duration of use was short-term
- Generalizability of patient population
What is a time-series? What is the major goal?
A set of observations on the values that a variable takes at diff. times
Goal: to develop a mathematical model that describes pattern to allow future prediction.
How can you figure out random fluctuations in a time series?
Take the trend, and subtract the cyclical components, the remaining portion are the random fluctuations.
What is autocorrelation? What is an example?
error term associated with any observation isrelated to error term of other observations.
- ex. Durbin-Watson statistic
- Ljung-box statistic
What is stationarity?
mean and variance of stochastic processes are consistent overtime, and the covariance depends only on distance or lag between the 2 time periods.
ex. Augmented Dick-Fuller test
What is an inception cohort?
Excludes patients with history of disease, just recent diagnosis/onset
Strengths of a cohort study
- Time sequence; exposure precedes outcome
- collect info on all relevant predictions/confounders
- minimizes recall bias
- estimates incidence
- can assess multiple outcomes
Weaknesses of cohort studies
- Large # of people needed
- loss to follow up
- expensive, and resourceintensive
- status of subject may change (switch between control and exposure)
- may miss subclinical forms of outcome
- inefficient for studying rare outcomes
- Prospective: time consuming
- Retrospective: limited control over sampling and choice/quality of predictor variables
What are four aspects in which cohorts and RCTs differ? and how so?
- Population: diverse vs. highly selected
- Intervention: patient/provider vs randomized
- Follow-up: longer followup vs short follow-up
- Analysis: sophisticated multivariate techniques for confounding adjustments vs. simple-to-sophisticated
How can you tell if an exposure is a cause of an outcome?
Exposure is a cause of an outcome if exposure at a given level results in a different out come than would have occurred without that level of exposure
Why establish causality?
- Guide to predict, prevent, diagnose , treat
- Treatable or reversible cause
- Clinical and research
What is the bradford-Hill criteria for establishing causality?
- strength of association
- biological plausability
- specficity: Causation is likely if a very specific population at a specific site and disease with no other likely explanation.
How might you alter your cohort study research design to help establish causality?
- Ensure entire group not experienced outcome
- observed time period needs to be meaningful in disease context
- complete follow up
What are some ways you can distinguish causality from association?
- chance (random error)
- bias (in selection or measurement)-> systematic error
- confound (another type of bias)
- effect modifier (interactions)
Selection bias (according to cohort study lecture)
- systematic error in creating intervention groups, causing them to differ with respect to prognosis
- groups differ in baseline characteristics due to ways in which participants were selected for study or assigned to study groups
- -occurs at design phase, may impact external or internal validity
How can you control selection bias?
- sensitivity analysis
What is sampling bias?
Study sample taken from population of interest should be representative of the population
- -inclusion/exclusion criteria
- -volunteer bias
- -referral centre bias
What is assembly bias?
- one group is more susceptible to outcome than another (aka susceptibility bias)
- due to differences in extent of disease, presence of co-morbidity, prior treatments.
What is migration bias?
- patients drop out or move from one group to another
- loss-to-follow up
When can measurement bias occur?
What are some types of measurement bias?
Can occur at design, conduct, or analysis phase
- ascertainment : differential way data is collected
- screening bias (aka lead-time bias)
How can you control measurement bias?
- exposure and outcome data should be collected similarly
- standard outcome definition
- choose an objective/hard outcome
What is confounding?
Estimated intervention effect is biased because of some difference between comparison groups apart from the planned intervention, such as baseline characteristics, prognostic factors, or concomitant interventions
How to control for confounding?
- Actively exclude or control for confounding variables
- cohort studies: matching
- RCT: stratification
- controlling for confounding by measuring known confounders and including them as covariates in multivariate analysis.
How can you determine the importance of a bias?
Strength and direction
- strength: robustness of findings when bias is controlled
- direction: bias less important if association is significant despite bias in opposite direction
What are the implications of bias?
- real, spurious, indirect
- spurious: due to selection bias, measurement bias or chance
- indirect: due to confounding
How can you improve follow-up for cohort studies
- exclude those likely to be lost (moving, unwilling)
- obtain info to allow future tracking
- during follow-up: periodic contact with subjects
- checking vital stats from OHIP/registries
What is a hazard ratio?
- Similar to RR but time-to-event analysis, has an associated time value.
- Usually shown as a survival rate
"Twice as many people developed HT in 6 months"
What are some critical appraisal tools for cohort studies?
Random info about STROBE statements
- Strengthening the Reporting of Observational Studies in Epid
- -international collaborative initiatives of epid, methodologists, statisticians involved in conduct and disseminiation of observational studies
Random info about NOS for cohort studies
aim is to assess the quality of non-randomize studies with its design, content, and ease-of-use
- uses a star system which judges a study in 3 perspectives:
- selection of study groups
- comparability of groups
- ascertainment of either exposure or outcome of interest, for case-control or cohort studies, respectively.
What is the goal of the NOS?
to developan instrument providing an easy and convenient tool for quality assessment of non-RCT to be used in systematic review
Who is richard doll?
linked smoking to health problems using case-control study
If you get similar results/estimates for OR when you compare your cases to different controls, then this is evidence _____ _____
What is a nested-case control study? What are its advantages? What is an example?
- Case-control nested within a cohort study
- most data are collected before outcome occurrence, so less likely to be impacted by recall bias
- ex. peanut allergy study by AVON longitudinal study group
What are some advantages of case-control studies?
- study a risk factor/multiple risk factors
- disease of interest is rare (impractical for cohort)
- long latency period between exposure and disease
- random assignment is unethical or impossible
- need an answer quickly (bypassses need to collect info on large # of people who wont get the outcome)
Case-control lets you study ___ while cohort lets you study ____
Which type of study is more susceptible to recall bias? Case-control or cohort?
When does OR approximate the RR?
When the baseline probabilities of the outcome are low (<0.1-0.2)
What is a primary base?
- Thepopulation that the investigator wishes to target, with the cases being the subjects who develop the disease within the base. The population is defined geographically and temporally
- adv: easier to sample for controls from a primary base
- disadv: it is challenging/impractical to ascertain all cases in a primary base
What is a challenge in using secondary bases?
determination of the study base
What is the ideal way of selecting cases?
- choose all incident cases in the source population
- -recall of past exposures may be more accurate
- -temporal sequence is easier to assess.
What is a problem with using prevalent cases?
- biases towards longer survival among participants
- differential recall of risks
- exposures status may change with onset of each disease
What is the ideal way to select controls?
Direct random sampling from source population
What are some different types of controls?
- hospital/medical practice controls
- neighbourhood/friends controls
- proxy respondents
- deceased controls
When might it be important to include hospital/medical practice controls?
in cases where the probability of disease diagnosis depends on access to medical care, then it would be appropriate to include hospital controls
Population based controls
- Random sample of population from which cases came
- usually defined by geographic borders
- valid if cases consist of all individuals who developed disease of interest in a defined population
- usually patients seeking care for other conditions (controls should have a variety of conditions)
- disadv: not always possible to identify source population of the cases
- -additional biases of specialized medical facilities
- source population of cases poorly defined and presumably healthy controls are desirable
- neighbours tend to seek care at similar places
- neighbours are similar with respect to socioeconomic status and determinants of health
What is the comparable accuracy principle?
measurement of exposure should be comparable in cases and controls
What is the deconfounding principle?
How can you deconfound?
- confounding should not be allowed to distort estimation of effect
- restriction, matching, adjustment
What is restriction?
what does it help control?
what is it susceptible to?
- Limits #of eligible subjects (ex.only males aged 40-50)
- helps control selection bias
- susceptible to residual confounding effect (risk still varies within age 40-50)
- limits generalizability and does not allow evaluation of the restricted factors
What is matching
what does it help control
- Ensure that grous do not differ for confounders (ex. for every active male between 40-50, there will be an inactive male aged 40-50)
- helps control selection bias
- requires control of confounding at both design and analysis of study.
- advantages: allows matching geographically to control socioeconomic/ethnic factors
- diadvantages: overmatching: match on factors that may themselves be related to exposures.
What is recall bias?
- Occurs when a survey respondent answers a question
- influenced by correct answer, but also respondent's memory
- ex. response bias: socially acceptable response
What is misclassification bias?
- Type of information bias orginated when sensitivity and specificity of procedure to detect exposure and/or effect is not perfect.
- i.e. exposed/diseased classified as non-exposed/non-diseased or vice-versa
What are some sample size considerations in case-control studies?
very little improvement past case:control of 4:1
remember that increasing sample size only improves precision, not validity.
What are some advantages of case-control studies?
- Good for studying rare condition/disease
- Good for long latency between predictor and outcome
- Useful for hypothesis generation since relatively inexpensive and fast as disease already developed
- can evaluate multiple etiological factors/exposures at once
Weakness of case-control studies?
- cannot estimate incidence/prevalence
- can only study one outcome, as that is how popn is sampled
- prone to various errors and biases
Define survey? What determines whether it's analytic or comparative?
- detailed and quantified description of a population
- systematic collection of data
- The research question does
Measure characteristics of a particular population, either at a fixed point in time, or comparatively over time
attemptto test a theory; i.e. explore and test associations between variables.
for surveys: define independent, dependent, and uncontrolled variables:
- subject of study, gains or losses produced by impact of research study
- 'cause' of changes in the dependent variable that will be manipulated or observed, then measured in dependent variable
- includes error variables that may confound results of the study (ideally you want these variables to be randomly distributed)
how can you control extraneous variables in surveys?
- hold them constant
- exclusion: only use females to diminish effects of gender
Stages in survey process
Research questions-> decide on information needed -> decide on prelim. analysis [examine resources, review existing lit. -> decide on sample+ chose survey method -> Design questionnaire -> Pilot survey -> amend questionnaire and sample -> main survey -> edit code and tabulate -> analyze -> final report
What are self-administered surveys? What are interviewer administered?
SA: postal, online, delivery and collection
I: structured interview, telephone questionnaire, focus group
When are postal surveys good?
- sample widely distributed geographically
- subjects need to be given time to think
- subjects have moderate/high interest in subject
- questions are written close-ended
When are Delivery and collection surveys good?
- Delivered by hand to each respondent and collected later
- -allows direct contact with potential respondents, which may lead to higher response
Why are online suveys good? Bad?
Adv: low cost, easy to design and administer, anonymous-> honesty
- Disadv: volunteer sample (little control over who responds)
- sampling error (non-internet users)
Structured surveys.,advantages and disadvantages
adv: higher response rates, can ask open ended questions for detailed responses, additional probes can be asked
disadv: time consuming, expensive, sensitive topics-> less accurate
Focus groups: advantages and disadvantages
- allow for a variety of views to emerge
- group dynamics can often allow for stimulation of new perspective (allows for the basis of a survey)
telephone surveys, advantages and disadvantages
- most people have a phone
- higher response rates
questions need to be short
what is a probability sample, and examples of probability samples:
- each member of population has known non-zero probability of being selected
- allows calculation of sampling error
- ex. random, systematic stratified
what is a non-probability sample, and examples of non-probability samples:
- members selected in non-random way
- sampling error unknown
- ex. convenience, judgement, quota, snowball
Describe each of the probability samples
- Random: each member has equal and known chance of selection
- Systematic: also called 9th name selection [as long as no hidden order in list of eligibile participants, just as good as random]
Stratified sampling in surveys:
- reduces sampling error even compared to random sampling
- identify relevant strata, then random sampling to select subjects from each stratum until # of participants is proportional to its frequency in the population
What is a stratum
subset of population that shares atleast one common characteristic
Explain the non-probability sampling techniques
- convenience: mostly during preliminary phases (ex. friends and family)
- judgement: an extension of convenience; researcher makes a judgement that this one sample is representative of the entire population (ex. use one city to sample the country)
- quota: non-probability equivalent of stratified sampling: identify your strata and proportions as represented in population, then use judgement sampling to select required number from each stratum
- snowball: used when desired sample characteristics are rare; relies on referrals from initial participiants to find new participants
Reducing sampling error in surveys
- contact members of sampling frame and ascertain whether they belong to the required sample
- design questionnaire/interview in such a way that ineligible respondents are identified early and screened out
When do you need a scale?
- If there is no easily applied "objective" measurement tool
- ex. determining a patient's status, assessing quality of life, enhancing educational programs
What are 4 question types in surveys?
- Objective: can be verified, real answer exists
- Subjective: ask about personal perception, no factual answer.
- Open-ended: allows information gathering, tough to analyze and time consuming
- Close-ended: useful for hypothesis testing, but may lose important info. time effecient, lots of response formats
Response formats for close-ended questions
- verbal frequency scale (all the time, fairly often, never etc. )
- likert scale
- graphical rating (continuous scale on a line)
What are some random facts about Likert scales?
- Measures strength of feeling or perception
- optimal number is 7+/- 2
- Often analyzed as interval data, but is actually nominal
What type of wording should you avoid in questions?
- Double negatives
- Double Barelled: ask for thoughts about 2 things at once
- Leading: suggests the answer the research is looking for (may end in weren't you, don't you, etc. )
- Loaded: contains a controversial assumption
Who do you pre-test on? Who do you pilot test on?
- Pre-test on convenient other
- Pilot on small sample from target popn.
What are the four dimensions of a good scale?
Lack of reliability in a tool does what?
- produces error variance
- places upper limit on validity (requires reliability for validity,but not the opposite)
What is classic reliability theory?
- raw score= true score + error
- sources of error: misinterpretation, biases, inexperience, inter-rater differences
Why are reliability tests constructed? What is the prof's watered down equation for a reliability test?
Tests constructed to differentiate between objects
Implications of reliability based on her equation
- reliability is not a fixed property of the scale
- change popn-> change reliability
- more items/raters increase reliability
- need trepeat measures to estimate error
- test that doesn't discriminate is useless!
What is the goal for internal consistency in reliability? For stability?
>0.8: based on one time administration of survey, is there good correlation between related domains? - does not account for day-to-day or observer variation into account so usually an optimistic reliability
- Stability >0.5:
- reproducibility when administered on different occasions:
- ex. test-retest: same result on 2 occasions
- intra-rater: agreement between ratings made by same rater on diff. occasions
- inter-rater: agreement between 2 different raters
Whats problematic about invalid tools? List some types of validity
- Wrong conclusions, unethical
- Content: does it tap relevant + disregard irrelevant ideas (based on expert opnion, existing tools etc. )
- Criterion: are results consistent with other measures [divided into concurrent and predictive]
- Construct: can we predict differences based on constructs? requires first 3 types of validity
Criterion validity types
- Concurrent: examine relationship between criterion measure and scale at time of administration
- predictive: examines relationship between scale and future outcomes (ex. compared to a gold standard)
What are some ways to test validity?
- extreme groups: t-tests
- change: 2 way anova
- criterion: pearson's correlation
What are some concerns with feasibility?
- Avoid ambiguity
- Consider reading level
- How much training required?
- How easy to score (avoid weighing responses, be aware of unintentional weighting)
What are some concerns with acceptability?
- be brief (but remember reliability proportional to # of items)
- use only items for which there is a variety of responses
- be aware of social desirability bias
- those who administer the scale willing to do so?
How to reduce item non-response
- Avoid intrusive questions
- emphasize confidentiality
How to reduce error associated with poor response rate
- identify most appropriate respondents
- use multiple forms of contact
- develop easy to complete questionnaire with instructions
- conduct on-site interiews to taylor question to participant cognitive ability
- be cautious about financial incentives
What are the cognitive requirements of responding:
- understand question
- recall relevant attitude, belief, behaviour
- inference and estimation (decomposition and extrapolation; end-digit bias)
- Map answer on response alternatives
- edit the answer (what people think and what they tell you)
Types of response bias:
- Social desirability
- End aversion: no extreme responses
- Positive skew: favourable response
- Halo effect
- framing: choice between 2 alternatives depends on how they're framed
Minimizing response bias:
- keep task simple
- maintain motivation of respondents (choose only those interested, motivation higher earlier so keep it short, ask people to explain their answers)