Course Content
Epidemiological Data
0/1
Emergency Psychiatry
0/2
Psychiatric Services
0/1
Critical Review
0/54
MRCPsych Paper B Made Easy
    Bias in Research Studies - Study Notes

    BIAS

    Validity in Research Studies

    When reviewing a research paper, a primary concern is whether the study's findings are trustworthy and meaningful. This relates to the concept of validity.

    Internal Validity

    Internal validity refers to how accurately the results of a study reflect the true situation within the specific group of participants studied. If a study has high internal validity, it means the observed effects (e.g., the difference between a treatment group and a placebo group) are likely due to the intervention being tested and not because of flaws in how the study was designed or conducted. It addresses the question: "Are the conclusions drawn about the study participants themselves correct?"

    • Example: Imagine a study finds that a new medication significantly reduces anxiety symptoms compared to a placebo within the 100 participants enrolled. If the study was well-designed, carefully executed, and free from major errors, we would say it has high internal validity. We can be reasonably confident that the medication did cause the reduction in anxiety within that specific group.

    External Validity (Generalisability)

    External validity, also known as generalisability, refers to the extent to which the results of a study can be applied to other populations, settings, or situations beyond the specific context of the study. It addresses the question: "Can these findings be applied to my patients or to the broader population?"

    • Example: Following the anxiety medication study above, external validity asks whether the positive results seen in the 100 participants would also apply to other people with anxiety who weren't in the study – perhaps people in different countries, of different age groups, or with co-existing medical conditions. If the study participants were very narrowly selected (e.g., only males aged 20-30 with no other health problems), the results might not be generalisable to a wider range of patients, indicating lower external validity.

    Threats to Study Validity

    The validity of a study's findings can be compromised, meaning the results might not be accurate or applicable. There are three main reasons why this can happen:

    1. Random Error (Chance or Imprecision): This occurs due to unpredictable variations.
    2. Systematic Error (Bias): This involves errors that occur in a consistent direction.
    3. Confounding: (While mentioned as a threat, confounding involves the influence of an external variable distorting the relationship between the exposure and outcome; the provided text focuses on error).

    Random Error

    Random error arises purely from the play of chance. Most research studies involve examining a sample (a smaller group) to make inferences about a larger population. Because a sample never perfectly represents the entire population, there's always a chance that the results observed in the sample differ from the true state in the population simply due to random variation – like drawing an unusual number of red marbles just by chance when sampling from a bag containing red and blue marbles.

    • Characteristics:
      • Unpredictable: It doesn't consistently push results in one specific direction.
      • Due to Sampling Variability: Arises because we study a sample, not the whole population.
      • Reduces Precision: Makes the study findings less precise or certain.
    • Assessment: Statistical methods like p-values (which estimate the probability that an observed effect occurred by chance) and confidence intervals (which provide a range of values likely to contain the true population value) are used to quantify the potential impact of random error.
    • Reduction: Random error can be reduced by increasing the sample size (larger samples tend to be more representative) or by repeating the study (replication). If similar results are found across multiple studies, it becomes less likely that the findings are due to chance alone.
    • Example: A study comparing two treatments in small groups of 10 patients each might find Treatment A looks slightly better than Treatment B just by chance, perhaps because a few patients in the Treatment A group happened to be slightly healthier at the start. If the study were repeated with 500 patients per group, such chance differences would be less likely to influence the overall result significantly.

    Systematic Error (Bias)

    Systematic error, or bias, is different from random error. It refers to flaws in the study's design, conduct, or analysis that cause the results to deviate from the truth in a predictable direction. Bias introduces a systematic distortion, leading to findings that are consistently higher or lower than the true value.

    Random Error vs. Systematic Error (Bias)

    • Random Error: Due to chance, affects precision (scatter), unpredictable direction, reduced by larger sample size.
    • Systematic Error (Bias): Due to flaws, affects accuracy (off-target), predictable direction, not reduced by larger sample size.
    • Characteristics:
      • Systematic: Pushes results consistently in a particular direction (e.g., always making the treatment look better than it is).
      • Due to Flaws: Arises from errors in how participants are selected, how information is collected, or how data is analysed.
      • Reduces Accuracy: Makes the study findings inaccurate.
    • Contrast with Random Error: Unlike random error, bias is not reduced simply by increasing the sample size or repeating the study using the same flawed methods. The same systematic error will likely be repeated, leading to consistently inaccurate results.

    Types of Bias (based on origin):

    1. Selection Bias: Errors in the process of selecting participants for the study or in factors influencing their retention, leading to groups that are not truly comparable.
      • Example: In a study comparing smokers and non-smokers for a certain disease, if researchers recruit smokers from a hospital clinic and non-smokers from a local sports club, the groups might differ in many ways other than smoking (e.g., general health, lifestyle), potentially biasing the results regarding the effect of smoking on the disease.
    2. Measurement Bias (Information Bias): Errors in how exposure, outcome, or other data are measured or recorded, leading to inaccurate information.
      • Example: In a study asking participants about past dietary habits, people who have developed a disease might recall their past unhealthy habits more vividly (or feel more inclined to report them) than healthy controls, leading to a biased estimate of the link between diet and disease (this specific type is called recall bias). Or, if an interviewer knows which participants received the active drug, they might probe more thoroughly for positive effects in that group (observer bias).
    3. Analysis Bias: Errors occurring during the statistical analysis of the data.
      • Example: If researchers decide to exclude participants with unfavourable outcomes from the analysis in the treatment group but not the control group, this would bias the results in favour of the treatment. Another example is choosing a specific statistical test because it yields a desired significant result, rather than using the most appropriate test determined beforehand.

    Consequences of Bias

    Bias is a serious threat to the validity of research findings.

    • False Conclusions: It can lead researchers to draw conclusions that are incorrect - for instance, concluding a treatment is effective when it is not, or vice versa.
    • Inaccurate Effect Estimation: Bias can cause an overestimation (making an effect look larger than it is) or an underestimation (making an effect look smaller than it is) of the true relationship or treatment effect.
    • Origin: Bias often stems from weaknesses in the study design (e.g., poor randomization procedures, lack of blinding) or during data collection (e.g., using uncalibrated instruments, leading questions in interviews).
    • Irreversibility: Importantly, once bias has been introduced into a study through its design or conduct, it generally cannot be corrected or "controlled for" during the statistical analysis stage. The focus must be on preventing bias in the first place through careful study planning and execution.

    Direction of Bias

    Bias can alter the study outcome in specific directions:

    1. Toward the Null Bias (Negative Bias): This type of bias makes the observed effect appear smaller than it truly is. It pushes the results closer to the "null value," which represents a state of no difference or no association between the groups being compared.
      • Example: A study investigating a new antidepressant genuinely improves mood scores by 10 points compared to placebo. However, due to bias (perhaps participants on the active drug downplay their improvement), the study only shows an average improvement of 5 points. The observed effect (5 points) is closer to the null value (0 points difference) than the true effect (10 points).
    2. Away from the Null Bias (Positive Bias): This bias makes the observed effect appear larger than it truly is. It exaggerates the difference or association found in the study.
      • Example: In the same antidepressant study, imagine the assessors are aware of who received the active drug and subconsciously rate their mood improvement more favourably. If the true improvement is 10 points, the biased assessment might lead to a reported average improvement of 15 points. This observed effect (15 points) is further away from the null value (0 points) than the true effect (10 points).
    3. Switch-Over Bias: This is an extreme form of bias that not only changes the magnitude of the effect but also reverses its direction. What appears to be a beneficial effect might actually be harmful, or vice versa, due to the bias.
      • Example: A study aims to see if a certain lifestyle factor increases the risk of anxiety (True Odds Ratio > 1). Due to significant bias (e.g., severely flawed selection of participants), the study concludes that the factor actually decreases the risk of anxiety (Observed Odds Ratio < 1). The direction of the association has been switched.

    When Bias Can Occur

    Bias can be introduced at various stages of the research process:

    • During the selection of participants (Selection Bias)
    • During the measurement of exposures, outcomes, or other variables (Measurement Bias)
    • During the statistical analysis of the data (Analysis Bias)

    Selection Bias

    Selection bias occurs when the groups selected for comparison within a study differ systematically in ways that are related to the exposure or outcome being investigated, beyond the factors under study. This means the study participants are not representative of the broader population they are meant to represent (the target population), or the comparison groups are fundamentally different from each other at the start.

    • General Example: In an RCT comparing a new antipsychotic drug against a standard treatment, if the group receiving the new drug predominantly includes patients with less severe symptoms or shorter illness duration compared to the standard treatment group, any observed difference in outcome might be due to these initial differences rather than the drug itself. This initial imbalance represents selection bias. Adequate randomization is the primary method to minimize this type of bias in RCTs by creating comparable groups.
    • Example (Diagnostic Procedures): If in a population, individuals with a specific risk factor (e.g., family history of depression) are more likely to undergo thorough diagnostic assessments for depression than those without the risk factor, a study might find a spurious association between the risk factor and depression simply because cases are more likely to be detected in the group with the risk factor.

    Specific types of selection bias include:

    • Berkson Bias (Admission Rate Bias): This bias occurs primarily in hospital-based case-control studies. It arises when the combination of the exposure and the disease influences the likelihood of being admitted to the hospital, leading to systematically different characteristics between cases and controls drawn from the hospital population compared to the general population.
      • Example: A hospital-based case-control study investigates the link between smoking (exposure) and dementia (disease). Cases are patients with dementia admitted to the hospital. Controls are patients admitted for other reasons. Since smoking increases the risk of many diseases requiring hospitalization (e.g., heart disease, lung cancer), smokers are generally overrepresented in the hospital population compared to the community. If controls are selected from this hospital population, they will have a higher rate of smoking than the general population from which the dementia cases arose. This can make the association between smoking and dementia appear weaker, or even absent, because the control group has an artificially high prevalence of the exposure (smoking).
    • Neyman Bias (Incidence-Prevalence Bias): This bias occurs when there is a gap in time between exposure and disease assessment, particularly affecting studies that use prevalent (existing) cases rather than incident (newly diagnosed) cases. If the exposure is related to the duration of the disease (e.g., causing rapid fatality or quick recovery), using prevalent cases can lead to a distorted measure of association.
      • Example: A case-control study examines the association between a specific genetic marker and rapidly fatal schizophrenia. The study recruits prevalent cases (people currently living with schizophrenia). If the genetic marker is strongly associated with very rapid death after onset, individuals with this marker will be less likely to be found among the prevalent cases (as they would have died sooner). This underrepresentation of exposed individuals among the prevalent cases will weaken or obscure the true association between the marker and the risk of developing schizophrenia. Studying incident (newly diagnosed) cases is preferred for investigating causation.
    • Response Bias (Volunteer Bias): This occurs when the individuals who agree to participate in a study (responders) are systematically different from those who decline (non-responders). The characteristics of the study sample may then differ significantly from the target population.
      • Example: A study invites people from the community to participate in research on stress levels, requiring completion of questionnaires and a clinic visit. Individuals who volunteer might be more health-conscious, have more free time, or be less stressed than those who do not volunteer (who might be too busy or stressed to participate). If the study aims to estimate average stress levels in the general population, relying solely on these volunteers could lead to an underestimation of the true average stress level. This is especially relevant when evaluating screening tests, as volunteers might have different underlying disease prevalence or health behaviours than the general population.
    • Unmasking Bias (Detection Signal Bias): This occurs when an exposure does not directly cause a disease but leads to symptoms or signs that trigger detection of a pre-existing, previously undiagnosed condition. The exposure appears to be associated with the disease simply because it brought the disease to light.
      • Example: A medication (exposure) is known to sometimes cause mild uterine spotting. A study investigates if this medication increases the risk of endometrial cancer (disease). It's possible that women taking the medication who experience spotting are more likely to seek medical attention and undergo investigations (like biopsies) compared to women not taking the medication. These investigations might detect pre-existing, asymptomatic endometrial cancer more frequently in the exposed group. The medication didn't cause the cancer but 'unmasked' it, leading to a biased association suggesting it did.
    • Lead-Time Bias: This bias specifically affects the evaluation of screening programs and survival time. Lead-time is the extra time a patient is considered 'alive with the disease' simply because the disease was detected earlier by screening, compared to when it would have been diagnosed based on symptoms alone. If this extra time is not accounted for, screening programs may appear to improve survival even if they don't actually extend life.
      • Example: A screening test detects a slow-growing cancer 2 years earlier than it would have been diagnosed based on symptoms. Patients diagnosed via screening will appear to survive, say, 7 years post-diagnosis, while those diagnosed symptomatically survive 5 years post-diagnosis. This difference of 2 years might be entirely due to the earlier diagnosis (the lead time) and not because the screening led to more effective treatment or actually prolonged life. Comparing survival times directly without adjusting for lead time results in lead-time bias, making the screening seem more beneficial than it is.
    • Referral Bias: This bias arises because the characteristics of patients seen in specialized settings (like tertiary care hospitals) often differ significantly from those seen in primary care or the general community, particularly regarding the prevalence of rare exposures or diseases, or disease severity.
      • Example: A study on the effectiveness of a highly specialized therapy for treatment-resistant depression is conducted at a national referral centre. The patients referred to this centre likely have more severe, complex, or co-morbid conditions than the average person with depression seen by a general practitioner. The results of the therapy in this highly selected group might not be generalizable to the broader population of depressed patients.
    • Diagnostic Purity Bias: This occurs when studies, particularly RCTs, exclude participants who have co-morbid conditions to create a 'pure' sample focusing only on the primary diagnosis. While this might increase internal consistency, it results in a study population that is not representative of real-world patients, who often have multiple health issues.
      • Example: An RCT for a new anxiety medication excludes individuals who also have a diagnosis of depression or a substance use disorder. The findings regarding the medication's efficacy and safety might not apply to the many patients in clinical practice who present with anxiety alongside these common co-morbidities.
    • Membership Bias: This bias arises when study participants are recruited from specific groups or organizations (e.g., patient advocacy groups, online forums, self-help groups). Members of such groups often differ systematically (e.g., in disease severity, health literacy, socioeconomic status, treatment adherence) from the general population of individuals with the condition.
      • Example: A study assessing the quality of life of individuals with bipolar disorder recruits participants through a local support group. Members of this group might be more proactive about managing their illness or have better social support compared to individuals with bipolar disorder who are not part of such groups. Findings based on this sample may not accurately reflect the quality of life of the broader population with bipolar disorder.

    Measurement Bias (Information Bias)

    Measurement bias occurs when the methods used to collect data (measuring exposures, outcomes, or other participant information) are flawed and result in systematic errors. The information gathered is inaccurate or varies systematically between comparison groups.

    • General Example: In a study comparing outcomes between patients interviewed in a quiet clinic room (Group A) and patients interviewed over a noisy telephone line (Group B), the quality and detail of information gathered might differ systematically. If Group A consistently provides more detailed symptom reports, this difference in measurement method introduces bias. Blinding (keeping data collectors and/or participants unaware of group allocation or study hypotheses) is an important strategy to minimize measurement bias.

    Specific types of measurement bias include:

    • Recall Bias: This occurs when there are systematic differences in how accurately participants recall past exposures or events, and this difference is related to their current health status (e.g., having the disease vs. not having it). It is a major concern in case-control studies.
      • Example: In a case-control study investigating whether stressful life events in childhood (exposure) are linked to developing depression later in life (disease), individuals who currently have depression (cases) might search their memories more thoroughly for past negative events or interpret ambiguous past events more negatively than individuals without depression (controls). This differential recall can create a spurious association or exaggerate a true one.
    • Reporting Bias: This bias occurs when participants are systematically more or less likely to report certain information (e.g., exposures or outcomes) due to social pressures, beliefs, or personal attitudes.
      • Example: In a study asking about illicit drug use (exposure) and mental health problems (outcome), participants might under-report their drug use due to fear of stigma or legal consequences. If this under-reporting differs between those with and without mental health problems, it introduces reporting bias.
    • Observer Bias: This occurs when the person assessing the outcome or measuring variables (the observer or researcher) has expectations or knowledge that systematically influences how they interpret or record the information. This influence can be conscious or subconscious.
      • Example: In an RCT of a new medication for psychosis, the psychiatrist rating symptom severity (the observer) knows which patients are receiving the new drug and which are receiving placebo. Believing the new drug is effective, the psychiatrist might subconsciously rate symptoms as less severe in the active treatment group or probe more for improvement compared to the placebo group. Blinding the observer to the treatment allocation is crucial to prevent this.
    • Surveillance Bias (Detection Bias): This bias arises when one group of participants is monitored more closely or undergoes more frequent diagnostic testing than another group, leading to a higher chance of detecting the outcome in the more closely monitored group, irrespective of a true difference in risk.
      • Example: A study follows up women who used oral contraceptives (exposed group) and women who never used them (unexposed group) to assess the risk of a certain health condition. If the women who used oral contraceptives have more regular medical check-ups (perhaps related to obtaining prescriptions), they might have a higher chance of having the condition detected, even if their actual risk of developing it is no different from the unexposed group.
    • Work-up Bias (Verification Bias): This bias occurs specifically in studies evaluating the accuracy of a diagnostic test. It happens when the decision to perform the 'gold standard' or reference test (which confirms the true disease status) is influenced by the result of the new test being evaluated.
      • Example: A study assesses a new, non-invasive screening test for early-stage Alzheimer's disease. The gold standard confirmation requires a more invasive procedure (e.g., lumbar puncture or PET scan). If doctors are less likely to refer patients for the gold standard test when the new screening test result is negative, but more likely when it's positive, the accuracy of the new test (especially its sensitivity and negative predictive value) will be overestimated. The sample verified by the gold standard is not representative of all patients who underwent the new test.
    • Misclassification Bias: This occurs when participants are incorrectly categorized regarding their exposure status, outcome status, or both.
      • Differential Misclassification: The error in classification is different between the comparison groups (e.g., cases are more likely to be misclassified regarding exposure than controls, or exposed individuals are more likely to be misclassified regarding outcome than unexposed individuals). This type of misclassification is a bias and can lead to either an overestimation or an underestimation of the true association, potentially even reversing the direction.
        • Example: In a case-control study, if interviewers probe cases more intensely about past exposures than controls (a form of observer bias), cases might be more likely to be wrongly classified as 'exposed' compared to controls. This differential error distorts the association.
      • Non-differential Misclassification: The error in classification occurs with roughly equal frequency in all comparison groups. For instance, an imprecise measurement tool might lead to some exposed individuals being classified as unexposed, and some unexposed individuals being classified as exposed, and this happens similarly whether they develop the disease or not. Non-differential misclassification typically biases the results toward the null, making the observed association weaker than the true association. It reduces the study's ability to detect a true effect but doesn't usually create a spurious one.
        • Example: Using a self-report questionnaire with vague questions about diet might lead to both depressed and non-depressed participants inaccurately reporting their fat intake to a similar degree. This makes it harder to detect a true association between fat intake and depression.
    • Desirability Bias: This is a type of reporting bias where participants tend to answer questions in a way they believe is socially acceptable or favourable, rather than truthfully reflecting their behaviour or attitudes.
      • Example: When asked about alcohol consumption or adherence to medication in a research interview, participants might under-report alcohol intake or over-report adherence because they perceive heavy drinking or non-adherence as undesirable. This leads to inaccurate data collection.
    • Hawthorne Effect: This refers to a change in participants' behaviour simply because they are aware they are being observed or participating in a study. They might alter their behaviour to align with what they perceive as expected or 'normal', rather than acting naturally. This often occurs in observational studies, especially cross-sectional surveys using questionnaires.
      • Example: Workers in a factory might increase their productivity temporarily simply because researchers are present and observing them, not due to any specific intervention being tested. In a health context, participants in a diet study might adhere more strictly to healthy eating guidelines while they know their food intake is being recorded, compared to their usual behaviour.

    Analysis Bias

    Analysis bias occurs during the data analysis phase, often due to how participants are handled after randomization, particularly if they do not adhere to their assigned group or are lost from the study.

    • Contamination Bias: This occurs when participants in one group inadvertently receive the intervention meant for the other group (e.g., controls start using the active treatment). This reduces the difference between the groups.
      • Example: In a trial comparing a new therapy program (Group A) versus usual care (Group B), some participants in Group B might seek out and receive elements of the new therapy program from outside the study.
    • Attrition Bias: This occurs when participants who drop out of the study (lost to follow-up) are systematically different from those who remain. If the reasons for dropping out are related to both the exposure/intervention and the outcome, the results from the remaining participants may be biased.
      • Example: In an RCT for an antidepressant, if participants experiencing severe side effects from the active drug are more likely to drop out than those on placebo or those experiencing benefits, the final analysis based only on completers might overestimate the drug's effectiveness and underestimate its side effects.

    Minimizing Analysis Bias: Intention-to-Treat (ITT)

    A common method to minimize bias arising from contamination and attrition during analysis is Intention-to-Treat (ITT) analysis. In ITT, participants are analyzed according to the group they were originally randomized to, regardless of whether they adhered to the treatment, switched groups, or dropped out. This approach preserves the benefits of randomization and provides a more realistic estimate of the intervention's effect in a real-world setting where adherence is not perfect.

    Bias Concepts

    0% Complete