00
Correct
00
Incorrect
00 : 00 : 00
Session Time
00 : 00
Average Question Time ( Mins)
  • Question 1 - Which of the following statements accurately describes significance tests? ...

    Incorrect

    • Which of the following statements accurately describes significance tests?

      Your Answer: The dependent variable is something that an experimenter purposely changes over the course of a study

      Correct Answer: Chi-squared test is used to compare non-parametric data

      Explanation:

      The chi-squared test is a statistical test that does not rely on any assumptions about the underlying distribution of the data, making it a non-parametric test.

      Choosing the right statistical test can be challenging, but understanding the basic principles can help. Different tests have different assumptions, and using the wrong one can lead to inaccurate results. To identify the appropriate test, a flow chart can be used based on three main factors: the type of dependent variable, the type of data, and whether the groups/samples are independent of dependent. It is important to know which tests are parametric and non-parametric, as well as their alternatives. For example, the chi-squared test is used to assess differences in categorical variables and is non-parametric, while Pearson’s correlation coefficient measures linear correlation between two variables and is parametric. T-tests are used to compare means between two groups, and ANOVA is used to compare means between more than two groups. Non-parametric equivalents to ANOVA include the Kruskal-Wallis analysis of ranks, the Median test, Friedman’s two-way analysis of variance, and Cochran Q test. Understanding these tests and their assumptions can help researchers choose the appropriate statistical test for their data.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      27.5
      Seconds
  • Question 2 - Which odds ratio suggests that there is no significant variation in the odds...

    Incorrect

    • Which odds ratio suggests that there is no significant variation in the odds between two groups?

      Your Answer: 2

      Correct Answer: 1

      Explanation:

      Measures of Effect in Clinical Studies

      When conducting clinical studies, we often want to know the effect of treatments of exposures on health outcomes. Measures of effect are used in randomized controlled trials (RCTs) and include the odds ratio (of), risk ratio (RR), risk difference (RD), and number needed to treat (NNT). Dichotomous (binary) outcome data are common in clinical trials, where the outcome for each participant is one of two possibilities, such as dead of alive, of clinical improvement of no improvement.

      To understand the difference between of and RR, it’s important to know the difference between risks and odds. Risk is a proportion that describes the probability of a health outcome occurring, while odds is a ratio that compares the probability of an event occurring to the probability of it not occurring. Absolute risk is the basic risk, while risk difference is the difference between the absolute risk of an event in the intervention group and the absolute risk in the control group. Relative risk is the ratio of risk in the intervention group to the risk in the control group.

      The number needed to treat (NNT) is the number of patients who need to be treated for one to benefit. Odds are calculated by dividing the number of times an event happens by the number of times it does not happen. The odds ratio is the odds of an outcome given a particular exposure versus the odds of an outcome in the absence of the exposure. It is commonly used in case-control studies and can also be used in cross-sectional and cohort study designs. An odds ratio of 1 indicates no difference in risk between the two groups, while an odds ratio >1 indicates an increased risk and an odds ratio <1 indicates a reduced risk.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      59
      Seconds
  • Question 3 - Which study design is always considered observational? ...

    Incorrect

    • Which study design is always considered observational?

      Your Answer: Case study

      Correct Answer: Cohort study

      Explanation:

      Case-studies and case-series can have an experimental nature due to the potential involvement of interventions of treatments.

      Types of Primary Research Studies and Their Advantages and Disadvantages

      Primary research studies can be categorized into six types based on the research question they aim to address. The best type of study for each question type is listed in the table below. There are two main types of study design: experimental and observational. Experimental studies involve an intervention, while observational studies do not. The advantages and disadvantages of each study type are summarized in the table below.

      Type of Question Best Type of Study

      Therapy Randomized controlled trial (RCT), cohort, case control, case series
      Diagnosis Cohort studies with comparison to gold standard test
      Prognosis Cohort studies, case control, case series
      Etiology/Harm RCT, cohort studies, case control, case series
      Prevention RCT, cohort studies, case control, case series
      Cost Economic analysis

      Study Type Advantages Disadvantages

      Randomized Controlled Trial – Unbiased distribution of confounders – Blinding more likely – Randomization facilitates statistical analysis – Expensive – Time-consuming – Volunteer bias – Ethically problematic at times
      Cohort Study – Ethically safe – Subjects can be matched – Can establish timing and directionality of events – Eligibility criteria and outcome assessments can be standardized – Administratively easier and cheaper than RCT – Controls may be difficult to identify – Exposure may be linked to a hidden confounder – Blinding is difficult – Randomization not present – For rare disease, large sample sizes of long follow-up necessary
      Case-Control Study – Quick and cheap – Only feasible method for very rare disorders of those with long lag between exposure and outcome – Fewer subjects needed than cross-sectional studies – Reliance on recall of records to determine exposure status – Confounders – Selection of control groups is difficult – Potential bias: recall, selection
      Cross-Sectional Survey – Cheap and simple – Ethically safe – Establishes association at most, not causality – Recall bias susceptibility – Confounders may be unequally distributed – Neyman bias – Group sizes may be unequal
      Ecological Study – Cheap and simple – Ethically safe – Ecological fallacy (when relationships which exist for groups are assumed to also be true for individuals)

      In conclusion, the choice of study type depends on the research question being addressed. Each study type has its own advantages and disadvantages, and researchers should carefully consider these when designing their studies.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      18.1
      Seconds
  • Question 4 - For which of the following research areas are qualitative methods least effective? ...

    Incorrect

    • For which of the following research areas are qualitative methods least effective?

      Your Answer: Finding out user views

      Correct Answer: Treatment evaluation

      Explanation:

      While quantitative methods are typically used for treatment evaluation, qualitative studies can also provide valuable insights by interpreting, qualifying, of illuminating findings. This is especially beneficial when examining unexpected results, as they can help to test the primary hypothesis.

      Qualitative research is a method of inquiry that seeks to understand the meaning and experience dimensions of human lives and social worlds. There are different approaches to qualitative research, such as ethnography, phenomenology, and grounded theory, each with its own purpose, role of the researcher, stages of research, and method of data analysis. The most common methods used in healthcare research are interviews and focus groups. Sampling techniques include convenience sampling, purposive sampling, quota sampling, snowball sampling, and case study sampling. Sample size can be determined by data saturation, which occurs when new categories, themes, of explanations stop emerging from the data. Validity can be assessed through triangulation, respondent validation, bracketing, and reflexivity. Analytical approaches include content analysis and constant comparison.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      138.3
      Seconds
  • Question 5 - Which study design involves conducting an experiment? ...

    Incorrect

    • Which study design involves conducting an experiment?

      Your Answer: A cross-sectional study

      Correct Answer: A randomised control study

      Explanation:

      Types of Primary Research Studies and Their Advantages and Disadvantages

      Primary research studies can be categorized into six types based on the research question they aim to address. The best type of study for each question type is listed in the table below. There are two main types of study design: experimental and observational. Experimental studies involve an intervention, while observational studies do not. The advantages and disadvantages of each study type are summarized in the table below.

      Type of Question Best Type of Study

      Therapy Randomized controlled trial (RCT), cohort, case control, case series
      Diagnosis Cohort studies with comparison to gold standard test
      Prognosis Cohort studies, case control, case series
      Etiology/Harm RCT, cohort studies, case control, case series
      Prevention RCT, cohort studies, case control, case series
      Cost Economic analysis

      Study Type Advantages Disadvantages

      Randomized Controlled Trial – Unbiased distribution of confounders – Blinding more likely – Randomization facilitates statistical analysis – Expensive – Time-consuming – Volunteer bias – Ethically problematic at times
      Cohort Study – Ethically safe – Subjects can be matched – Can establish timing and directionality of events – Eligibility criteria and outcome assessments can be standardized – Administratively easier and cheaper than RCT – Controls may be difficult to identify – Exposure may be linked to a hidden confounder – Blinding is difficult – Randomization not present – For rare disease, large sample sizes of long follow-up necessary
      Case-Control Study – Quick and cheap – Only feasible method for very rare disorders of those with long lag between exposure and outcome – Fewer subjects needed than cross-sectional studies – Reliance on recall of records to determine exposure status – Confounders – Selection of control groups is difficult – Potential bias: recall, selection
      Cross-Sectional Survey – Cheap and simple – Ethically safe – Establishes association at most, not causality – Recall bias susceptibility – Confounders may be unequally distributed – Neyman bias – Group sizes may be unequal
      Ecological Study – Cheap and simple – Ethically safe – Ecological fallacy (when relationships which exist for groups are assumed to also be true for individuals)

      In conclusion, the choice of study type depends on the research question being addressed. Each study type has its own advantages and disadvantages, and researchers should carefully consider these when designing their studies.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      17.2
      Seconds
  • Question 6 - What study method would be most suitable for a researcher tasked with comparing...

    Incorrect

    • What study method would be most suitable for a researcher tasked with comparing the cost-effectiveness of olanzapine and haloperidol in reducing symptom severity of schizophrenia, as measured by the Positive and Negative Syndrome Scale?

      Your Answer:

      Correct Answer: Cost-effectiveness analysis

      Explanation:

      The task assigned to the researcher is to conduct a cost-effectiveness analysis, which involves comparing two interventions based on their costs and their impact on a single clinical measure of effectiveness, specifically the reduction in symptom severity as measured by the PANSS.

      Methods of Economic Evaluation

      There are four main methods of economic evaluation: cost-effectiveness analysis (CEA), cost-benefit analysis (CBA), cost-utility analysis (CUA), and cost-minimisation analysis (CMA). While all four methods capture costs, they differ in how they assess health effects.

      Cost-effectiveness analysis (CEA) compares interventions by relating costs to a single clinical measure of effectiveness, such as symptom reduction of improvement in activities of daily living. The cost-effectiveness ratio is calculated as total cost divided by units of effectiveness. CEA is typically used when CBA cannot be performed due to the inability to monetise benefits.

      Cost-benefit analysis (CBA) measures all costs and benefits of an intervention in monetary terms to establish which alternative has the greatest net benefit. CBA requires that all consequences of an intervention, such as life-years saved, treatment side-effects, symptom relief, disability, pain, and discomfort, are allocated a monetary value. CBA is rarely used in mental health service evaluation due to the difficulty in converting benefits from mental health programmes into monetary values.

      Cost-utility analysis (CUA) is a special form of CEA in which health benefits/outcomes are measured in broader, more generic ways, enabling comparisons between treatments for different diseases and conditions. Multidimensional health outcomes are measured by a single preference- of utility-based index such as the Quality-Adjusted-Life-Years (QALY). QALYs are a composite measure of gains in life expectancy and health-related quality of life. CUA allows for comparisons across treatments for different conditions.

      Cost-minimisation analysis (CMA) is an economic evaluation in which the consequences of competing interventions are the same, and only inputs, i.e. costs, are taken into consideration. The aim is to decide the least costly way of achieving the same outcome.

      Costs in Economic Evaluation Studies

      There are three main types of costs in economic evaluation studies: direct, indirect, and intangible. Direct costs are associated directly with the healthcare intervention, such as staff time, medical supplies, cost of travel for the patient, childcare costs for the patient, and costs falling on other social sectors such as domestic help from social services. Indirect costs are incurred by the reduced productivity of the patient, such as time off work, reduced work productivity, and time spent caring for the patient by relatives. Intangible costs are difficult to measure, such as pain of suffering on the part of the patient.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 7 - What is the accurate formula for determining the pre-test odds? ...

    Incorrect

    • What is the accurate formula for determining the pre-test odds?

      Your Answer:

      Correct Answer: Pre-test probability/ (1 - pre-test probability)

      Explanation:

      Clinical tests are used to determine the presence of absence of a disease of condition. To interpret test results, it is important to have a working knowledge of statistics used to describe them. Two by two tables are commonly used to calculate test statistics such as sensitivity and specificity. Sensitivity refers to the proportion of people with a condition that the test correctly identifies, while specificity refers to the proportion of people without a condition that the test correctly identifies. Accuracy tells us how closely a test measures to its true value, while predictive values help us understand the likelihood of having a disease based on a positive of negative test result. Likelihood ratios combine sensitivity and specificity into a single figure that can refine our estimation of the probability of a disease being present. Pre and post-test odds and probabilities can also be calculated to better understand the likelihood of having a disease before and after a test is carried out. Fagan’s nomogram is a useful tool for calculating post-test probabilities.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 8 - A case-control study was conducted to determine if exposure to passive smoking during...

    Incorrect

    • A case-control study was conducted to determine if exposure to passive smoking during childhood increases the risk of nicotine dependence. Two groups were recruited: 200 patients with nicotine dependence and 200 controls without nicotine dependence. Among the patients, 40 reported exposure to parental smoking during childhood, while among the controls, 20 reported such exposure. The odds ratio of developing nicotine dependence after being exposed to passive smoking is:

      Your Answer:

      Correct Answer: 2.25

      Explanation:

      Measures of Effect in Clinical Studies

      When conducting clinical studies, we often want to know the effect of treatments of exposures on health outcomes. Measures of effect are used in randomized controlled trials (RCTs) and include the odds ratio (of), risk ratio (RR), risk difference (RD), and number needed to treat (NNT). Dichotomous (binary) outcome data are common in clinical trials, where the outcome for each participant is one of two possibilities, such as dead of alive, of clinical improvement of no improvement.

      To understand the difference between of and RR, it’s important to know the difference between risks and odds. Risk is a proportion that describes the probability of a health outcome occurring, while odds is a ratio that compares the probability of an event occurring to the probability of it not occurring. Absolute risk is the basic risk, while risk difference is the difference between the absolute risk of an event in the intervention group and the absolute risk in the control group. Relative risk is the ratio of risk in the intervention group to the risk in the control group.

      The number needed to treat (NNT) is the number of patients who need to be treated for one to benefit. Odds are calculated by dividing the number of times an event happens by the number of times it does not happen. The odds ratio is the odds of an outcome given a particular exposure versus the odds of an outcome in the absence of the exposure. It is commonly used in case-control studies and can also be used in cross-sectional and cohort study designs. An odds ratio of 1 indicates no difference in risk between the two groups, while an odds ratio >1 indicates an increased risk and an odds ratio <1 indicates a reduced risk.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 9 - Which statement about disease rates is incorrect? ...

    Incorrect

    • Which statement about disease rates is incorrect?

      Your Answer:

      Correct Answer: The odds ratio is synonymous with the risk ratio

      Explanation:

      Disease Rates and Their Interpretation

      Disease rates are a measure of the occurrence of a disease in a population. They are used to establish causation, monitor interventions, and measure the impact of exposure on disease rates. The attributable risk is the difference in the rate of disease between the exposed and unexposed groups. It tells us what proportion of deaths in the exposed group were due to the exposure. The relative risk is the risk of an event relative to exposure. It is calculated by dividing the rate of disease in the exposed group by the rate of disease in the unexposed group. A relative risk of 1 means there is no difference between the two groups. A relative risk of <1 means that the event is less likely to occur in the exposed group, while a relative risk of >1 means that the event is more likely to occur in the exposed group. The population attributable risk is the reduction in incidence that would be observed if the population were entirely unexposed. It can be calculated by multiplying the attributable risk by the prevalence of exposure in the population. The attributable proportion is the proportion of the disease that would be eliminated in a population if its disease rate were reduced to that of the unexposed group.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 10 - An endocrinologist conducts a study to determine if there is a correlation between...

    Incorrect

    • An endocrinologist conducts a study to determine if there is a correlation between a patient's age and their blood pressure. Assuming both age and blood pressure are normally distributed, what statistical test would be most suitable to use?

      Your Answer:

      Correct Answer: Pearson's product-moment coefficient

      Explanation:

      Since the data is normally distributed and the study aims to evaluate the correlation between two variables, the most suitable test to use is Pearson’s product-moment coefficient. On the other hand, if the data is non-parametric, Spearman’s coefficient would be more appropriate.

      Choosing the right statistical test can be challenging, but understanding the basic principles can help. Different tests have different assumptions, and using the wrong one can lead to inaccurate results. To identify the appropriate test, a flow chart can be used based on three main factors: the type of dependent variable, the type of data, and whether the groups/samples are independent of dependent. It is important to know which tests are parametric and non-parametric, as well as their alternatives. For example, the chi-squared test is used to assess differences in categorical variables and is non-parametric, while Pearson’s correlation coefficient measures linear correlation between two variables and is parametric. T-tests are used to compare means between two groups, and ANOVA is used to compare means between more than two groups. Non-parametric equivalents to ANOVA include the Kruskal-Wallis analysis of ranks, the Median test, Friedman’s two-way analysis of variance, and Cochran Q test. Understanding these tests and their assumptions can help researchers choose the appropriate statistical test for their data.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 11 - What is the appropriate denominator for calculating the incidence rate? ...

    Incorrect

    • What is the appropriate denominator for calculating the incidence rate?

      Your Answer:

      Correct Answer: The total person time at risk during a specified time period

      Explanation:

      Measures of Disease Frequency: Incidence and Prevalence

      Incidence and prevalence are two important measures of disease frequency. Incidence measures the speed at which new cases of a disease are emerging, while prevalence measures the burden of disease within a population. Cumulative incidence and incidence rate are two types of incidence measures, while point prevalence and period prevalence are two types of prevalence measures.

      Cumulative incidence is the average risk of getting a disease over a certain period of time, while incidence rate is a measure of the speed at which new cases are emerging. Prevalence is a proportion and is a measure of the burden of disease within a population. Point prevalence measures the number of cases in a defined population at a specific point in time, while period prevalence measures the number of identified cases during a specified period of time.

      It is important to note that prevalence is equal to incidence multiplied by the duration of the condition. In chronic diseases, the prevalence is much greater than the incidence. The incidence rate is stated in units of person-time, while cumulative incidence is always a proportion. When describing cumulative incidence, it is necessary to give the follow-up period over which the risk is estimated. In acute diseases, the prevalence and incidence may be similar, while for conditions such as the common cold, the incidence may be greater than the prevalence.

      Incidence is a useful measure to study disease etiology and risk factors, while prevalence is useful for health resource planning. Understanding these measures of disease frequency is important for public health professionals and researchers in order to effectively monitor and address the burden of disease within populations.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 12 - Which data type does age in years belong to? ...

    Incorrect

    • Which data type does age in years belong to?

      Your Answer:

      Correct Answer: Ratio

      Explanation:

      Age is a type of measurement that follows a ratio scale, which means that the values can be compared as multiples of each other. For instance, if someone is 20 years old, they are twice as old as someone who is 10 years old.

      Scales of Measurement in Statistics

      In the 1940s, Stanley Smith Stevens introduced four scales of measurement to categorize data variables. Knowing the scale of measurement for a variable is crucial in selecting the appropriate statistical analysis. The four scales of measurement are ratio, interval, ordinal, and nominal.

      Ratio scales are similar to interval scales, but they have true zero points. Examples of ratio scales include weight, time, and length. Interval scales measure the difference between two values, and one unit on the scale represents the same magnitude on the trait of characteristic being measured across the whole range of the scale. The Fahrenheit scale for temperature is an example of an interval scale.

      Ordinal scales categorize observed values into set categories that can be ordered, but the intervals between each value are uncertain. Examples of ordinal scales include social class, education level, and income level. Nominal scales categorize observed values into set categories that have no particular order of hierarchy. Examples of nominal scales include genotype, blood type, and political party.

      Data can also be categorized as quantitative of qualitative. Quantitative variables take on numeric values and can be further classified into discrete and continuous types. Qualitative variables do not take on numerical values and are usually names. Some qualitative variables have an inherent order in their categories and are described as ordinal. Qualitative variables are also called categorical of nominal variables. When a qualitative variable has only two categories, it is called a binary variable.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 13 - Which of the following is an example of selection bias? ...

    Incorrect

    • Which of the following is an example of selection bias?

      Your Answer:

      Correct Answer: Berkson's bias

      Explanation:

      Types of Bias in Statistics

      Bias is a systematic error that can lead to incorrect conclusions. Confounding factors are variables that are associated with both the outcome and the exposure but have no causative role. Confounding can be addressed in the design and analysis stage of a study. The main method of controlling confounding in the analysis phase is stratification analysis. The main methods used in the design stage are matching, randomization, and restriction of participants.

      There are two main types of bias: selection bias and information bias. Selection bias occurs when the selected sample is not a representative sample of the reference population. Disease spectrum bias, self-selection bias, participation bias, incidence-prevalence bias, exclusion bias, publication of dissemination bias, citation bias, and Berkson’s bias are all subtypes of selection bias. Information bias occurs when gathered information about exposure, outcome, of both is not correct and there was an error in measurement. Detection bias, recall bias, lead time bias, interviewer/observer bias, verification and work-up bias, Hawthorne effect, and ecological fallacy are all subtypes of information bias.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 14 - What is the term used to describe how a person's age affects their...

    Incorrect

    • What is the term used to describe how a person's age affects their likelihood of reporting past exposure to a certain risk factor?

      Your Answer:

      Correct Answer: Recall bias

      Explanation:

      Recall bias pertains to how a person’s illness status can influence their tendency to report past exposure to a risk factor. Confounding arises when an additional variable is associated with both an independent and dependent variable. Observer bias refers to the possibility that researchers’ cognitive biases may unconsciously impact the results of a study. Publication bias refers to the tendency for studies with positive results to be more likely to be published. Selection bias occurs when certain individuals of groups are overrepresented, leading to inadequate randomization.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 15 - What is necessary for a study to confidently assert causation? ...

    Incorrect

    • What is necessary for a study to confidently assert causation?

      Your Answer:

      Correct Answer: Good internal validity

      Explanation:

      In order to make assertions about causation, strong internal validity is necessary.

      Validity in statistics refers to how accurately something measures what it claims to measure. There are two main types of validity: internal and external. Internal validity refers to the confidence we have in the cause and effect relationship in a study, while external validity refers to the degree to which the conclusions of a study can be applied to other people, places, and times. There are various threats to both internal and external validity, such as sampling, measurement instrument obtrusiveness, and reactive effects of setting. Additionally, there are several subtypes of validity, including face validity, content validity, criterion validity, and construct validity. Each subtype has its own specific focus and methods for testing validity.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 16 - A study comparing the benefit of two surgical procedures for patients over 65...

    Incorrect

    • A study comparing the benefit of two surgical procedures for patients over 65 concludes that the two procedures are equally effective. A researcher is then asked to conduct a cost analysis of the two procedures, considering only the financial expenses.

      What is the best way to describe this approach?

      Your Answer:

      Correct Answer: Cost-minimisation analysis

      Explanation:

      Methods of Economic Evaluation

      There are four main methods of economic evaluation: cost-effectiveness analysis (CEA), cost-benefit analysis (CBA), cost-utility analysis (CUA), and cost-minimisation analysis (CMA). While all four methods capture costs, they differ in how they assess health effects.

      Cost-effectiveness analysis (CEA) compares interventions by relating costs to a single clinical measure of effectiveness, such as symptom reduction of improvement in activities of daily living. The cost-effectiveness ratio is calculated as total cost divided by units of effectiveness. CEA is typically used when CBA cannot be performed due to the inability to monetise benefits.

      Cost-benefit analysis (CBA) measures all costs and benefits of an intervention in monetary terms to establish which alternative has the greatest net benefit. CBA requires that all consequences of an intervention, such as life-years saved, treatment side-effects, symptom relief, disability, pain, and discomfort, are allocated a monetary value. CBA is rarely used in mental health service evaluation due to the difficulty in converting benefits from mental health programmes into monetary values.

      Cost-utility analysis (CUA) is a special form of CEA in which health benefits/outcomes are measured in broader, more generic ways, enabling comparisons between treatments for different diseases and conditions. Multidimensional health outcomes are measured by a single preference- of utility-based index such as the Quality-Adjusted-Life-Years (QALY). QALYs are a composite measure of gains in life expectancy and health-related quality of life. CUA allows for comparisons across treatments for different conditions.

      Cost-minimisation analysis (CMA) is an economic evaluation in which the consequences of competing interventions are the same, and only inputs, i.e. costs, are taken into consideration. The aim is to decide the least costly way of achieving the same outcome.

      Costs in Economic Evaluation Studies

      There are three main types of costs in economic evaluation studies: direct, indirect, and intangible. Direct costs are associated directly with the healthcare intervention, such as staff time, medical supplies, cost of travel for the patient, childcare costs for the patient, and costs falling on other social sectors such as domestic help from social services. Indirect costs are incurred by the reduced productivity of the patient, such as time off work, reduced work productivity, and time spent caring for the patient by relatives. Intangible costs are difficult to measure, such as pain of suffering on the part of the patient.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 17 - If a study has a Type I error rate of <0.05 and a...

    Incorrect

    • If a study has a Type I error rate of <0.05 and a Type II error rate of 0.2, what is the power of the study?

      Your Answer:

      Correct Answer: 0.8

      Explanation:

      A study’s ability to correctly detect a true effect of difference may be calculated as Power = 1 – Type II error rate. In the given scenario, the power can be calculated as Power = 1 – 0.2 = 0.8. Type I error refers to a false positive, while Type II error refers to a false negative.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 18 - A team of scientists aimed to examine the prognosis of late-onset Alzheimer's disease...

    Incorrect

    • A team of scientists aimed to examine the prognosis of late-onset Alzheimer's disease using the available evidence. They intend to arrange the evidence in a hierarchy based on their study designs.
      What study design would be placed at the top of their hierarchy?

      Your Answer:

      Correct Answer: Systematic review of cohort studies

      Explanation:

      When investigating prognosis, the hierarchy of study designs starts with a systematic review of cohort studies, followed by a cohort study, follow-up of untreated patients from randomized controlled trials, case series, and expert opinion. The strength of evidence provided by a study depends on its ability to minimize bias and maximize attribution. The Agency for Healthcare Policy and Research hierarchy of study types is widely accepted as reliable, with systematic reviews and meta-analyses of randomized controlled trials at the top, followed by randomized controlled trials, non-randomized intervention studies, observational studies, and non-experimental studies.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 19 - What is the estimated range for the 95% confidence interval for the mean...

    Incorrect

    • What is the estimated range for the 95% confidence interval for the mean glucose levels in a population of people taking antipsychotics, given a sample mean of 7 mmol/L, a sample standard deviation of 6 mmol/L, and a sample size of 9 with a standard error of the mean of 2 mmol/L?

      Your Answer:

      Correct Answer: 3-11 mmol/L

      Explanation:

      It is important to note that confidence intervals are derived from standard errors, not standard deviation, despite the common misconception. It is crucial to avoid mixing up these two terms.

      Measures of dispersion are used to indicate the variation of spread of a data set, often in conjunction with a measure of central tendency such as the mean of median. The range, which is the difference between the largest and smallest value, is the simplest measure of dispersion. The interquartile range, which is the difference between the 3rd and 1st quartiles, is another useful measure. Quartiles divide a data set into quarters, and the interquartile range can provide additional information about the spread of the data. However, to get a more representative idea of spread, measures such as the variance and standard deviation are needed. The variance gives an indication of how much the items in the data set vary from the mean, while the standard deviation reflects the distribution of individual scores around their mean. The standard deviation is expressed in the same units as the data set and can be used to indicate how confident we are that data points lie within a particular range. The standard error of the mean is an inferential statistic used to estimate the population mean and is a measure of the spread expected for the mean of the observations. Confidence intervals are often presented alongside sample results such as the mean value, indicating a range that is likely to contain the true value.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 20 - You have been tasked with examining the potential advantage of establishing a program...

    Incorrect

    • You have been tasked with examining the potential advantage of establishing a program to assist elderly patients with panic disorder in the nearby region. What is the primary consideration in determining the amount of resources needed?

      Your Answer:

      Correct Answer: Prevalence

      Explanation:

      Measures of Disease Frequency: Incidence and Prevalence

      Incidence and prevalence are two important measures of disease frequency. Incidence measures the speed at which new cases of a disease are emerging, while prevalence measures the burden of disease within a population. Cumulative incidence and incidence rate are two types of incidence measures, while point prevalence and period prevalence are two types of prevalence measures.

      Cumulative incidence is the average risk of getting a disease over a certain period of time, while incidence rate is a measure of the speed at which new cases are emerging. Prevalence is a proportion and is a measure of the burden of disease within a population. Point prevalence measures the number of cases in a defined population at a specific point in time, while period prevalence measures the number of identified cases during a specified period of time.

      It is important to note that prevalence is equal to incidence multiplied by the duration of the condition. In chronic diseases, the prevalence is much greater than the incidence. The incidence rate is stated in units of person-time, while cumulative incidence is always a proportion. When describing cumulative incidence, it is necessary to give the follow-up period over which the risk is estimated. In acute diseases, the prevalence and incidence may be similar, while for conditions such as the common cold, the incidence may be greater than the prevalence.

      Incidence is a useful measure to study disease etiology and risk factors, while prevalence is useful for health resource planning. Understanding these measures of disease frequency is important for public health professionals and researchers in order to effectively monitor and address the burden of disease within populations.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 21 - A study is conducted to investigate whether a new exercise program has any...

    Incorrect

    • A study is conducted to investigate whether a new exercise program has any impact on weight loss. A total of 300 participants are enrolled from various locations and are randomly assigned to either the exercise group of the control group. Weight measurements are taken at the beginning of the study and at the end of a six-month period.

      What is the most effective method of visually presenting the data?

      Your Answer:

      Correct Answer: Kaplan-Meier plot

      Explanation:

      The Kaplan-Meier plot is the most effective graphical representation of survival probability. It presents the overall likelihood of an individual’s survival over time from a baseline, and the comparison of two lines on the plot can indicate whether there is a survival advantage. To determine if the distinction between the two groups is significant, a log rank test can be employed.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 22 - Which statement accurately describes box and whisker plots? ...

    Incorrect

    • Which statement accurately describes box and whisker plots?

      Your Answer:

      Correct Answer: Each whisker represents approximately 25% of the data

      Explanation:

      Box and whisker plots are a useful tool for displaying information about the range, median, and quartiles of a data set. The whiskers only contain values within 1.5 times the interquartile range (IQR), and any values outside of this range are considered outliers and displayed as dots. The IQR is the difference between the 3rd and 1st quartiles, which divide the data set into quarters. Quartiles can also be used to determine the percentage of observations that fall below a certain value. However, quartiles and ranges have limitations because they do not take into account every score in a data set. To get a more representative idea of spread, measures such as variance and standard deviation are needed. Box plots can also provide information about the shape of a data set, such as whether it is skewed or symmetric. Notched boxes on the plot represent the confidence intervals of the median values.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 23 - What type of bias could arise from using only one psychiatrist to diagnose...

    Incorrect

    • What type of bias could arise from using only one psychiatrist to diagnose all participants in a study?

      Your Answer:

      Correct Answer: Information bias

      Explanation:

      The scenario described above highlights the issue of information bias, which can arise due to errors in measuring, collecting, of interpreting data related to the exposure of disease. Specifically, interviewer/observer bias is a type of information bias that can occur when a single psychiatrist has a tendency to either over of under diagnose a condition, potentially skewing the study results.

      Types of Bias in Statistics

      Bias is a systematic error that can lead to incorrect conclusions. Confounding factors are variables that are associated with both the outcome and the exposure but have no causative role. Confounding can be addressed in the design and analysis stage of a study. The main method of controlling confounding in the analysis phase is stratification analysis. The main methods used in the design stage are matching, randomization, and restriction of participants.

      There are two main types of bias: selection bias and information bias. Selection bias occurs when the selected sample is not a representative sample of the reference population. Disease spectrum bias, self-selection bias, participation bias, incidence-prevalence bias, exclusion bias, publication of dissemination bias, citation bias, and Berkson’s bias are all subtypes of selection bias. Information bias occurs when gathered information about exposure, outcome, of both is not correct and there was an error in measurement. Detection bias, recall bias, lead time bias, interviewer/observer bias, verification and work-up bias, Hawthorne effect, and ecological fallacy are all subtypes of information bias.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 24 - What is the typical measure of outcome in a case-control study investigating the...

    Incorrect

    • What is the typical measure of outcome in a case-control study investigating the potential association between autism and a recently developed varicella vaccine?

      Your Answer:

      Correct Answer: Odds ratio

      Explanation:

      The odds ratio is used in case-control studies to measure the association between exposure and outcome, while the relative risk is used in cohort studies to measure the risk of developing an outcome in the exposed group compared to the unexposed group. To convert the odds ratio to a relative risk, one can use the formula: relative risk = odds ratio / (1 – incidence in the unexposed group x odds ratio).

      Types of Primary Research Studies and Their Advantages and Disadvantages

      Primary research studies can be categorized into six types based on the research question they aim to address. The best type of study for each question type is listed in the table below. There are two main types of study design: experimental and observational. Experimental studies involve an intervention, while observational studies do not. The advantages and disadvantages of each study type are summarized in the table below.

      Type of Question Best Type of Study

      Therapy Randomized controlled trial (RCT), cohort, case control, case series
      Diagnosis Cohort studies with comparison to gold standard test
      Prognosis Cohort studies, case control, case series
      Etiology/Harm RCT, cohort studies, case control, case series
      Prevention RCT, cohort studies, case control, case series
      Cost Economic analysis

      Study Type Advantages Disadvantages

      Randomized Controlled Trial – Unbiased distribution of confounders – Blinding more likely – Randomization facilitates statistical analysis – Expensive – Time-consuming – Volunteer bias – Ethically problematic at times
      Cohort Study – Ethically safe – Subjects can be matched – Can establish timing and directionality of events – Eligibility criteria and outcome assessments can be standardized – Administratively easier and cheaper than RCT – Controls may be difficult to identify – Exposure may be linked to a hidden confounder – Blinding is difficult – Randomization not present – For rare disease, large sample sizes of long follow-up necessary
      Case-Control Study – Quick and cheap – Only feasible method for very rare disorders of those with long lag between exposure and outcome – Fewer subjects needed than cross-sectional studies – Reliance on recall of records to determine exposure status – Confounders – Selection of control groups is difficult – Potential bias: recall, selection
      Cross-Sectional Survey – Cheap and simple – Ethically safe – Establishes association at most, not causality – Recall bias susceptibility – Confounders may be unequally distributed – Neyman bias – Group sizes may be unequal
      Ecological Study – Cheap and simple – Ethically safe – Ecological fallacy (when relationships which exist for groups are assumed to also be true for individuals)

      In conclusion, the choice of study type depends on the research question being addressed. Each study type has its own advantages and disadvantages, and researchers should carefully consider these when designing their studies.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 25 - What is a true statement about measures of effect? ...

    Incorrect

    • What is a true statement about measures of effect?

      Your Answer:

      Correct Answer: Relative risk can be used to measure effect in randomised control trials

      Explanation:

      The use of relative risk is applicable in cohort, cross-sectional, and randomized control trials, but not in case-control studies. In situations where there are no events in the control group, neither the risk ratio nor the odds ratio can be computed. It is important to note that the odds ratio tends to overestimate effects and is always more extreme than the relative risk, moving away from the null value of 1.

      Measures of Effect in Clinical Studies

      When conducting clinical studies, we often want to know the effect of treatments of exposures on health outcomes. Measures of effect are used in randomized controlled trials (RCTs) and include the odds ratio (of), risk ratio (RR), risk difference (RD), and number needed to treat (NNT). Dichotomous (binary) outcome data are common in clinical trials, where the outcome for each participant is one of two possibilities, such as dead of alive, of clinical improvement of no improvement.

      To understand the difference between of and RR, it’s important to know the difference between risks and odds. Risk is a proportion that describes the probability of a health outcome occurring, while odds is a ratio that compares the probability of an event occurring to the probability of it not occurring. Absolute risk is the basic risk, while risk difference is the difference between the absolute risk of an event in the intervention group and the absolute risk in the control group. Relative risk is the ratio of risk in the intervention group to the risk in the control group.

      The number needed to treat (NNT) is the number of patients who need to be treated for one to benefit. Odds are calculated by dividing the number of times an event happens by the number of times it does not happen. The odds ratio is the odds of an outcome given a particular exposure versus the odds of an outcome in the absence of the exposure. It is commonly used in case-control studies and can also be used in cross-sectional and cohort study designs. An odds ratio of 1 indicates no difference in risk between the two groups, while an odds ratio >1 indicates an increased risk and an odds ratio <1 indicates a reduced risk.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 26 - What is the purpose of using bracketing as a method in qualitative research?...

    Incorrect

    • What is the purpose of using bracketing as a method in qualitative research?

      Your Answer:

      Correct Answer: Assessing validity

      Explanation:

      Qualitative research is a method of inquiry that seeks to understand the meaning and experience dimensions of human lives and social worlds. There are different approaches to qualitative research, such as ethnography, phenomenology, and grounded theory, each with its own purpose, role of the researcher, stages of research, and method of data analysis. The most common methods used in healthcare research are interviews and focus groups. Sampling techniques include convenience sampling, purposive sampling, quota sampling, snowball sampling, and case study sampling. Sample size can be determined by data saturation, which occurs when new categories, themes, of explanations stop emerging from the data. Validity can be assessed through triangulation, respondent validation, bracketing, and reflexivity. Analytical approaches include content analysis and constant comparison.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 27 - Which of the following is not a valid type of validity? ...

    Incorrect

    • Which of the following is not a valid type of validity?

      Your Answer:

      Correct Answer: Internal consistency

      Explanation:

      Validity in statistics refers to how accurately something measures what it claims to measure. There are two main types of validity: internal and external. Internal validity refers to the confidence we have in the cause and effect relationship in a study, while external validity refers to the degree to which the conclusions of a study can be applied to other people, places, and times. There are various threats to both internal and external validity, such as sampling, measurement instrument obtrusiveness, and reactive effects of setting. Additionally, there are several subtypes of validity, including face validity, content validity, criterion validity, and construct validity. Each subtype has its own specific focus and methods for testing validity.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 28 - If a patient follows a new healthy eating campaign for 2 years, with...

    Incorrect

    • If a patient follows a new healthy eating campaign for 2 years, with an average weight loss of 18 kg and a standard deviation of 3 kg, what is the probability that their weight loss will fall between 9 and 27 kg?

      Your Answer:

      Correct Answer: 99.70%

      Explanation:

      The mean weight is 18kg with a standard deviation of 3kg. Three standard deviations below the mean is 9kg and three standard deviations above the mean is 27kg.

      Standard Deviation and Standard Error of the Mean

      Standard deviation (SD) and standard error of the mean (SEM) are two important statistical measures used to describe data. SD is a measure of how much the data varies, while SEM is a measure of how precisely we know the true mean of the population. The normal distribution, also known as the Gaussian distribution, is a symmetrical bell-shaped curve that describes the spread of many biological and clinical measurements.

      68.3% of the data lies within 1 SD of the mean, 95.4% of the data lies within 2 SD of the mean, and 99.7% of the data lies within 3 SD of the mean. The SD is calculated by taking the square root of the variance and is expressed in the same units as the data set. A low SD indicates that data points tend to be very close to the mean.

      On the other hand, SEM is an inferential statistic that quantifies the precision of the mean. It is expressed in the same units as the data and is calculated by dividing the SD of the sample mean by the square root of the sample size. The SEM gets smaller as the sample size increases, and it takes into account both the value of the SD and the sample size.

      Both SD and SEM are important measures in statistical analysis, and they are used to calculate confidence intervals and test hypotheses. While SD quantifies scatter, SEM quantifies precision, and both are essential in understanding and interpreting data.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 29 - What is the conventional cutoff for a p-value of 0.05 and what does...

    Incorrect

    • What is the conventional cutoff for a p-value of 0.05 and what does it mean in terms of the likelihood of detecting a difference by chance?

      Your Answer:

      Correct Answer: 1 in 14 times

      Explanation:

      The probability of detecting a difference by chance is 1 in 20 times when the p-value is 0.05, which is the conventional cutoff. In this case, the answer is 1 in 14 times, which is equivalent to a p-value of 0.07.

      Understanding Hypothesis Testing in Statistics

      In statistics, it is not feasible to investigate hypotheses on entire populations. Therefore, researchers take samples and use them to make estimates about the population they are drawn from. However, this leads to uncertainty as there is no guarantee that the sample taken will be truly representative of the population, resulting in potential errors. Statistical hypothesis testing is the process used to determine if claims from samples to populations can be made and with what certainty.

      The null hypothesis (Ho) is the claim that there is no real difference between two groups, while the alternative hypothesis (H1 of Ha) suggests that any difference is due to some non-random chance. The alternative hypothesis can be one-tailed of two-tailed, depending on whether it seeks to establish a difference of a change in one direction.

      Two types of errors may occur when testing the null hypothesis: Type I and Type II errors. Type I error occurs when the null hypothesis is rejected when it is true, while Type II error occurs when the null hypothesis is accepted when it is false. The power of a study is the probability of correctly rejecting the null hypothesis when it is false, and it can be increased by increasing the sample size.

      P-values provide information on statistical significance and help researchers decide if study results have occurred due to chance. The p-value is the probability of obtaining a result that is as large of larger when in reality there is no difference between two groups. The cutoff for the p-value is called the significance level (alpha level), typically set at 0.05. If the p-value is less than the cutoff, the null hypothesis is rejected, and if it is greater or equal to the cut off, the null hypothesis is not rejected. However, the p-value does not indicate clinical significance, which may be too small to be meaningful.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 30 - Which statement accurately describes bar charts? ...

    Incorrect

    • Which statement accurately describes bar charts?

      Your Answer:

      Correct Answer: The height of the bar indicates the frequency

      Explanation:

      The frequency of each category of characteristic is displayed through the height of the bars in a bar chart. When dealing with discrete data, it is typically organized into distinct categories and presented in a bar chart. On the other hand, continuous data covers a range and the categories are not separate but rather blend into one another. This type of data is best represented through a histogram, which is similar to a bar chart but with bars that are connected.

      Differences between Bar Charts and Histograms

      Bar charts and histograms are both used to represent data, but they differ in their application and design. Bar charts are used to summarize nominal of ordinal data, while histograms are used for quantitative data. In a bar chart, the x-axis represents categories without a scale, and the y-axis represents frequencies. The columns are of equal width, and the height of the bar indicates the frequency. On the other hand, histograms have a scale on both axes, with the y-axis representing the relative frequency of frequency density. The width of the columns in a histogram can vary, and the area of the column indicates the true frequency. Overall, bar charts and histograms are useful tools for visualizing data, but their differences in design and application make them better suited for different types of data.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds

SESSION STATS - PERFORMANCE PER SPECIALTY

Research Methods, Statistics, Critical Review And Evidence-Based Practice (0/5) 0%
Passmed