00
Correct
00
Incorrect
00 : 00 : 00
Session Time
00 : 00
Average Question Time ( Mins)
  • Question 1 - What is the nature of the hypothesis that a researcher wants to test...

    Incorrect

    • What is the nature of the hypothesis that a researcher wants to test regarding the effect of a drug on a person's heart rate?

      Your Answer: Null hypothesis

      Correct Answer: One-tailed alternative hypothesis

      Explanation:

      A one-tailed hypothesis indicates a specific direction of association between groups. The researcher not only declares that there will be a distinction between the groups but also defines the direction in which the difference will occur.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      47.6
      Seconds
  • Question 2 - What standardized mortality ratio indicates a lower mortality rate in a sample group...

    Correct

    • What standardized mortality ratio indicates a lower mortality rate in a sample group compared to a reference group?

      Your Answer: 0.5

      Explanation:

      A negative SMR is not possible. An SMR less than 1.0 suggests that there were fewer deaths than expected in the study population, while an SMR of 1.0 indicates that the observed and expected deaths were equal. An SMR greater than 1.0 indicates that there were excess deaths in the study population.

      Calculation of Standardised Mortality Ratio (SMR)

      To calculate the SMR, age and sex-specific death rates in the standard population are obtained. An estimate for the number of people in each category for both the standard and study populations is needed. The number of expected deaths in each age-sex group of the study population is calculated by multiplying the age-sex-specific rates in the standard population by the number of people in each category of the study population. The sum of all age- and sex-specific expected deaths gives the expected number of deaths for the whole study population. The observed number of deaths is then divided by the expected number of deaths to obtain the SMR.

      The SMR can be standardised using the direct of indirect method. The direct method is used when the age-sex-specific rates for the study population and the age-sex-structure of the standard population are known. The indirect method is used when the age-specific rates for the study population are unknown of not available. This method uses the observed number of deaths in the study population and compares it to the number of deaths that would be expected if the age distribution was the same as that of the standard population.

      The SMR can be interpreted as follows: an SMR less than 1.0 indicates fewer than expected deaths in the study population, an SMR of 1.0 indicates the number of observed deaths equals the number of expected deaths in the study population, and an SMR greater than 1.0 indicates more than expected deaths in the study population (excess deaths). It is sometimes expressed after multiplying by 100.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      32
      Seconds
  • Question 3 - Which of the following can be used to represent the overall number of...

    Correct

    • Which of the following can be used to represent the overall number of individuals affected by a disease during a specific period?

      Your Answer: Period prevalence

      Explanation:

      Measures of Disease Frequency: Incidence and Prevalence

      Incidence and prevalence are two important measures of disease frequency. Incidence measures the speed at which new cases of a disease are emerging, while prevalence measures the burden of disease within a population. Cumulative incidence and incidence rate are two types of incidence measures, while point prevalence and period prevalence are two types of prevalence measures.

      Cumulative incidence is the average risk of getting a disease over a certain period of time, while incidence rate is a measure of the speed at which new cases are emerging. Prevalence is a proportion and is a measure of the burden of disease within a population. Point prevalence measures the number of cases in a defined population at a specific point in time, while period prevalence measures the number of identified cases during a specified period of time.

      It is important to note that prevalence is equal to incidence multiplied by the duration of the condition. In chronic diseases, the prevalence is much greater than the incidence. The incidence rate is stated in units of person-time, while cumulative incidence is always a proportion. When describing cumulative incidence, it is necessary to give the follow-up period over which the risk is estimated. In acute diseases, the prevalence and incidence may be similar, while for conditions such as the common cold, the incidence may be greater than the prevalence.

      Incidence is a useful measure to study disease etiology and risk factors, while prevalence is useful for health resource planning. Understanding these measures of disease frequency is important for public health professionals and researchers in order to effectively monitor and address the burden of disease within populations.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      50.8
      Seconds
  • Question 4 - What method did the researchers use to ensure the accuracy and credibility of...

    Incorrect

    • What method did the researchers use to ensure the accuracy and credibility of their findings in the qualitative study on antidepressants?

      Your Answer: Triangulation

      Correct Answer: Member checking

      Explanation:

      To ensure validity in qualitative studies, a technique called member checking of respondent validation is used. This involves interviewing a subset of the participants (typically around 11) to confirm that their perspectives align with the study’s findings.

      Qualitative research is a method of inquiry that seeks to understand the meaning and experience dimensions of human lives and social worlds. There are different approaches to qualitative research, such as ethnography, phenomenology, and grounded theory, each with its own purpose, role of the researcher, stages of research, and method of data analysis. The most common methods used in healthcare research are interviews and focus groups. Sampling techniques include convenience sampling, purposive sampling, quota sampling, snowball sampling, and case study sampling. Sample size can be determined by data saturation, which occurs when new categories, themes, of explanations stop emerging from the data. Validity can be assessed through triangulation, respondent validation, bracketing, and reflexivity. Analytical approaches include content analysis and constant comparison.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      22.5
      Seconds
  • Question 5 - What is the probability that a person who tests negative on the new...

    Correct

    • What is the probability that a person who tests negative on the new Mephedrone screening test does not actually use Mephedrone?

      Your Answer: 172/177

      Explanation:

      Negative predictive value = 172 / 177

      Clinical tests are used to determine the presence of absence of a disease of condition. To interpret test results, it is important to have a working knowledge of statistics used to describe them. Two by two tables are commonly used to calculate test statistics such as sensitivity and specificity. Sensitivity refers to the proportion of people with a condition that the test correctly identifies, while specificity refers to the proportion of people without a condition that the test correctly identifies. Accuracy tells us how closely a test measures to its true value, while predictive values help us understand the likelihood of having a disease based on a positive of negative test result. Likelihood ratios combine sensitivity and specificity into a single figure that can refine our estimation of the probability of a disease being present. Pre and post-test odds and probabilities can also be calculated to better understand the likelihood of having a disease before and after a test is carried out. Fagan’s nomogram is a useful tool for calculating post-test probabilities.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      17.2
      Seconds
  • Question 6 - What factors affect the statistical power of a study? ...

    Correct

    • What factors affect the statistical power of a study?

      Your Answer: Sample size

      Explanation:

      A study that has a greater sample size is considered to have higher power, meaning it is capable of detecting a significant difference of effect that is clinically relevant.

      The Importance of Power in Statistical Analysis

      Power is a crucial concept in statistical analysis as it helps researchers determine the number of participants needed in a study to detect a clinically significant difference of effect. It represents the probability of correctly rejecting the null hypothesis when it is false, which means avoiding a Type II error. Power values range from 0 to 1, with 0 indicating 0% and 1 indicating 100%. A power of 0.80 is generally considered the minimum acceptable level.

      Several factors influence the power of a study, including sample size, effect size, and significance level. Larger sample sizes lead to more precise parameter estimations and increase the study’s ability to detect a significant effect. Effect size, which is determined at the beginning of a study, refers to the size of the difference between two means that leads to rejecting the null hypothesis. Finally, the significance level, also known as the alpha level, represents the probability of a Type I error. By considering these factors, researchers can optimize the power of their studies and increase the likelihood of detecting meaningful effects.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      7.7
      Seconds
  • Question 7 - The research team is studying the effectiveness of a new treatment for a...

    Incorrect

    • The research team is studying the effectiveness of a new treatment for a certain medical condition. They have found that the brand name medication Y and its generic version Y1 have similar efficacy. They approach you for guidance on what type of analysis to conduct next. What would you suggest?

      Your Answer: Cost utility analysis

      Correct Answer: Cost minimisation analysis

      Explanation:

      Cost minimisation analysis is employed to compare net costs when the observed effects of health care interventions are similar. To conduct this analysis, it is necessary to have clinical evidence that demonstrates the differences in health effects between alternatives are negligible of insignificant. This approach is commonly used by institutions like the National Institute for Health and Care Excellence (NICE).

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      5548.4
      Seconds
  • Question 8 - What is the term used to describe the point at which a researcher...

    Incorrect

    • What is the term used to describe the point at which a researcher chooses to reject a null hypothesis?

      Your Answer: Theta level

      Correct Answer: Alpha level

      Explanation:

      If the p-value is lower than the predetermined alpha level of 0.05, the outcome is considered significant.

      Understanding Hypothesis Testing in Statistics

      In statistics, it is not feasible to investigate hypotheses on entire populations. Therefore, researchers take samples and use them to make estimates about the population they are drawn from. However, this leads to uncertainty as there is no guarantee that the sample taken will be truly representative of the population, resulting in potential errors. Statistical hypothesis testing is the process used to determine if claims from samples to populations can be made and with what certainty.

      The null hypothesis (Ho) is the claim that there is no real difference between two groups, while the alternative hypothesis (H1 of Ha) suggests that any difference is due to some non-random chance. The alternative hypothesis can be one-tailed of two-tailed, depending on whether it seeks to establish a difference of a change in one direction.

      Two types of errors may occur when testing the null hypothesis: Type I and Type II errors. Type I error occurs when the null hypothesis is rejected when it is true, while Type II error occurs when the null hypothesis is accepted when it is false. The power of a study is the probability of correctly rejecting the null hypothesis when it is false, and it can be increased by increasing the sample size.

      P-values provide information on statistical significance and help researchers decide if study results have occurred due to chance. The p-value is the probability of obtaining a result that is as large of larger when in reality there is no difference between two groups. The cutoff for the p-value is called the significance level (alpha level), typically set at 0.05. If the p-value is less than the cutoff, the null hypothesis is rejected, and if it is greater or equal to the cut off, the null hypothesis is not rejected. However, the p-value does not indicate clinical significance, which may be too small to be meaningful.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      50.2
      Seconds
  • Question 9 - What percentage of values fall within a range of 3 standard deviations above...

    Correct

    • What percentage of values fall within a range of 3 standard deviations above and below the mean?

      Your Answer: 99.70%

      Explanation:

      Measures of dispersion are used to indicate the variation of spread of a data set, often in conjunction with a measure of central tendency such as the mean of median. The range, which is the difference between the largest and smallest value, is the simplest measure of dispersion. The interquartile range, which is the difference between the 3rd and 1st quartiles, is another useful measure. Quartiles divide a data set into quarters, and the interquartile range can provide additional information about the spread of the data. However, to get a more representative idea of spread, measures such as the variance and standard deviation are needed. The variance gives an indication of how much the items in the data set vary from the mean, while the standard deviation reflects the distribution of individual scores around their mean. The standard deviation is expressed in the same units as the data set and can be used to indicate how confident we are that data points lie within a particular range. The standard error of the mean is an inferential statistic used to estimate the population mean and is a measure of the spread expected for the mean of the observations. Confidence intervals are often presented alongside sample results such as the mean value, indicating a range that is likely to contain the true value.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      15
      Seconds
  • Question 10 - A masters student had noticed that nearly all of her patients with arthritis...

    Incorrect

    • A masters student had noticed that nearly all of her patients with arthritis were over the age of 50. She was keen to investigate this further to see if there was an association.
      She selected 100 patients with arthritis and 100 controls. of the 100 patients with arthritis, 90 were over the age of 50. of the 100 controls, only 40 were over the age of 50.
      What is the odds ratio?

      Your Answer:

      Correct Answer: 3.77

      Explanation:

      The odds of being married are 3.77 times higher in individuals with panic disorder compared to controls.

      Measures of Effect in Clinical Studies

      When conducting clinical studies, we often want to know the effect of treatments of exposures on health outcomes. Measures of effect are used in randomized controlled trials (RCTs) and include the odds ratio (of), risk ratio (RR), risk difference (RD), and number needed to treat (NNT). Dichotomous (binary) outcome data are common in clinical trials, where the outcome for each participant is one of two possibilities, such as dead of alive, of clinical improvement of no improvement.

      To understand the difference between of and RR, it’s important to know the difference between risks and odds. Risk is a proportion that describes the probability of a health outcome occurring, while odds is a ratio that compares the probability of an event occurring to the probability of it not occurring. Absolute risk is the basic risk, while risk difference is the difference between the absolute risk of an event in the intervention group and the absolute risk in the control group. Relative risk is the ratio of risk in the intervention group to the risk in the control group.

      The number needed to treat (NNT) is the number of patients who need to be treated for one to benefit. Odds are calculated by dividing the number of times an event happens by the number of times it does not happen. The odds ratio is the odds of an outcome given a particular exposure versus the odds of an outcome in the absence of the exposure. It is commonly used in case-control studies and can also be used in cross-sectional and cohort study designs. An odds ratio of 1 indicates no difference in risk between the two groups, while an odds ratio >1 indicates an increased risk and an odds ratio <1 indicates a reduced risk.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 11 - The national health organization has a team of analysts to compare the effectiveness...

    Incorrect

    • The national health organization has a team of analysts to compare the effectiveness of two different cancer treatments in terms of cost and patient outcomes. They have gathered data on the number of years of life gained by each treatment and are seeking your recommendation on what type of analysis to conduct next. What analysis would you suggest they undertake?

      Your Answer:

      Correct Answer: Cost utility analysis

      Explanation:

      Cost utility analysis is a method used in health economics to determine the cost-effectiveness of a health intervention by comparing the cost of the intervention to the benefit it provides in terms of the number of years lived in full health. The cost is measured in monetary units, while the benefit is quantified using a measure that assigns values to different health states, including those that are less desirable than full health. In health technology assessments, this measure is typically expressed as quality-adjusted life years (QALYs).

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 12 - What is the appropriate denominator for calculating the incidence rate? ...

    Incorrect

    • What is the appropriate denominator for calculating the incidence rate?

      Your Answer:

      Correct Answer: The total person time at risk during a specified time period

      Explanation:

      Measures of Disease Frequency: Incidence and Prevalence

      Incidence and prevalence are two important measures of disease frequency. Incidence measures the speed at which new cases of a disease are emerging, while prevalence measures the burden of disease within a population. Cumulative incidence and incidence rate are two types of incidence measures, while point prevalence and period prevalence are two types of prevalence measures.

      Cumulative incidence is the average risk of getting a disease over a certain period of time, while incidence rate is a measure of the speed at which new cases are emerging. Prevalence is a proportion and is a measure of the burden of disease within a population. Point prevalence measures the number of cases in a defined population at a specific point in time, while period prevalence measures the number of identified cases during a specified period of time.

      It is important to note that prevalence is equal to incidence multiplied by the duration of the condition. In chronic diseases, the prevalence is much greater than the incidence. The incidence rate is stated in units of person-time, while cumulative incidence is always a proportion. When describing cumulative incidence, it is necessary to give the follow-up period over which the risk is estimated. In acute diseases, the prevalence and incidence may be similar, while for conditions such as the common cold, the incidence may be greater than the prevalence.

      Incidence is a useful measure to study disease etiology and risk factors, while prevalence is useful for health resource planning. Understanding these measures of disease frequency is important for public health professionals and researchers in order to effectively monitor and address the burden of disease within populations.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 13 - Researchers have conducted a study comparing a new blood pressure medication with a...

    Incorrect

    • Researchers have conducted a study comparing a new blood pressure medication with a standard blood pressure medication. 200 patients are divided equally between the two groups. Over the course of one year, 20 patients in the treatment group experienced a significant reduction in blood pressure, compared to 35 patients in the control group.

      What is the number needed to treat (NNT)?

      Your Answer:

      Correct Answer: 7

      Explanation:

      The Relative Risk Reduction (RRR) is calculated by subtracting the experimental event rate (EER) from the control event rate (CER), dividing the result by the CER, and then multiplying by 100 to get a percentage. In this case, the RRR is (35-20)÷35 = 0.4285 of 42.85%.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 14 - What qualitative research approach aims to understand individuals' inner experiences and perspectives? ...

    Incorrect

    • What qualitative research approach aims to understand individuals' inner experiences and perspectives?

      Your Answer:

      Correct Answer: Phenomenology

      Explanation:

      Qualitative research is a method of inquiry that seeks to understand the meaning and experience dimensions of human lives and social worlds. There are different approaches to qualitative research, such as ethnography, phenomenology, and grounded theory, each with its own purpose, role of the researcher, stages of research, and method of data analysis. The most common methods used in healthcare research are interviews and focus groups. Sampling techniques include convenience sampling, purposive sampling, quota sampling, snowball sampling, and case study sampling. Sample size can be determined by data saturation, which occurs when new categories, themes, of explanations stop emerging from the data. Validity can be assessed through triangulation, respondent validation, bracketing, and reflexivity. Analytical approaches include content analysis and constant comparison.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 15 - What is another term used to refer to a type II error in...

    Incorrect

    • What is another term used to refer to a type II error in hypothesis testing?

      Your Answer:

      Correct Answer: False negative

      Explanation:

      Hypothesis testing involves the possibility of two types of errors: type I and type II errors. A type I error occurs when the null hypothesis is wrongly rejected of the alternative hypothesis is wrongly accepted. This error is also referred to as an alpha error, error of the first kind, of a false positive. On the other hand, a type II error occurs when the null hypothesis is wrongly accepted. This error is also known as the beta error, error of the second kind, of the false negative.

      Understanding Hypothesis Testing in Statistics

      In statistics, it is not feasible to investigate hypotheses on entire populations. Therefore, researchers take samples and use them to make estimates about the population they are drawn from. However, this leads to uncertainty as there is no guarantee that the sample taken will be truly representative of the population, resulting in potential errors. Statistical hypothesis testing is the process used to determine if claims from samples to populations can be made and with what certainty.

      The null hypothesis (Ho) is the claim that there is no real difference between two groups, while the alternative hypothesis (H1 of Ha) suggests that any difference is due to some non-random chance. The alternative hypothesis can be one-tailed of two-tailed, depending on whether it seeks to establish a difference of a change in one direction.

      Two types of errors may occur when testing the null hypothesis: Type I and Type II errors. Type I error occurs when the null hypothesis is rejected when it is true, while Type II error occurs when the null hypothesis is accepted when it is false. The power of a study is the probability of correctly rejecting the null hypothesis when it is false, and it can be increased by increasing the sample size.

      P-values provide information on statistical significance and help researchers decide if study results have occurred due to chance. The p-value is the probability of obtaining a result that is as large of larger when in reality there is no difference between two groups. The cutoff for the p-value is called the significance level (alpha level), typically set at 0.05. If the p-value is less than the cutoff, the null hypothesis is rejected, and if it is greater or equal to the cut off, the null hypothesis is not rejected. However, the p-value does not indicate clinical significance, which may be too small to be meaningful.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 16 - What statement accurately describes measures of dispersion? ...

    Incorrect

    • What statement accurately describes measures of dispersion?

      Your Answer:

      Correct Answer: The standard error indicates how close the statistical mean is to the population mean

      Explanation:

      Measures of dispersion are used to indicate the variation of spread of a data set, often in conjunction with a measure of central tendency such as the mean of median. The range, which is the difference between the largest and smallest value, is the simplest measure of dispersion. The interquartile range, which is the difference between the 3rd and 1st quartiles, is another useful measure. Quartiles divide a data set into quarters, and the interquartile range can provide additional information about the spread of the data. However, to get a more representative idea of spread, measures such as the variance and standard deviation are needed. The variance gives an indication of how much the items in the data set vary from the mean, while the standard deviation reflects the distribution of individual scores around their mean. The standard deviation is expressed in the same units as the data set and can be used to indicate how confident we are that data points lie within a particular range. The standard error of the mean is an inferential statistic used to estimate the population mean and is a measure of the spread expected for the mean of the observations. Confidence intervals are often presented alongside sample results such as the mean value, indicating a range that is likely to contain the true value.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 17 - What is a common tool used to help determine the appropriate sample size...

    Incorrect

    • What is a common tool used to help determine the appropriate sample size for qualitative research?

      Your Answer:

      Correct Answer: Saturation

      Explanation:

      Qualitative research is a method of inquiry that seeks to understand the meaning and experience dimensions of human lives and social worlds. There are different approaches to qualitative research, such as ethnography, phenomenology, and grounded theory, each with its own purpose, role of the researcher, stages of research, and method of data analysis. The most common methods used in healthcare research are interviews and focus groups. Sampling techniques include convenience sampling, purposive sampling, quota sampling, snowball sampling, and case study sampling. Sample size can be determined by data saturation, which occurs when new categories, themes, of explanations stop emerging from the data. Validity can be assessed through triangulation, respondent validation, bracketing, and reflexivity. Analytical approaches include content analysis and constant comparison.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 18 - What is a true statement about measures of effect? ...

    Incorrect

    • What is a true statement about measures of effect?

      Your Answer:

      Correct Answer: Relative risk can be used to measure effect in randomised control trials

      Explanation:

      The use of relative risk is applicable in cohort, cross-sectional, and randomized control trials, but not in case-control studies. In situations where there are no events in the control group, neither the risk ratio nor the odds ratio can be computed. It is important to note that the odds ratio tends to overestimate effects and is always more extreme than the relative risk, moving away from the null value of 1.

      Measures of Effect in Clinical Studies

      When conducting clinical studies, we often want to know the effect of treatments of exposures on health outcomes. Measures of effect are used in randomized controlled trials (RCTs) and include the odds ratio (of), risk ratio (RR), risk difference (RD), and number needed to treat (NNT). Dichotomous (binary) outcome data are common in clinical trials, where the outcome for each participant is one of two possibilities, such as dead of alive, of clinical improvement of no improvement.

      To understand the difference between of and RR, it’s important to know the difference between risks and odds. Risk is a proportion that describes the probability of a health outcome occurring, while odds is a ratio that compares the probability of an event occurring to the probability of it not occurring. Absolute risk is the basic risk, while risk difference is the difference between the absolute risk of an event in the intervention group and the absolute risk in the control group. Relative risk is the ratio of risk in the intervention group to the risk in the control group.

      The number needed to treat (NNT) is the number of patients who need to be treated for one to benefit. Odds are calculated by dividing the number of times an event happens by the number of times it does not happen. The odds ratio is the odds of an outcome given a particular exposure versus the odds of an outcome in the absence of the exposure. It is commonly used in case-control studies and can also be used in cross-sectional and cohort study designs. An odds ratio of 1 indicates no difference in risk between the two groups, while an odds ratio >1 indicates an increased risk and an odds ratio <1 indicates a reduced risk.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 19 - What is another name for admission rate bias? ...

    Incorrect

    • What is another name for admission rate bias?

      Your Answer:

      Correct Answer: Berkson's bias

      Explanation:

      Types of Bias in Statistics

      Bias is a systematic error that can lead to incorrect conclusions. Confounding factors are variables that are associated with both the outcome and the exposure but have no causative role. Confounding can be addressed in the design and analysis stage of a study. The main method of controlling confounding in the analysis phase is stratification analysis. The main methods used in the design stage are matching, randomization, and restriction of participants.

      There are two main types of bias: selection bias and information bias. Selection bias occurs when the selected sample is not a representative sample of the reference population. Disease spectrum bias, self-selection bias, participation bias, incidence-prevalence bias, exclusion bias, publication of dissemination bias, citation bias, and Berkson’s bias are all subtypes of selection bias. Information bias occurs when gathered information about exposure, outcome, of both is not correct and there was an error in measurement. Detection bias, recall bias, lead time bias, interviewer/observer bias, verification and work-up bias, Hawthorne effect, and ecological fallacy are all subtypes of information bias.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 20 - Which of the following resources has been filtered? ...

    Incorrect

    • Which of the following resources has been filtered?

      Your Answer:

      Correct Answer: DARE

      Explanation:

      The main focus of the Database of Abstracts of Reviews of Effect (DARE) is on systematic reviews that assess the impact of healthcare interventions and the management and provision of healthcare services. In order to be considered for inclusion, reviews must satisfy several requirements.

      Evidence-based medicine involves four basic steps: developing a focused clinical question, searching for the best evidence, critically appraising the evidence, and applying the evidence and evaluating the outcome. When developing a question, it is important to understand the difference between background and foreground questions. Background questions are general questions about conditions, illnesses, syndromes, and pathophysiology, while foreground questions are more often about issues of care. The PICO system is often used to define the components of a foreground question: patient group of interest, intervention of interest, comparison, and primary outcome.

      When searching for evidence, it is important to have a basic understanding of the types of evidence and sources of information. Scientific literature is divided into two basic categories: primary (empirical research) and secondary (interpretation and analysis of primary sources). Unfiltered sources are large databases of articles that have not been pre-screened for quality, while filtered resources summarize and appraise evidence from several studies.

      There are several databases and search engines that can be used to search for evidence, including Medline and PubMed, Embase, the Cochrane Library, PsycINFO, CINAHL, and OpenGrey. Boolean logic can be used to combine search terms in PubMed, and phrase searching and truncation can also be used. Medical Subject Headings (MeSH) are used by indexers to describe articles for MEDLINE records, and the MeSH Database is like a thesaurus that enables exploration of this vocabulary.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 21 - In a cohort study investigating the association between smoking and Alzheimer's dementia, what...

    Incorrect

    • In a cohort study investigating the association between smoking and Alzheimer's dementia, what is the typical variable used to measure the outcome?

      Your Answer:

      Correct Answer: Relative risk

      Explanation:

      The odds ratio is used in case-control studies to measure the association between exposure and outcome, while the relative risk is used in cohort studies to measure the risk of developing an outcome in the exposed group compared to the unexposed group. To convert the odds ratio to a relative risk, one can use the formula: relative risk = odds ratio / (1 – incidence in the unexposed group x odds ratio).

      Types of Primary Research Studies and Their Advantages and Disadvantages

      Primary research studies can be categorized into six types based on the research question they aim to address. The best type of study for each question type is listed in the table below. There are two main types of study design: experimental and observational. Experimental studies involve an intervention, while observational studies do not. The advantages and disadvantages of each study type are summarized in the table below.

      Type of Question Best Type of Study

      Therapy Randomized controlled trial (RCT), cohort, case control, case series
      Diagnosis Cohort studies with comparison to gold standard test
      Prognosis Cohort studies, case control, case series
      Etiology/Harm RCT, cohort studies, case control, case series
      Prevention RCT, cohort studies, case control, case series
      Cost Economic analysis

      Study Type Advantages Disadvantages

      Randomized Controlled Trial – Unbiased distribution of confounders – Blinding more likely – Randomization facilitates statistical analysis – Expensive – Time-consuming – Volunteer bias – Ethically problematic at times
      Cohort Study – Ethically safe – Subjects can be matched – Can establish timing and directionality of events – Eligibility criteria and outcome assessments can be standardized – Administratively easier and cheaper than RCT – Controls may be difficult to identify – Exposure may be linked to a hidden confounder – Blinding is difficult – Randomization not present – For rare disease, large sample sizes of long follow-up necessary
      Case-Control Study – Quick and cheap – Only feasible method for very rare disorders of those with long lag between exposure and outcome – Fewer subjects needed than cross-sectional studies – Reliance on recall of records to determine exposure status – Confounders – Selection of control groups is difficult – Potential bias: recall, selection
      Cross-Sectional Survey – Cheap and simple – Ethically safe – Establishes association at most, not causality – Recall bias susceptibility – Confounders may be unequally distributed – Neyman bias – Group sizes may be unequal
      Ecological Study – Cheap and simple – Ethically safe – Ecological fallacy (when relationships which exist for groups are assumed to also be true for individuals)

      In conclusion, the choice of study type depends on the research question being addressed. Each study type has its own advantages and disadvantages, and researchers should carefully consider these when designing their studies.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 22 - What does the term external validity in a study refer to? ...

    Incorrect

    • What does the term external validity in a study refer to?

      Your Answer:

      Correct Answer: The degree to which the conclusions in a study would hold for other persons in other places and at other times

      Explanation:

      Validity in statistics refers to how accurately something measures what it claims to measure. There are two main types of validity: internal and external. Internal validity refers to the confidence we have in the cause and effect relationship in a study, while external validity refers to the degree to which the conclusions of a study can be applied to other people, places, and times. There are various threats to both internal and external validity, such as sampling, measurement instrument obtrusiveness, and reactive effects of setting. Additionally, there are several subtypes of validity, including face validity, content validity, criterion validity, and construct validity. Each subtype has its own specific focus and methods for testing validity.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 23 - What is the name of the database that focuses on literature created by...

    Incorrect

    • What is the name of the database that focuses on literature created by non-traditional commercial of academic publishing and distribution channels?

      Your Answer:

      Correct Answer: OpenGrey

      Explanation:

      SIGLE is a database that specializes in collecting and indexing grey literature.

      Evidence-based medicine involves four basic steps: developing a focused clinical question, searching for the best evidence, critically appraising the evidence, and applying the evidence and evaluating the outcome. When developing a question, it is important to understand the difference between background and foreground questions. Background questions are general questions about conditions, illnesses, syndromes, and pathophysiology, while foreground questions are more often about issues of care. The PICO system is often used to define the components of a foreground question: patient group of interest, intervention of interest, comparison, and primary outcome.

      When searching for evidence, it is important to have a basic understanding of the types of evidence and sources of information. Scientific literature is divided into two basic categories: primary (empirical research) and secondary (interpretation and analysis of primary sources). Unfiltered sources are large databases of articles that have not been pre-screened for quality, while filtered resources summarize and appraise evidence from several studies.

      There are several databases and search engines that can be used to search for evidence, including Medline and PubMed, Embase, the Cochrane Library, PsycINFO, CINAHL, and OpenGrey. Boolean logic can be used to combine search terms in PubMed, and phrase searching and truncation can also be used. Medical Subject Headings (MeSH) are used by indexers to describe articles for MEDLINE records, and the MeSH Database is like a thesaurus that enables exploration of this vocabulary.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 24 - Which odds ratio suggests that there is no significant variation in the odds...

    Incorrect

    • Which odds ratio suggests that there is no significant variation in the odds between two groups?

      Your Answer:

      Correct Answer: 1

      Explanation:

      Measures of Effect in Clinical Studies

      When conducting clinical studies, we often want to know the effect of treatments of exposures on health outcomes. Measures of effect are used in randomized controlled trials (RCTs) and include the odds ratio (of), risk ratio (RR), risk difference (RD), and number needed to treat (NNT). Dichotomous (binary) outcome data are common in clinical trials, where the outcome for each participant is one of two possibilities, such as dead of alive, of clinical improvement of no improvement.

      To understand the difference between of and RR, it’s important to know the difference between risks and odds. Risk is a proportion that describes the probability of a health outcome occurring, while odds is a ratio that compares the probability of an event occurring to the probability of it not occurring. Absolute risk is the basic risk, while risk difference is the difference between the absolute risk of an event in the intervention group and the absolute risk in the control group. Relative risk is the ratio of risk in the intervention group to the risk in the control group.

      The number needed to treat (NNT) is the number of patients who need to be treated for one to benefit. Odds are calculated by dividing the number of times an event happens by the number of times it does not happen. The odds ratio is the odds of an outcome given a particular exposure versus the odds of an outcome in the absence of the exposure. It is commonly used in case-control studies and can also be used in cross-sectional and cohort study designs. An odds ratio of 1 indicates no difference in risk between the two groups, while an odds ratio >1 indicates an increased risk and an odds ratio <1 indicates a reduced risk.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 25 - What is a true statement about standardised mortality ratios? ...

    Incorrect

    • What is a true statement about standardised mortality ratios?

      Your Answer:

      Correct Answer: Direct standardisation requires that we know the age-specific rates of mortality in all the populations under study

      Explanation:

      Calculation of Standardised Mortality Ratio (SMR)

      To calculate the SMR, age and sex-specific death rates in the standard population are obtained. An estimate for the number of people in each category for both the standard and study populations is needed. The number of expected deaths in each age-sex group of the study population is calculated by multiplying the age-sex-specific rates in the standard population by the number of people in each category of the study population. The sum of all age- and sex-specific expected deaths gives the expected number of deaths for the whole study population. The observed number of deaths is then divided by the expected number of deaths to obtain the SMR.

      The SMR can be standardised using the direct of indirect method. The direct method is used when the age-sex-specific rates for the study population and the age-sex-structure of the standard population are known. The indirect method is used when the age-specific rates for the study population are unknown of not available. This method uses the observed number of deaths in the study population and compares it to the number of deaths that would be expected if the age distribution was the same as that of the standard population.

      The SMR can be interpreted as follows: an SMR less than 1.0 indicates fewer than expected deaths in the study population, an SMR of 1.0 indicates the number of observed deaths equals the number of expected deaths in the study population, and an SMR greater than 1.0 indicates more than expected deaths in the study population (excess deaths). It is sometimes expressed after multiplying by 100.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 26 - Which statement accurately describes the measurement of serum potassium in 1,000 patients with...

    Incorrect

    • Which statement accurately describes the measurement of serum potassium in 1,000 patients with anorexia nervosa, where the mean potassium is 4.6 mmol/l and the standard deviation is 0.3 mmol/l?

      Your Answer:

      Correct Answer: 68.3% of values lie between 4.3 and 4.9 mmol/l

      Explanation:

      Standard Deviation and Standard Error of the Mean

      Standard deviation (SD) and standard error of the mean (SEM) are two important statistical measures used to describe data. SD is a measure of how much the data varies, while SEM is a measure of how precisely we know the true mean of the population. The normal distribution, also known as the Gaussian distribution, is a symmetrical bell-shaped curve that describes the spread of many biological and clinical measurements.

      68.3% of the data lies within 1 SD of the mean, 95.4% of the data lies within 2 SD of the mean, and 99.7% of the data lies within 3 SD of the mean. The SD is calculated by taking the square root of the variance and is expressed in the same units as the data set. A low SD indicates that data points tend to be very close to the mean.

      On the other hand, SEM is an inferential statistic that quantifies the precision of the mean. It is expressed in the same units as the data and is calculated by dividing the SD of the sample mean by the square root of the sample size. The SEM gets smaller as the sample size increases, and it takes into account both the value of the SD and the sample size.

      Both SD and SEM are important measures in statistical analysis, and they are used to calculate confidence intervals and test hypotheses. While SD quantifies scatter, SEM quantifies precision, and both are essential in understanding and interpreting data.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 27 - What is the negative predictive value of the blood test for bowel cancer,...

    Incorrect

    • What is the negative predictive value of the blood test for bowel cancer, given a sensitivity of 60% and a specificity of 80% and a negative test result for a patient?

      Your Answer:

      Correct Answer: 0.5

      Explanation:

      Clinical tests are used to determine the presence of absence of a disease of condition. To interpret test results, it is important to have a working knowledge of statistics used to describe them. Two by two tables are commonly used to calculate test statistics such as sensitivity and specificity. Sensitivity refers to the proportion of people with a condition that the test correctly identifies, while specificity refers to the proportion of people without a condition that the test correctly identifies. Accuracy tells us how closely a test measures to its true value, while predictive values help us understand the likelihood of having a disease based on a positive of negative test result. Likelihood ratios combine sensitivity and specificity into a single figure that can refine our estimation of the probability of a disease being present. Pre and post-test odds and probabilities can also be calculated to better understand the likelihood of having a disease before and after a test is carried out. Fagan’s nomogram is a useful tool for calculating post-test probabilities.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 28 - A team of scientists aims to conduct a systematic review on the effectiveness...

    Incorrect

    • A team of scientists aims to conduct a systematic review on the effectiveness of a new medication for elderly patients with dementia. They decide to search for studies published in languages other than English, as they know that positive results are more likely to be published in English-language journals, while negative results are more likely to be published in non-English language journals. What type of bias are they trying to prevent?

      Your Answer:

      Correct Answer: Tower of Babel bias

      Explanation:

      When conducting a systematic review, restricting the selection of studies to those published only in English may introduce a bias known as the Tower of Babel effect. This occurs because studies conducted in non-English speaking countries that report positive results are more likely to be published in English language journals, while those with negative results are more likely to be published in non-English language journals.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 29 - Which data type does age in years belong to? ...

    Incorrect

    • Which data type does age in years belong to?

      Your Answer:

      Correct Answer: Ratio

      Explanation:

      Age is a type of measurement that follows a ratio scale, which means that the values can be compared as multiples of each other. For instance, if someone is 20 years old, they are twice as old as someone who is 10 years old.

      Scales of Measurement in Statistics

      In the 1940s, Stanley Smith Stevens introduced four scales of measurement to categorize data variables. Knowing the scale of measurement for a variable is crucial in selecting the appropriate statistical analysis. The four scales of measurement are ratio, interval, ordinal, and nominal.

      Ratio scales are similar to interval scales, but they have true zero points. Examples of ratio scales include weight, time, and length. Interval scales measure the difference between two values, and one unit on the scale represents the same magnitude on the trait of characteristic being measured across the whole range of the scale. The Fahrenheit scale for temperature is an example of an interval scale.

      Ordinal scales categorize observed values into set categories that can be ordered, but the intervals between each value are uncertain. Examples of ordinal scales include social class, education level, and income level. Nominal scales categorize observed values into set categories that have no particular order of hierarchy. Examples of nominal scales include genotype, blood type, and political party.

      Data can also be categorized as quantitative of qualitative. Quantitative variables take on numeric values and can be further classified into discrete and continuous types. Qualitative variables do not take on numerical values and are usually names. Some qualitative variables have an inherent order in their categories and are described as ordinal. Qualitative variables are also called categorical of nominal variables. When a qualitative variable has only two categories, it is called a binary variable.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 30 - Which of the following statements accurately describes the concept of study power? ...

    Incorrect

    • Which of the following statements accurately describes the concept of study power?

      Your Answer:

      Correct Answer: Is the probability of rejecting the null hypothesis when it is false

      Explanation:

      The Importance of Power in Statistical Analysis

      Power is a crucial concept in statistical analysis as it helps researchers determine the number of participants needed in a study to detect a clinically significant difference of effect. It represents the probability of correctly rejecting the null hypothesis when it is false, which means avoiding a Type II error. Power values range from 0 to 1, with 0 indicating 0% and 1 indicating 100%. A power of 0.80 is generally considered the minimum acceptable level.

      Several factors influence the power of a study, including sample size, effect size, and significance level. Larger sample sizes lead to more precise parameter estimations and increase the study’s ability to detect a significant effect. Effect size, which is determined at the beginning of a study, refers to the size of the difference between two means that leads to rejecting the null hypothesis. Finally, the significance level, also known as the alpha level, represents the probability of a Type I error. By considering these factors, researchers can optimize the power of their studies and increase the likelihood of detecting meaningful effects.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 31 - The QALY is utilized in which of the following approaches for economic assessment?...

    Incorrect

    • The QALY is utilized in which of the following approaches for economic assessment?

      Your Answer:

      Correct Answer: Cost-utility analysis

      Explanation:

      Methods of Economic Evaluation

      There are four main methods of economic evaluation: cost-effectiveness analysis (CEA), cost-benefit analysis (CBA), cost-utility analysis (CUA), and cost-minimisation analysis (CMA). While all four methods capture costs, they differ in how they assess health effects.

      Cost-effectiveness analysis (CEA) compares interventions by relating costs to a single clinical measure of effectiveness, such as symptom reduction of improvement in activities of daily living. The cost-effectiveness ratio is calculated as total cost divided by units of effectiveness. CEA is typically used when CBA cannot be performed due to the inability to monetise benefits.

      Cost-benefit analysis (CBA) measures all costs and benefits of an intervention in monetary terms to establish which alternative has the greatest net benefit. CBA requires that all consequences of an intervention, such as life-years saved, treatment side-effects, symptom relief, disability, pain, and discomfort, are allocated a monetary value. CBA is rarely used in mental health service evaluation due to the difficulty in converting benefits from mental health programmes into monetary values.

      Cost-utility analysis (CUA) is a special form of CEA in which health benefits/outcomes are measured in broader, more generic ways, enabling comparisons between treatments for different diseases and conditions. Multidimensional health outcomes are measured by a single preference- of utility-based index such as the Quality-Adjusted-Life-Years (QALY). QALYs are a composite measure of gains in life expectancy and health-related quality of life. CUA allows for comparisons across treatments for different conditions.

      Cost-minimisation analysis (CMA) is an economic evaluation in which the consequences of competing interventions are the same, and only inputs, i.e. costs, are taken into consideration. The aim is to decide the least costly way of achieving the same outcome.

      Costs in Economic Evaluation Studies

      There are three main types of costs in economic evaluation studies: direct, indirect, and intangible. Direct costs are associated directly with the healthcare intervention, such as staff time, medical supplies, cost of travel for the patient, childcare costs for the patient, and costs falling on other social sectors such as domestic help from social services. Indirect costs are incurred by the reduced productivity of the patient, such as time off work, reduced work productivity, and time spent caring for the patient by relatives. Intangible costs are difficult to measure, such as pain of suffering on the part of the patient.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 32 - A study was conducted to investigate the correlation between body mass index (BMI)...

    Incorrect

    • A study was conducted to investigate the correlation between body mass index (BMI) and mortality in patients with schizophrenia. The study involved a cohort of 1000 patients with schizophrenia who were evaluated by measuring their weight and height, and calculating their BMI. The participants were then monitored for up to 15 years after the study commenced. The BMI levels were classified into three categories (high, average, low). The findings revealed that, after adjusting for age, gender, treatment method, and comorbidities, a high BMI at the beginning of the study was linked to a twofold increase in mortality.
      How is this study best described?

      Your Answer:

      Correct Answer:

      Explanation:

      The study is a prospective cohort study that observes the effect of BMI as an exposure on the group over time, without manipulating any risk factors of interventions.

      Types of Primary Research Studies and Their Advantages and Disadvantages

      Primary research studies can be categorized into six types based on the research question they aim to address. The best type of study for each question type is listed in the table below. There are two main types of study design: experimental and observational. Experimental studies involve an intervention, while observational studies do not. The advantages and disadvantages of each study type are summarized in the table below.

      Type of Question Best Type of Study

      Therapy Randomized controlled trial (RCT), cohort, case control, case series
      Diagnosis Cohort studies with comparison to gold standard test
      Prognosis Cohort studies, case control, case series
      Etiology/Harm RCT, cohort studies, case control, case series
      Prevention RCT, cohort studies, case control, case series
      Cost Economic analysis

      Study Type Advantages Disadvantages

      Randomized Controlled Trial – Unbiased distribution of confounders – Blinding more likely – Randomization facilitates statistical analysis – Expensive – Time-consuming – Volunteer bias – Ethically problematic at times
      Cohort Study – Ethically safe – Subjects can be matched – Can establish timing and directionality of events – Eligibility criteria and outcome assessments can be standardized – Administratively easier and cheaper than RCT – Controls may be difficult to identify – Exposure may be linked to a hidden confounder – Blinding is difficult – Randomization not present – For rare disease, large sample sizes of long follow-up necessary
      Case-Control Study – Quick and cheap – Only feasible method for very rare disorders of those with long lag between exposure and outcome – Fewer subjects needed than cross-sectional studies – Reliance on recall of records to determine exposure status – Confounders – Selection of control groups is difficult – Potential bias: recall, selection
      Cross-Sectional Survey – Cheap and simple – Ethically safe – Establishes association at most, not causality – Recall bias susceptibility – Confounders may be unequally distributed – Neyman bias – Group sizes may be unequal
      Ecological Study – Cheap and simple – Ethically safe – Ecological fallacy (when relationships which exist for groups are assumed to also be true for individuals)

      In conclusion, the choice of study type depends on the research question being addressed. Each study type has its own advantages and disadvantages, and researchers should carefully consider these when designing their studies.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 33 - What is a characteristic of skewed data? ...

    Incorrect

    • What is a characteristic of skewed data?

      Your Answer:

      Correct Answer: For positively skewed data the mean is greater than the mode

      Explanation:

      Skewed Data: Understanding the Relationship between Mean, Median, and Mode

      When analyzing a data set, it is important to consider the shape of the distribution. In a normally distributed data set, the curve is symmetrical and bell-shaped, with the median, mode, and mean all equal. However, in skewed data sets, the distribution is asymmetrical, with the bulk of the data concentrated on one side of the figure.

      In a negatively skewed distribution, the left tail is longer, and the bulk of the data is concentrated to the right of the figure. In contrast, a positively skewed distribution has a longer right tail, with the bulk of the data concentrated to the left of the figure. In both cases, the median is positioned between the mode and the mean, as it represents the halfway point of the distribution.

      However, the mean is affected by extreme values of outliers, causing it to move away from the median in the direction of the tail. In positively skewed data, the mean is greater than the median, which is greater than the mode. In negatively skewed data, the mode is greater than the median, which is greater than the mean.

      Understanding the relationship between mean, median, and mode in skewed data sets is crucial for accurate data analysis and interpretation. By recognizing the shape of the distribution, researchers can make informed decisions about which measures of central tendency to use and how to interpret their results.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 34 - What is the primary purpose of funnel plots? ...

    Incorrect

    • What is the primary purpose of funnel plots?

      Your Answer:

      Correct Answer: Demonstrate the existence of publication bias in meta-analyses

      Explanation:

      Stats Publication Bias

      Publication bias refers to the tendency for studies with positive findings to be published more than studies with negative findings, leading to incomplete data sets in meta-analyses and erroneous conclusions. Graphical methods such as funnel plots, Galbraith plots, ordered forest plots, and normal quantile plots can be used to detect publication bias. Funnel plots are the most commonly used and offer an easy visual way to ensure that published literature is evenly weighted. The x-axis represents the effect size, and the y-axis represents the study size. A symmetrical, inverted funnel shape indicates that publication bias is unlikely, while an asymmetrical funnel indicates a relationship between treatment effect and study size, indicating either publication bias of small study effects.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 35 - In what way can the study on depression be deemed as having limited...

    Incorrect

    • In what way can the study on depression be deemed as having limited applicability to the average patient population?

      Your Answer:

      Correct Answer: External validity

      Explanation:

      When a study has good external validity, its findings can be applied to other populations with confidence.

      Validity in statistics refers to how accurately something measures what it claims to measure. There are two main types of validity: internal and external. Internal validity refers to the confidence we have in the cause and effect relationship in a study, while external validity refers to the degree to which the conclusions of a study can be applied to other people, places, and times. There are various threats to both internal and external validity, such as sampling, measurement instrument obtrusiveness, and reactive effects of setting. Additionally, there are several subtypes of validity, including face validity, content validity, criterion validity, and construct validity. Each subtype has its own specific focus and methods for testing validity.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 36 - Which odds ratio, along with its confidence interval, indicates a statistically significant reduction...

    Incorrect

    • Which odds ratio, along with its confidence interval, indicates a statistically significant reduction in the odds?

      Your Answer:

      Correct Answer: 0.7 (0.1 - 0.8)

      Explanation:

      Measures of Effect in Clinical Studies

      When conducting clinical studies, we often want to know the effect of treatments of exposures on health outcomes. Measures of effect are used in randomized controlled trials (RCTs) and include the odds ratio (of), risk ratio (RR), risk difference (RD), and number needed to treat (NNT). Dichotomous (binary) outcome data are common in clinical trials, where the outcome for each participant is one of two possibilities, such as dead of alive, of clinical improvement of no improvement.

      To understand the difference between of and RR, it’s important to know the difference between risks and odds. Risk is a proportion that describes the probability of a health outcome occurring, while odds is a ratio that compares the probability of an event occurring to the probability of it not occurring. Absolute risk is the basic risk, while risk difference is the difference between the absolute risk of an event in the intervention group and the absolute risk in the control group. Relative risk is the ratio of risk in the intervention group to the risk in the control group.

      The number needed to treat (NNT) is the number of patients who need to be treated for one to benefit. Odds are calculated by dividing the number of times an event happens by the number of times it does not happen. The odds ratio is the odds of an outcome given a particular exposure versus the odds of an outcome in the absence of the exposure. It is commonly used in case-control studies and can also be used in cross-sectional and cohort study designs. An odds ratio of 1 indicates no difference in risk between the two groups, while an odds ratio >1 indicates an increased risk and an odds ratio <1 indicates a reduced risk.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 37 - Calculate the median value from the following values:
    1, 3, 3, 3, 4, 5,...

    Incorrect

    • Calculate the median value from the following values:
      1, 3, 3, 3, 4, 5, 5, 6, 6, 6, 6

      Your Answer:

      Correct Answer: 5

      Explanation:

      Measures of Central Tendency

      Measures of central tendency are used in descriptive statistics to summarize the middle of typical value of a data set. There are three common measures of central tendency: the mean, median, and mode.

      The median is the middle value in a data set that has been arranged in numerical order. It is not affected by outliers and is used for ordinal data. The mode is the most frequent value in a data set and is used for categorical data. The mean is calculated by adding all the values in a data set and dividing by the number of values. It is sensitive to outliers and is used for interval and ratio data.

      The appropriate measure of central tendency depends on the measurement scale of the data. For nominal and categorical data, the mode is used. For ordinal data, the median of mode is used. For interval data with a normal distribution, the mean is preferable, but the median of mode can also be used. For interval data with skewed distribution, the median is used. For ratio data, the mean is preferable, but the median of mode can also be used for skewed data.

      In addition to measures of central tendency, the range is also used to describe the spread of a data set. It is calculated by subtracting the smallest value from the largest value.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 38 - Which of the following options is not a possible value for Pearson's correlation...

    Incorrect

    • Which of the following options is not a possible value for Pearson's correlation coefficient?

      Your Answer:

      Correct Answer: 1.5

      Explanation:

      Stats: Correlation and Regression

      Correlation and regression are related but not interchangeable terms. Correlation is used to test for association between variables, while regression is used to predict values of dependent variables from independent variables. Correlation can be linear, non-linear, of non-existent, and can be strong, moderate, of weak. The strength of a linear relationship is measured by the correlation coefficient, which can be positive of negative and ranges from very weak to very strong. However, the interpretation of a correlation coefficient depends on the context and purposes. Correlation can suggest association but cannot prove of disprove causation. Linear regression, on the other hand, can be used to predict how much one variable changes when a second variable is changed. Scatter graphs are used in correlation and regression analyses to visually determine if variables are associated and to detect outliers. When constructing a scatter graph, the dependent variable is typically placed on the vertical axis and the independent variable on the horizontal axis.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 39 - Which statistical test is appropriate for analyzing normally distributed data that is measured?...

    Incorrect

    • Which statistical test is appropriate for analyzing normally distributed data that is measured?

      Your Answer:

      Correct Answer: Independent t-test

      Explanation:

      The t-test is appropriate for analyzing data that meets parametric assumptions, while other tests are more suitable for non-parametric data.

      Choosing the right statistical test can be challenging, but understanding the basic principles can help. Different tests have different assumptions, and using the wrong one can lead to inaccurate results. To identify the appropriate test, a flow chart can be used based on three main factors: the type of dependent variable, the type of data, and whether the groups/samples are independent of dependent. It is important to know which tests are parametric and non-parametric, as well as their alternatives. For example, the chi-squared test is used to assess differences in categorical variables and is non-parametric, while Pearson’s correlation coefficient measures linear correlation between two variables and is parametric. T-tests are used to compare means between two groups, and ANOVA is used to compare means between more than two groups. Non-parametric equivalents to ANOVA include the Kruskal-Wallis analysis of ranks, the Median test, Friedman’s two-way analysis of variance, and Cochran Q test. Understanding these tests and their assumptions can help researchers choose the appropriate statistical test for their data.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 40 - In scientific research, what variable type has traditionally been used to record the...

    Incorrect

    • In scientific research, what variable type has traditionally been used to record the age of study participants?

      Your Answer:

      Correct Answer: Binary

      Explanation:

      Gender has traditionally been recorded as either male of female, creating a binary of dichotomous variable. Other categorical variables, such as eye color and ethnicity, can be grouped into two or more categories. Continuous variables, such as temperature, height, weight, and age, can be placed anywhere on a scale and have mathematical properties. Ordinal variables allow for ranking, but do not allow for direct mathematical comparisons between values.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 41 - What methods are most effective in determining interobserver agreement? ...

    Incorrect

    • What methods are most effective in determining interobserver agreement?

      Your Answer:

      Correct Answer: Kappa

      Explanation:

      Kappa is used to assess the consistency of reliability between different raters.

      Understanding the Kappa Statistic for Measuring Interobserver Variation

      The kappa statistic, also known as Cohen’s kappa coefficient, is a useful tool for quantifying the level of agreement between independent observers. This measure can be applied in any situation where multiple observers are evaluating the same thing, such as in medical diagnoses of research studies. The kappa coefficient ranges from 0 to 1, with 0 indicating complete disagreement and 1 indicating perfect agreement. By using the kappa statistic, researchers and practitioners can gain insight into the level of interobserver variation present in their data, which can help to improve the accuracy and reliability of their findings. Overall, the kappa statistic is a valuable tool for understanding and measuring interobserver variation in a variety of contexts.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 42 - What is the appropriate denominator for calculating cumulative incidence? ...

    Incorrect

    • What is the appropriate denominator for calculating cumulative incidence?

      Your Answer:

      Correct Answer: The number of disease free people at the beginning of a specified time period

      Explanation:

      Measures of Disease Frequency: Incidence and Prevalence

      Incidence and prevalence are two important measures of disease frequency. Incidence measures the speed at which new cases of a disease are emerging, while prevalence measures the burden of disease within a population. Cumulative incidence and incidence rate are two types of incidence measures, while point prevalence and period prevalence are two types of prevalence measures.

      Cumulative incidence is the average risk of getting a disease over a certain period of time, while incidence rate is a measure of the speed at which new cases are emerging. Prevalence is a proportion and is a measure of the burden of disease within a population. Point prevalence measures the number of cases in a defined population at a specific point in time, while period prevalence measures the number of identified cases during a specified period of time.

      It is important to note that prevalence is equal to incidence multiplied by the duration of the condition. In chronic diseases, the prevalence is much greater than the incidence. The incidence rate is stated in units of person-time, while cumulative incidence is always a proportion. When describing cumulative incidence, it is necessary to give the follow-up period over which the risk is estimated. In acute diseases, the prevalence and incidence may be similar, while for conditions such as the common cold, the incidence may be greater than the prevalence.

      Incidence is a useful measure to study disease etiology and risk factors, while prevalence is useful for health resource planning. Understanding these measures of disease frequency is important for public health professionals and researchers in order to effectively monitor and address the burden of disease within populations.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 43 - Which of the following is the correct description of construct validity? ...

    Incorrect

    • Which of the following is the correct description of construct validity?

      Your Answer:

      Correct Answer: A test has good construct validity if it has a high correlation with another test that measures the same construct

      Explanation:

      Validity in statistics refers to how accurately something measures what it claims to measure. There are two main types of validity: internal and external. Internal validity refers to the confidence we have in the cause and effect relationship in a study, while external validity refers to the degree to which the conclusions of a study can be applied to other people, places, and times. There are various threats to both internal and external validity, such as sampling, measurement instrument obtrusiveness, and reactive effects of setting. Additionally, there are several subtypes of validity, including face validity, content validity, criterion validity, and construct validity. Each subtype has its own specific focus and methods for testing validity.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 44 - What is the term used to describe the likelihood of correctly rejecting the...

    Incorrect

    • What is the term used to describe the likelihood of correctly rejecting the null hypothesis when it is actually false?

      Your Answer:

      Correct Answer: Power of the test

      Explanation:

      Understanding Hypothesis Testing in Statistics

      In statistics, it is not feasible to investigate hypotheses on entire populations. Therefore, researchers take samples and use them to make estimates about the population they are drawn from. However, this leads to uncertainty as there is no guarantee that the sample taken will be truly representative of the population, resulting in potential errors. Statistical hypothesis testing is the process used to determine if claims from samples to populations can be made and with what certainty.

      The null hypothesis (Ho) is the claim that there is no real difference between two groups, while the alternative hypothesis (H1 of Ha) suggests that any difference is due to some non-random chance. The alternative hypothesis can be one-tailed of two-tailed, depending on whether it seeks to establish a difference of a change in one direction.

      Two types of errors may occur when testing the null hypothesis: Type I and Type II errors. Type I error occurs when the null hypothesis is rejected when it is true, while Type II error occurs when the null hypothesis is accepted when it is false. The power of a study is the probability of correctly rejecting the null hypothesis when it is false, and it can be increased by increasing the sample size.

      P-values provide information on statistical significance and help researchers decide if study results have occurred due to chance. The p-value is the probability of obtaining a result that is as large of larger when in reality there is no difference between two groups. The cutoff for the p-value is called the significance level (alpha level), typically set at 0.05. If the p-value is less than the cutoff, the null hypothesis is rejected, and if it is greater or equal to the cut off, the null hypothesis is not rejected. However, the p-value does not indicate clinical significance, which may be too small to be meaningful.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 45 - What statement accurately describes the process of searching a database? ...

    Incorrect

    • What statement accurately describes the process of searching a database?

      Your Answer:

      Correct Answer: New references are added to PubMed more quickly than they are to MEDLINE

      Explanation:

      PubMed receives new references faster than MEDLINE because they do not need to undergo indexing, such as adding MeSH headings and checking tags. While an increasing number of MEDLINE citations have a link to the complete article, not all of them do. Since 2010, Embased has included all MEDLINE citations in its database, but it does not have all citations from before that year.

      Evidence-based medicine involves four basic steps: developing a focused clinical question, searching for the best evidence, critically appraising the evidence, and applying the evidence and evaluating the outcome. When developing a question, it is important to understand the difference between background and foreground questions. Background questions are general questions about conditions, illnesses, syndromes, and pathophysiology, while foreground questions are more often about issues of care. The PICO system is often used to define the components of a foreground question: patient group of interest, intervention of interest, comparison, and primary outcome.

      When searching for evidence, it is important to have a basic understanding of the types of evidence and sources of information. Scientific literature is divided into two basic categories: primary (empirical research) and secondary (interpretation and analysis of primary sources). Unfiltered sources are large databases of articles that have not been pre-screened for quality, while filtered resources summarize and appraise evidence from several studies.

      There are several databases and search engines that can be used to search for evidence, including Medline and PubMed, Embase, the Cochrane Library, PsycINFO, CINAHL, and OpenGrey. Boolean logic can be used to combine search terms in PubMed, and phrase searching and truncation can also be used. Medical Subject Headings (MeSH) are used by indexers to describe articles for MEDLINE records, and the MeSH Database is like a thesaurus that enables exploration of this vocabulary.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 46 - Which study design involves conducting an experiment? ...

    Incorrect

    • Which study design involves conducting an experiment?

      Your Answer:

      Correct Answer: A randomised control study

      Explanation:

      Types of Primary Research Studies and Their Advantages and Disadvantages

      Primary research studies can be categorized into six types based on the research question they aim to address. The best type of study for each question type is listed in the table below. There are two main types of study design: experimental and observational. Experimental studies involve an intervention, while observational studies do not. The advantages and disadvantages of each study type are summarized in the table below.

      Type of Question Best Type of Study

      Therapy Randomized controlled trial (RCT), cohort, case control, case series
      Diagnosis Cohort studies with comparison to gold standard test
      Prognosis Cohort studies, case control, case series
      Etiology/Harm RCT, cohort studies, case control, case series
      Prevention RCT, cohort studies, case control, case series
      Cost Economic analysis

      Study Type Advantages Disadvantages

      Randomized Controlled Trial – Unbiased distribution of confounders – Blinding more likely – Randomization facilitates statistical analysis – Expensive – Time-consuming – Volunteer bias – Ethically problematic at times
      Cohort Study – Ethically safe – Subjects can be matched – Can establish timing and directionality of events – Eligibility criteria and outcome assessments can be standardized – Administratively easier and cheaper than RCT – Controls may be difficult to identify – Exposure may be linked to a hidden confounder – Blinding is difficult – Randomization not present – For rare disease, large sample sizes of long follow-up necessary
      Case-Control Study – Quick and cheap – Only feasible method for very rare disorders of those with long lag between exposure and outcome – Fewer subjects needed than cross-sectional studies – Reliance on recall of records to determine exposure status – Confounders – Selection of control groups is difficult – Potential bias: recall, selection
      Cross-Sectional Survey – Cheap and simple – Ethically safe – Establishes association at most, not causality – Recall bias susceptibility – Confounders may be unequally distributed – Neyman bias – Group sizes may be unequal
      Ecological Study – Cheap and simple – Ethically safe – Ecological fallacy (when relationships which exist for groups are assumed to also be true for individuals)

      In conclusion, the choice of study type depends on the research question being addressed. Each study type has its own advantages and disadvantages, and researchers should carefully consider these when designing their studies.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 47 - Which option is not a type of descriptive statistic? ...

    Incorrect

    • Which option is not a type of descriptive statistic?

      Your Answer:

      Correct Answer: Student's t-test

      Explanation:

      A t-test is a statistical method used to determine if there is a significant difference between the means of two groups. It is a type of statistical inference.

      Types of Statistics: Descriptive and Inferential

      Statistics can be divided into two categories: descriptive and inferential. Descriptive statistics are used to describe and summarize data without making any generalizations beyond the data at hand. On the other hand, inferential statistics are used to make inferences about a population based on sample data.

      Descriptive statistics are useful for identifying patterns and trends in data. Common measures used to describe a data set include measures of central tendency (such as the mean, median, and mode) and measures of variability of dispersion (such as the standard deviation of variance).

      Inferential statistics, on the other hand, are used to make predictions of draw conclusions about a population based on sample data. These statistics are also used to determine the probability that observed differences between groups are reliable and not due to chance.

      Overall, both descriptive and inferential statistics play important roles in analyzing and interpreting data. Descriptive statistics help us understand the characteristics of a data set, while inferential statistics allow us to make predictions and draw conclusions about larger populations.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 48 - What is the term used to describe the study design where a margin...

    Incorrect

    • What is the term used to describe the study design where a margin is set for the mean reduction of PANSS score, and if the confidence interval of the difference between the new drug and olanzapine falls within this margin, the trial is considered successful?

      Your Answer:

      Correct Answer: Equivalence trial

      Explanation:

      Study Designs for New Drugs: Options and Considerations

      When launching a new drug, there are various study design options available. One common approach is a placebo-controlled trial, which can provide strong evidence but may be deemed unethical if established treatments are available. Additionally, it does not allow for a comparison with standard treatments. Therefore, statisticians must decide whether the trial aims to demonstrate superiority, equivalence, of non-inferiority to an existing treatment.

      Superiority trials may seem like the obvious choice, but they require a large sample size to show a significant benefit over an existing treatment. Equivalence trials define an equivalence margin on a specified outcome, and if the confidence interval of the difference between the two drugs falls within this margin, the drugs are assumed to have a similar effect. Non-inferiority trials are similar to equivalence trials, but only the lower confidence interval needs to fall within the equivalence margin. These trials require smaller sample sizes, and once a drug has been shown to be non-inferior, larger studies may be conducted to demonstrate superiority.

      It is important to note that drug companies may not necessarily aim to show superiority over an existing product. If they can demonstrate that their product is equivalent of even non-inferior, they may compete on price of convenience. Overall, the choice of study design depends on various factors, including ethical considerations, sample size, and the desired outcome.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 49 - What is another name for the incidence rate? ...

    Incorrect

    • What is another name for the incidence rate?

      Your Answer:

      Correct Answer: Incidence density

      Explanation:

      Measures of Disease Frequency: Incidence and Prevalence

      Incidence and prevalence are two important measures of disease frequency. Incidence measures the speed at which new cases of a disease are emerging, while prevalence measures the burden of disease within a population. Cumulative incidence and incidence rate are two types of incidence measures, while point prevalence and period prevalence are two types of prevalence measures.

      Cumulative incidence is the average risk of getting a disease over a certain period of time, while incidence rate is a measure of the speed at which new cases are emerging. Prevalence is a proportion and is a measure of the burden of disease within a population. Point prevalence measures the number of cases in a defined population at a specific point in time, while period prevalence measures the number of identified cases during a specified period of time.

      It is important to note that prevalence is equal to incidence multiplied by the duration of the condition. In chronic diseases, the prevalence is much greater than the incidence. The incidence rate is stated in units of person-time, while cumulative incidence is always a proportion. When describing cumulative incidence, it is necessary to give the follow-up period over which the risk is estimated. In acute diseases, the prevalence and incidence may be similar, while for conditions such as the common cold, the incidence may be greater than the prevalence.

      Incidence is a useful measure to study disease etiology and risk factors, while prevalence is useful for health resource planning. Understanding these measures of disease frequency is important for public health professionals and researchers in order to effectively monitor and address the burden of disease within populations.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds
  • Question 50 - What is the accurate formula for determining the likelihood ratio of a negative...

    Incorrect

    • What is the accurate formula for determining the likelihood ratio of a negative test result?

      Your Answer:

      Correct Answer: (1 - sensitivity) / specificity

      Explanation:

      Clinical tests are used to determine the presence of absence of a disease of condition. To interpret test results, it is important to have a working knowledge of statistics used to describe them. Two by two tables are commonly used to calculate test statistics such as sensitivity and specificity. Sensitivity refers to the proportion of people with a condition that the test correctly identifies, while specificity refers to the proportion of people without a condition that the test correctly identifies. Accuracy tells us how closely a test measures to its true value, while predictive values help us understand the likelihood of having a disease based on a positive of negative test result. Likelihood ratios combine sensitivity and specificity into a single figure that can refine our estimation of the probability of a disease being present. Pre and post-test odds and probabilities can also be calculated to better understand the likelihood of having a disease before and after a test is carried out. Fagan’s nomogram is a useful tool for calculating post-test probabilities.

    • This question is part of the following fields:

      • Research Methods, Statistics, Critical Review And Evidence-Based Practice
      0
      Seconds

SESSION STATS - PERFORMANCE PER SPECIALTY

Research Methods, Statistics, Critical Review And Evidence-Based Practice (5/9) 56%
Passmed