-
Question 1
Correct
-
A team of scientists plans to carry out a randomized controlled study to assess the effectiveness of a new medication for treating anxiety in elderly patients. To prevent any potential biases, they intend to enroll participants through online portals, ensuring that neither the patients nor the researchers are aware of the group assignment. What type of bias are they seeking to eliminate?
Your Answer: Selection bias
Explanation:The use of allocation concealment is being implemented by the researchers to prevent interference from investigators of patients in the randomisation process. This is important as knowledge of group allocation can lead to patient refusal to participate of researchers manipulating the allocation process. By using distant call centres for allocation concealment, the risk of selection bias, which refers to systematic differences between comparison groups, is reduced.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 2
Correct
-
A masters student had noticed that nearly all of her patients with arthritis were over the age of 50. She was keen to investigate this further to see if there was an association.
She selected 100 patients with arthritis and 100 controls. of the 100 patients with arthritis, 90 were over the age of 50. of the 100 controls, only 40 were over the age of 50.
What is the odds ratio?Your Answer: 3.77
Explanation:The odds of being married are 3.77 times higher in individuals with panic disorder compared to controls.
Measures of Effect in Clinical Studies
When conducting clinical studies, we often want to know the effect of treatments of exposures on health outcomes. Measures of effect are used in randomized controlled trials (RCTs) and include the odds ratio (of), risk ratio (RR), risk difference (RD), and number needed to treat (NNT). Dichotomous (binary) outcome data are common in clinical trials, where the outcome for each participant is one of two possibilities, such as dead of alive, of clinical improvement of no improvement.
To understand the difference between of and RR, it’s important to know the difference between risks and odds. Risk is a proportion that describes the probability of a health outcome occurring, while odds is a ratio that compares the probability of an event occurring to the probability of it not occurring. Absolute risk is the basic risk, while risk difference is the difference between the absolute risk of an event in the intervention group and the absolute risk in the control group. Relative risk is the ratio of risk in the intervention group to the risk in the control group.
The number needed to treat (NNT) is the number of patients who need to be treated for one to benefit. Odds are calculated by dividing the number of times an event happens by the number of times it does not happen. The odds ratio is the odds of an outcome given a particular exposure versus the odds of an outcome in the absence of the exposure. It is commonly used in case-control studies and can also be used in cross-sectional and cohort study designs. An odds ratio of 1 indicates no difference in risk between the two groups, while an odds ratio >1 indicates an increased risk and an odds ratio <1 indicates a reduced risk.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 3
Correct
-
What statement accurately describes the mean?
Your Answer: Is sensitive to a change in any value in the data set
Explanation:Measures of Central Tendency
Measures of central tendency are used in descriptive statistics to summarize the middle of typical value of a data set. There are three common measures of central tendency: the mean, median, and mode.
The median is the middle value in a data set that has been arranged in numerical order. It is not affected by outliers and is used for ordinal data. The mode is the most frequent value in a data set and is used for categorical data. The mean is calculated by adding all the values in a data set and dividing by the number of values. It is sensitive to outliers and is used for interval and ratio data.
The appropriate measure of central tendency depends on the measurement scale of the data. For nominal and categorical data, the mode is used. For ordinal data, the median of mode is used. For interval data with a normal distribution, the mean is preferable, but the median of mode can also be used. For interval data with skewed distribution, the median is used. For ratio data, the mean is preferable, but the median of mode can also be used for skewed data.
In addition to measures of central tendency, the range is also used to describe the spread of a data set. It is calculated by subtracting the smallest value from the largest value.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 4
Correct
-
What is the standard deviation of the sample mean weight of 64 patients diagnosed with paranoid schizophrenia, given that the average weight is 81 kg and the standard deviation is 12 kg?
Your Answer: 1.5
Explanation:– The standard error of the mean is calculated using the formula: standard deviation / square root (number of patients).
– In this case, the standard error of the mean is 12 / square root (64).
– Simplifying this equation gives a standard error of the mean of 12 / 8.Measures of dispersion are used to indicate the variation of spread of a data set, often in conjunction with a measure of central tendency such as the mean of median. The range, which is the difference between the largest and smallest value, is the simplest measure of dispersion. The interquartile range, which is the difference between the 3rd and 1st quartiles, is another useful measure. Quartiles divide a data set into quarters, and the interquartile range can provide additional information about the spread of the data. However, to get a more representative idea of spread, measures such as the variance and standard deviation are needed. The variance gives an indication of how much the items in the data set vary from the mean, while the standard deviation reflects the distribution of individual scores around their mean. The standard deviation is expressed in the same units as the data set and can be used to indicate how confident we are that data points lie within a particular range. The standard error of the mean is an inferential statistic used to estimate the population mean and is a measure of the spread expected for the mean of the observations. Confidence intervals are often presented alongside sample results such as the mean value, indicating a range that is likely to contain the true value.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 5
Incorrect
-
Which of the following is another term for the average of squared deviations from the mean?
Your Answer: Standard deviation
Correct Answer: Variance
Explanation:The variance can be expressed as the mean of the squared differences between each value and the mean.
Measures of dispersion are used to indicate the variation of spread of a data set, often in conjunction with a measure of central tendency such as the mean of median. The range, which is the difference between the largest and smallest value, is the simplest measure of dispersion. The interquartile range, which is the difference between the 3rd and 1st quartiles, is another useful measure. Quartiles divide a data set into quarters, and the interquartile range can provide additional information about the spread of the data. However, to get a more representative idea of spread, measures such as the variance and standard deviation are needed. The variance gives an indication of how much the items in the data set vary from the mean, while the standard deviation reflects the distribution of individual scores around their mean. The standard deviation is expressed in the same units as the data set and can be used to indicate how confident we are that data points lie within a particular range. The standard error of the mean is an inferential statistic used to estimate the population mean and is a measure of the spread expected for the mean of the observations. Confidence intervals are often presented alongside sample results such as the mean value, indicating a range that is likely to contain the true value.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 6
Incorrect
-
What is a correct statement about funnel plots?
Your Answer: An asymmetric funnel shape confirms the presence of publication bias
Correct Answer: Each dot represents a separate study result
Explanation:An asymmetric funnel plot may indicate the presence of publication bias, although this is not a definitive confirmation. The x-axis typically represents a measure of effect, such as the risk ratio of odds ratio, although other measures may also be used.
Stats Publication Bias
Publication bias refers to the tendency for studies with positive findings to be published more than studies with negative findings, leading to incomplete data sets in meta-analyses and erroneous conclusions. Graphical methods such as funnel plots, Galbraith plots, ordered forest plots, and normal quantile plots can be used to detect publication bias. Funnel plots are the most commonly used and offer an easy visual way to ensure that published literature is evenly weighted. The x-axis represents the effect size, and the y-axis represents the study size. A symmetrical, inverted funnel shape indicates that publication bias is unlikely, while an asymmetrical funnel indicates a relationship between treatment effect and study size, indicating either publication bias of small study effects.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 7
Incorrect
-
What statement accurately describes measures of dispersion?
Your Answer: The units of variance are expressed as the same as the data set from which it is calculated
Correct Answer: The standard error indicates how close the statistical mean is to the population mean
Explanation:Measures of dispersion are used to indicate the variation of spread of a data set, often in conjunction with a measure of central tendency such as the mean of median. The range, which is the difference between the largest and smallest value, is the simplest measure of dispersion. The interquartile range, which is the difference between the 3rd and 1st quartiles, is another useful measure. Quartiles divide a data set into quarters, and the interquartile range can provide additional information about the spread of the data. However, to get a more representative idea of spread, measures such as the variance and standard deviation are needed. The variance gives an indication of how much the items in the data set vary from the mean, while the standard deviation reflects the distribution of individual scores around their mean. The standard deviation is expressed in the same units as the data set and can be used to indicate how confident we are that data points lie within a particular range. The standard error of the mean is an inferential statistic used to estimate the population mean and is a measure of the spread expected for the mean of the observations. Confidence intervals are often presented alongside sample results such as the mean value, indicating a range that is likely to contain the true value.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 8
Correct
-
What is the nature of the hypothesis that a researcher wants to test regarding the effect of a drug on a person's heart rate?
Your Answer: One-tailed alternative hypothesis
Explanation:A one-tailed hypothesis indicates a specific direction of association between groups. The researcher not only declares that there will be a distinction between the groups but also defines the direction in which the difference will occur.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 9
Incorrect
-
After creating a scatter plot of the data, what would be the next step for the researcher to determine if there is a linear relationship between a person's age and blood pressure?
Your Answer: T-test
Correct Answer: Pearson's coefficient
Explanation:Choosing the right statistical test can be challenging, but understanding the basic principles can help. Different tests have different assumptions, and using the wrong one can lead to inaccurate results. To identify the appropriate test, a flow chart can be used based on three main factors: the type of dependent variable, the type of data, and whether the groups/samples are independent of dependent. It is important to know which tests are parametric and non-parametric, as well as their alternatives. For example, the chi-squared test is used to assess differences in categorical variables and is non-parametric, while Pearson’s correlation coefficient measures linear correlation between two variables and is parametric. T-tests are used to compare means between two groups, and ANOVA is used to compare means between more than two groups. Non-parametric equivalents to ANOVA include the Kruskal-Wallis analysis of ranks, the Median test, Friedman’s two-way analysis of variance, and Cochran Q test. Understanding these tests and their assumptions can help researchers choose the appropriate statistical test for their data.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 10
Correct
-
Which variable classification is not included in Stevens' typology?
Your Answer: Ranked
Explanation:Stevens suggested that scales can be categorized into one of four types based on measurements.
Scales of Measurement in Statistics
In the 1940s, Stanley Smith Stevens introduced four scales of measurement to categorize data variables. Knowing the scale of measurement for a variable is crucial in selecting the appropriate statistical analysis. The four scales of measurement are ratio, interval, ordinal, and nominal.
Ratio scales are similar to interval scales, but they have true zero points. Examples of ratio scales include weight, time, and length. Interval scales measure the difference between two values, and one unit on the scale represents the same magnitude on the trait of characteristic being measured across the whole range of the scale. The Fahrenheit scale for temperature is an example of an interval scale.
Ordinal scales categorize observed values into set categories that can be ordered, but the intervals between each value are uncertain. Examples of ordinal scales include social class, education level, and income level. Nominal scales categorize observed values into set categories that have no particular order of hierarchy. Examples of nominal scales include genotype, blood type, and political party.
Data can also be categorized as quantitative of qualitative. Quantitative variables take on numeric values and can be further classified into discrete and continuous types. Qualitative variables do not take on numerical values and are usually names. Some qualitative variables have an inherent order in their categories and are described as ordinal. Qualitative variables are also called categorical of nominal variables. When a qualitative variable has only two categories, it is called a binary variable.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 11
Incorrect
-
In a study of a new statin therapy for primary prevention of ischaemic heart disease in a diabetic population over a five year period, 1000 patients were randomly assigned to receive the new therapy and 1000 were given a placebo. The results showed that 150 patients in the placebo group had a myocardial infarction (MI) compared to 100 patients in the statin group. What is the number needed to treat (NNT) to prevent one MI in this population?
Your Answer: 40
Correct Answer: 20
Explanation:– Treating 1000 patients with a new statin for five years prevented 50 MIs.
– The number needed to treat (NNT) to prevent one MI is 20 (1000/50).
– NNT provides information on treatment efficacy beyond statistical significance.
– Based on these data, treating as few as 20 patients over five years may prevent an infarct.
– Cost economic data can be calculated by factoring in drug costs and costs of treating and rehabilitating a patient with an MI. -
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 12
Incorrect
-
A team of scientists aimed to examine the prognosis of late-onset Alzheimer's disease using the available evidence. They intend to arrange the evidence in a hierarchy based on their study designs.
What study design would be placed at the top of their hierarchy?Your Answer: Expert opinion
Correct Answer: Systematic review of cohort studies
Explanation:When investigating prognosis, the hierarchy of study designs starts with a systematic review of cohort studies, followed by a cohort study, follow-up of untreated patients from randomized controlled trials, case series, and expert opinion. The strength of evidence provided by a study depends on its ability to minimize bias and maximize attribution. The Agency for Healthcare Policy and Research hierarchy of study types is widely accepted as reliable, with systematic reviews and meta-analyses of randomized controlled trials at the top, followed by randomized controlled trials, non-randomized intervention studies, observational studies, and non-experimental studies.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 13
Incorrect
-
A researcher wants to compare the mean age of two groups of participants who were randomly assigned to either a standard exercise program of a standard exercise program + new supplement. The data collected is parametric and continuous. What is the most appropriate statistical test to use?
Your Answer: Chi square test
Correct Answer: Unpaired t test
Explanation:The two sample unpaired t test is utilized to examine whether the null hypothesis that the two populations related to the two random samples are equivalent is true of not. When dealing with continuous data that is believed to conform to the normal distribution, a t test is suitable, making it appropriate for comparing weight loss between two groups. In contrast, a paired t test is used when the data is dependent, meaning there is a direct correlation between the values in the two samples. This could include the same subject being measured before and after a process change of at different times.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 14
Incorrect
-
What term is used to describe an association between two variables that is influenced by a confounding factor?
Your Answer: Direct
Correct Answer: Indirect
Explanation:Stats Association and Causation
When two variables are found to be more commonly present together, they are said to be associated. However, this association can be of three types: spurious, indirect, of direct. Spurious association is one that has arisen by chance and is not real, while indirect association is due to the presence of another factor, known as a confounding variable. Direct association, on the other hand, is a true association not linked by a third variable.
Once an association has been established, the next question is whether it is causal. To determine causation, the Bradford Hill Causal Criteria are used. These criteria include strength, temporality, specificity, coherence, and consistency. The stronger the association, the more likely it is to be truly causal. Temporality refers to whether the exposure precedes the outcome. Specificity asks whether the suspected cause is associated with a specific outcome of disease. Coherence refers to whether the association fits with other biological knowledge. Finally, consistency asks whether the same association is found in many studies.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 15
Incorrect
-
What is the purpose of using Cohen's kappa coefficient?
Your Answer: Internal consistency
Correct Answer: Inter-rater reliability
Explanation:Kappa is used to assess the consistency of agreement between different raters.
Understanding the Kappa Statistic for Measuring Interobserver Variation
The kappa statistic, also known as Cohen’s kappa coefficient, is a useful tool for quantifying the level of agreement between independent observers. This measure can be applied in any situation where multiple observers are evaluating the same thing, such as in medical diagnoses of research studies. The kappa coefficient ranges from 0 to 1, with 0 indicating complete disagreement and 1 indicating perfect agreement. By using the kappa statistic, researchers and practitioners can gain insight into the level of interobserver variation present in their data, which can help to improve the accuracy and reliability of their findings. Overall, the kappa statistic is a valuable tool for understanding and measuring interobserver variation in a variety of contexts.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 16
Incorrect
-
What is the significance of the cut off of 5 on the MDQ in diagnosing depression?
Your Answer: The false positive rate
Correct Answer: The optimal threshold
Explanation:The threshold score that results in the lowest misclassification rate, achieved by minimizing both false positive and false negative rates, is known as the optimal threshold. Based on the findings of the previous study, the ideal cut off for identifying caseness on the MDQ is five, making it the optimal threshold.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 17
Incorrect
-
What is the GRADE approach used in evidence based medicine and what are its characteristics?
Your Answer: A high probability of reporting bias increases the grade quality of a study
Correct Answer: The system can be applied to observational studies
Explanation:Levels and Grades of Evidence in Evidence-Based Medicine
To evaluate the quality of evidence on a subject of question, levels of grades are used. The traditional hierarchy approach places systematic reviews of randomized control trials at the top and case-series/report at the bottom. However, this approach is overly simplistic as certain research questions cannot be answered using RCTs. To address this, the Oxford Centre for Evidence-Based Medicine introduced their 2011 Levels of Evidence system, which separates the type of study questions and gives a hierarchy for each.
The grading approach to be aware of is the GRADE system, which classifies the quality of evidence as high, moderate, low, of very low. The process begins by formulating a study question and identifying specific outcomes. Outcomes are then graded as critical of important. The evidence is then gathered and criteria are used to grade the evidence, with the type of evidence being a significant factor. Evidence can be promoted of downgraded based on certain criteria, such as limitations to study quality, inconsistency, uncertainty about directness, imprecise of sparse data, and reporting bias. The GRADE system allows for the promotion of observational studies to high-quality evidence under the right circumstances.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 18
Incorrect
-
What is the standard deviation of the sample mean height of 100 adults who were administered steroids during childhood, given that the average height of the adults is 169cm and the standard deviation is 16cm?
Your Answer: 1.3
Correct Answer: 1.6
Explanation:The standard error of the mean is 1.6, calculated by dividing the standard deviation of 16 by the square root of the number of patients, which is 100.
Measures of dispersion are used to indicate the variation of spread of a data set, often in conjunction with a measure of central tendency such as the mean of median. The range, which is the difference between the largest and smallest value, is the simplest measure of dispersion. The interquartile range, which is the difference between the 3rd and 1st quartiles, is another useful measure. Quartiles divide a data set into quarters, and the interquartile range can provide additional information about the spread of the data. However, to get a more representative idea of spread, measures such as the variance and standard deviation are needed. The variance gives an indication of how much the items in the data set vary from the mean, while the standard deviation reflects the distribution of individual scores around their mean. The standard deviation is expressed in the same units as the data set and can be used to indicate how confident we are that data points lie within a particular range. The standard error of the mean is an inferential statistic used to estimate the population mean and is a measure of the spread expected for the mean of the observations. Confidence intervals are often presented alongside sample results such as the mean value, indicating a range that is likely to contain the true value.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 19
Incorrect
-
You are asked to design a study to assess whether living near electricity pylons is a risk factor for adult leukemia. What is the most appropriate type of study design?:
Your Answer: Randomised controlled trial
Correct Answer: Case-control study
Explanation:Due to the low incidence of childhood leukaemia, a cohort study would require a significant amount of time to yield meaningful findings.
Types of Primary Research Studies and Their Advantages and Disadvantages
Primary research studies can be categorized into six types based on the research question they aim to address. The best type of study for each question type is listed in the table below. There are two main types of study design: experimental and observational. Experimental studies involve an intervention, while observational studies do not. The advantages and disadvantages of each study type are summarized in the table below.
Type of Question Best Type of Study
Therapy Randomized controlled trial (RCT), cohort, case control, case series
Diagnosis Cohort studies with comparison to gold standard test
Prognosis Cohort studies, case control, case series
Etiology/Harm RCT, cohort studies, case control, case series
Prevention RCT, cohort studies, case control, case series
Cost Economic analysisStudy Type Advantages Disadvantages
Randomized Controlled Trial – Unbiased distribution of confounders – Blinding more likely – Randomization facilitates statistical analysis – Expensive – Time-consuming – Volunteer bias – Ethically problematic at times
Cohort Study – Ethically safe – Subjects can be matched – Can establish timing and directionality of events – Eligibility criteria and outcome assessments can be standardized – Administratively easier and cheaper than RCT – Controls may be difficult to identify – Exposure may be linked to a hidden confounder – Blinding is difficult – Randomization not present – For rare disease, large sample sizes of long follow-up necessary
Case-Control Study – Quick and cheap – Only feasible method for very rare disorders of those with long lag between exposure and outcome – Fewer subjects needed than cross-sectional studies – Reliance on recall of records to determine exposure status – Confounders – Selection of control groups is difficult – Potential bias: recall, selection
Cross-Sectional Survey – Cheap and simple – Ethically safe – Establishes association at most, not causality – Recall bias susceptibility – Confounders may be unequally distributed – Neyman bias – Group sizes may be unequal
Ecological Study – Cheap and simple – Ethically safe – Ecological fallacy (when relationships which exist for groups are assumed to also be true for individuals)In conclusion, the choice of study type depends on the research question being addressed. Each study type has its own advantages and disadvantages, and researchers should carefully consider these when designing their studies.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 20
Incorrect
-
Which odds ratio suggests that there is no significant variation in the odds between two groups?
Your Answer: -1.5
Correct Answer: 1
Explanation:Measures of Effect in Clinical Studies
When conducting clinical studies, we often want to know the effect of treatments of exposures on health outcomes. Measures of effect are used in randomized controlled trials (RCTs) and include the odds ratio (of), risk ratio (RR), risk difference (RD), and number needed to treat (NNT). Dichotomous (binary) outcome data are common in clinical trials, where the outcome for each participant is one of two possibilities, such as dead of alive, of clinical improvement of no improvement.
To understand the difference between of and RR, it’s important to know the difference between risks and odds. Risk is a proportion that describes the probability of a health outcome occurring, while odds is a ratio that compares the probability of an event occurring to the probability of it not occurring. Absolute risk is the basic risk, while risk difference is the difference between the absolute risk of an event in the intervention group and the absolute risk in the control group. Relative risk is the ratio of risk in the intervention group to the risk in the control group.
The number needed to treat (NNT) is the number of patients who need to be treated for one to benefit. Odds are calculated by dividing the number of times an event happens by the number of times it does not happen. The odds ratio is the odds of an outcome given a particular exposure versus the odds of an outcome in the absence of the exposure. It is commonly used in case-control studies and can also be used in cross-sectional and cohort study designs. An odds ratio of 1 indicates no difference in risk between the two groups, while an odds ratio >1 indicates an increased risk and an odds ratio <1 indicates a reduced risk.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 21
Incorrect
-
Which category does convenience sampling fall under?
Your Answer: Cluster sampling
Correct Answer: Non-probabilistic sampling
Explanation:Sampling Methods in Statistics
When collecting data from a population, it is often impractical and unnecessary to gather information from every single member. Instead, taking a sample is preferred. However, it is crucial that the sample accurately represents the population from which it is drawn. There are two main types of sampling methods: probability (random) sampling and non-probability (non-random) sampling.
Non-probability sampling methods, also known as judgement samples, are based on human choice rather than random selection. These samples are convenient and cheaper than probability sampling methods. Examples of non-probability sampling methods include voluntary sampling, convenience sampling, snowball sampling, and quota sampling.
Probability sampling methods give a more representative sample of the population than non-probability sampling. In each probability sampling technique, each population element has a known (non-zero) chance of being selected for the sample. Examples of probability sampling methods include simple random sampling, systematic sampling, cluster sampling, stratified sampling, and multistage sampling.
Simple random sampling is a sample in which every member of the population has an equal chance of being chosen. Systematic sampling involves selecting every kth member of the population. Cluster sampling involves dividing a population into separate groups (called clusters) and selecting a random sample of clusters. Stratified sampling involves dividing a population into groups (strata) and taking a random sample from each strata. Multistage sampling is a more complex method that involves several stages and combines two of more sampling methods.
Overall, probability sampling methods give a more representative sample of the population, but non-probability sampling methods are often more convenient and cheaper. It is important to choose the appropriate sampling method based on the research question and available resources.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 22
Incorrect
-
What is the appropriate denominator for calculating the incidence rate?
Your Answer: The number of new cases in a specified time period
Correct Answer: The total person time at risk during a specified time period
Explanation:Measures of Disease Frequency: Incidence and Prevalence
Incidence and prevalence are two important measures of disease frequency. Incidence measures the speed at which new cases of a disease are emerging, while prevalence measures the burden of disease within a population. Cumulative incidence and incidence rate are two types of incidence measures, while point prevalence and period prevalence are two types of prevalence measures.
Cumulative incidence is the average risk of getting a disease over a certain period of time, while incidence rate is a measure of the speed at which new cases are emerging. Prevalence is a proportion and is a measure of the burden of disease within a population. Point prevalence measures the number of cases in a defined population at a specific point in time, while period prevalence measures the number of identified cases during a specified period of time.
It is important to note that prevalence is equal to incidence multiplied by the duration of the condition. In chronic diseases, the prevalence is much greater than the incidence. The incidence rate is stated in units of person-time, while cumulative incidence is always a proportion. When describing cumulative incidence, it is necessary to give the follow-up period over which the risk is estimated. In acute diseases, the prevalence and incidence may be similar, while for conditions such as the common cold, the incidence may be greater than the prevalence.
Incidence is a useful measure to study disease etiology and risk factors, while prevalence is useful for health resource planning. Understanding these measures of disease frequency is important for public health professionals and researchers in order to effectively monitor and address the burden of disease within populations.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 23
Incorrect
-
What is the term coined by Robert Rosenthal that refers to the bias that can result from the non-publication of a few studies with negative of inconclusive results, leading to a significant impact on research in a specific field?
Your Answer: Positive results bias
Correct Answer: File drawer problem
Explanation:Publication bias refers to the tendency of researchers, editors, and pharmaceutical companies to favor the publication of studies with positive results over those with negative of inconclusive results. This bias can have various causes and can result in a skewed representation of the literature. The file drawer problem refers to the phenomenon of unpublished negative studies. HARKing, of hypothesizing after the results are known, is a form of outcome reporting bias where outcomes are selectively reported based on the strength and direction of observed associations. Begg’s funnel plot is an analytical tool used to quantify the presence of publication bias.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 24
Incorrect
-
A new antihypertensive medication is trialled for adults with high blood pressure. There are 500 adults in the control group and 300 adults assigned to take the new medication. After 6 months, 200 adults in the control group had high blood pressure compared to 30 adults in the group taking the new medication. What is the relative risk reduction?
Your Answer: 30%
Correct Answer: 75%
Explanation:The RRR (Relative Risk Reduction) is calculated by dividing the ARR (Absolute Risk Reduction) by the CER (Control Event Rate). The CER is determined by dividing the number of control events by the total number of participants, which in this case is 200/500 of 0.4. The EER (Experimental Event Rate) is determined by dividing the number of events in the experimental group by the total number of participants, which in this case is 30/300 of 0.1. The ARR is calculated by subtracting the EER from the CER, which is 0.4 – 0.1 = 0.3. Finally, the RRR is calculated by dividing the ARR by the CER, which is 0.3/0.4 of 0.75 (of 75%).
Measures of Effect in Clinical Studies
When conducting clinical studies, we often want to know the effect of treatments of exposures on health outcomes. Measures of effect are used in randomized controlled trials (RCTs) and include the odds ratio (of), risk ratio (RR), risk difference (RD), and number needed to treat (NNT). Dichotomous (binary) outcome data are common in clinical trials, where the outcome for each participant is one of two possibilities, such as dead of alive, of clinical improvement of no improvement.
To understand the difference between of and RR, it’s important to know the difference between risks and odds. Risk is a proportion that describes the probability of a health outcome occurring, while odds is a ratio that compares the probability of an event occurring to the probability of it not occurring. Absolute risk is the basic risk, while risk difference is the difference between the absolute risk of an event in the intervention group and the absolute risk in the control group. Relative risk is the ratio of risk in the intervention group to the risk in the control group.
The number needed to treat (NNT) is the number of patients who need to be treated for one to benefit. Odds are calculated by dividing the number of times an event happens by the number of times it does not happen. The odds ratio is the odds of an outcome given a particular exposure versus the odds of an outcome in the absence of the exposure. It is commonly used in case-control studies and can also be used in cross-sectional and cohort study designs. An odds ratio of 1 indicates no difference in risk between the two groups, while an odds ratio >1 indicates an increased risk and an odds ratio <1 indicates a reduced risk.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 25
Incorrect
-
What statement accurately describes population parameters?
Your Answer: A parameter is a statistic
Correct Answer: Parameters tend to have normal distributions
Explanation:Parametric vs Non-Parametric Statistics
Statistics are used to draw conclusions about a population based on a sample. A parameter is a numerical value that describes a population characteristic, but it is often impossible to know the true value of a parameter without collecting data from every individual in the population. Instead, we take a sample and use statistics to estimate the parameters.
Parametric statistical procedures assume that the population distribution is normal and that the parameters (such as means and standard deviations) are known. Examples of parametric tests include the t-test, ANOVA, and Pearson coefficient of correlation.
Non-parametric statistical procedures make few of no assumptions about the population distribution of parameters. Examples of non-parametric tests include the Mann-Whitney Test, Wilcoxon Signed-Rank Test, Kruskal-Wallis Test, and Fisher Exact Probability test.
Overall, the choice between parametric and non-parametric tests depends on the nature of the data and the research question being asked.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 26
Incorrect
-
A team of scientists aims to prevent bias in their study on the effectiveness of a new medication for elderly patients with hypertension. They randomly assign 80 patients to the treatment group, of which 60 complete the 12-week trial. Another 80 patients are assigned to the placebo group, with 75 completing the trial. The researchers agree to conduct an intention-to-treat (ITT) analysis using the LOCF method. What type of bias are they attempting to eliminate?
Your Answer: Selection bias
Correct Answer: Attrition bias
Explanation:To address the issue of drop-outs in a study, an intention to treat (ITT) analysis can be employed. Drop-outs can lead to attrition bias, which creates systematic differences in attrition across treatment groups. In an ITT analysis, all patients are included in the groups they were initially assigned to through random allocation. To handle missing data, two common methods are last observation carried forward and worst case scenario analysis.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 27
Incorrect
-
What is the mathematical operation used to determine the value of the square root of the variance?
Your Answer: Median
Correct Answer: Standard deviation
Explanation:Measures of dispersion are used to indicate the variation of spread of a data set, often in conjunction with a measure of central tendency such as the mean of median. The range, which is the difference between the largest and smallest value, is the simplest measure of dispersion. The interquartile range, which is the difference between the 3rd and 1st quartiles, is another useful measure. Quartiles divide a data set into quarters, and the interquartile range can provide additional information about the spread of the data. However, to get a more representative idea of spread, measures such as the variance and standard deviation are needed. The variance gives an indication of how much the items in the data set vary from the mean, while the standard deviation reflects the distribution of individual scores around their mean. The standard deviation is expressed in the same units as the data set and can be used to indicate how confident we are that data points lie within a particular range. The standard error of the mean is an inferential statistic used to estimate the population mean and is a measure of the spread expected for the mean of the observations. Confidence intervals are often presented alongside sample results such as the mean value, indicating a range that is likely to contain the true value.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 28
Incorrect
-
A new medication aimed at preventing age-related macular degeneration (AMD) is being tested in clinical trials. One hundred patients over the age of 60 with early signs of AMD are given the new medication. Over a three month period, 10 of these patients experience progression of their AMD. In the control group, there are 300 patients over the age of 60 with early signs of AMD who are given a placebo. During the same time period, 50 of these patients experience progression of their AMD. What is the relative risk of AMD progression while taking the new medication?
Your Answer: 0.2
Correct Answer: 0.6
Explanation:The relative risk (RR) is calculated by dividing the exposure event rate (EER) by the control event rate (CER). In this case, the EER is 10 out of 100 (0.10) and the CER is 50 out of 300 (0.166). Therefore, the RR is calculated as 0.10 divided by 0.166, which equals 0.6.
Measures of Effect in Clinical Studies
When conducting clinical studies, we often want to know the effect of treatments of exposures on health outcomes. Measures of effect are used in randomized controlled trials (RCTs) and include the odds ratio (of), risk ratio (RR), risk difference (RD), and number needed to treat (NNT). Dichotomous (binary) outcome data are common in clinical trials, where the outcome for each participant is one of two possibilities, such as dead of alive, of clinical improvement of no improvement.
To understand the difference between of and RR, it’s important to know the difference between risks and odds. Risk is a proportion that describes the probability of a health outcome occurring, while odds is a ratio that compares the probability of an event occurring to the probability of it not occurring. Absolute risk is the basic risk, while risk difference is the difference between the absolute risk of an event in the intervention group and the absolute risk in the control group. Relative risk is the ratio of risk in the intervention group to the risk in the control group.
The number needed to treat (NNT) is the number of patients who need to be treated for one to benefit. Odds are calculated by dividing the number of times an event happens by the number of times it does not happen. The odds ratio is the odds of an outcome given a particular exposure versus the odds of an outcome in the absence of the exposure. It is commonly used in case-control studies and can also be used in cross-sectional and cohort study designs. An odds ratio of 1 indicates no difference in risk between the two groups, while an odds ratio >1 indicates an increased risk and an odds ratio <1 indicates a reduced risk.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 29
Correct
-
Which value of r indicates the highest degree of correlation?
Your Answer: -0.8
Explanation:It is important to distinguish between the direction of the correlation (the slope of the line) and its strength (the spread of the data). To emphasize this difference, the correct answer to this question is a negative value.
Stats: Correlation and Regression
Correlation and regression are related but not interchangeable terms. Correlation is used to test for association between variables, while regression is used to predict values of dependent variables from independent variables. Correlation can be linear, non-linear, of non-existent, and can be strong, moderate, of weak. The strength of a linear relationship is measured by the correlation coefficient, which can be positive of negative and ranges from very weak to very strong. However, the interpretation of a correlation coefficient depends on the context and purposes. Correlation can suggest association but cannot prove of disprove causation. Linear regression, on the other hand, can be used to predict how much one variable changes when a second variable is changed. Scatter graphs are used in correlation and regression analyses to visually determine if variables are associated and to detect outliers. When constructing a scatter graph, the dependent variable is typically placed on the vertical axis and the independent variable on the horizontal axis.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 30
Incorrect
-
The regional Health Authority has requested your expertise in determining whether to establish a new 12 bed pediatric ward of a six bed adolescent psychiatric unit. Your task is to conduct an economic analysis that evaluates the financial advantages and disadvantages of both proposals.
Your Answer: Cost utility analysis
Correct Answer: Cost benefit analysis
Explanation:A cost benefit analysis is a method of evaluating whether the benefits of an intervention outweigh its costs, using monetary units as the common measurement. Typically, this type of analysis is employed by funding bodies to make decisions about financing, such as whether to allocate resources for a new delivery suite of electroconvulsive therapy suite.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 31
Incorrect
-
If you anticipate that a drug will result in more side-effects than a placebo, what would be your estimated relative risk of side-effects occurring in the group receiving the drug?
Your Answer: <1
Correct Answer: >1
Explanation:Disease Rates and Their Interpretation
Disease rates are a measure of the occurrence of a disease in a population. They are used to establish causation, monitor interventions, and measure the impact of exposure on disease rates. The attributable risk is the difference in the rate of disease between the exposed and unexposed groups. It tells us what proportion of deaths in the exposed group were due to the exposure. The relative risk is the risk of an event relative to exposure. It is calculated by dividing the rate of disease in the exposed group by the rate of disease in the unexposed group. A relative risk of 1 means there is no difference between the two groups. A relative risk of <1 means that the event is less likely to occur in the exposed group, while a relative risk of >1 means that the event is more likely to occur in the exposed group. The population attributable risk is the reduction in incidence that would be observed if the population were entirely unexposed. It can be calculated by multiplying the attributable risk by the prevalence of exposure in the population. The attributable proportion is the proportion of the disease that would be eliminated in a population if its disease rate were reduced to that of the unexposed group.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 32
Incorrect
-
What type of bias is present in a study evaluating the accuracy of a new diagnostic test for epilepsy if not all patients undergo the established gold-standard test?
Your Answer: Instrument bias
Correct Answer: Work-up bias
Explanation:When comparing new diagnostic tests with gold standard tests, work-up bias can be a concern. Clinicians may be hesitant to order the gold standard test unless the new test yields a positive result, as the gold standard test may involve invasive procedures like tissue biopsy. This can significantly skew the study’s findings and affect metrics such as sensitivity and specificity. While it may not always be possible to eliminate work-up bias, researchers must account for it in their analysis.
Types of Bias in Statistics
Bias is a systematic error that can lead to incorrect conclusions. Confounding factors are variables that are associated with both the outcome and the exposure but have no causative role. Confounding can be addressed in the design and analysis stage of a study. The main method of controlling confounding in the analysis phase is stratification analysis. The main methods used in the design stage are matching, randomization, and restriction of participants.
There are two main types of bias: selection bias and information bias. Selection bias occurs when the selected sample is not a representative sample of the reference population. Disease spectrum bias, self-selection bias, participation bias, incidence-prevalence bias, exclusion bias, publication of dissemination bias, citation bias, and Berkson’s bias are all subtypes of selection bias. Information bias occurs when gathered information about exposure, outcome, of both is not correct and there was an error in measurement. Detection bias, recall bias, lead time bias, interviewer/observer bias, verification and work-up bias, Hawthorne effect, and ecological fallacy are all subtypes of information bias.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 33
Incorrect
-
What is the term used to describe a graph that can be utilized to identify publication bias?
Your Answer: Forest plot
Correct Answer: Funnel plot
Explanation:Stats Publication Bias
Publication bias refers to the tendency for studies with positive findings to be published more than studies with negative findings, leading to incomplete data sets in meta-analyses and erroneous conclusions. Graphical methods such as funnel plots, Galbraith plots, ordered forest plots, and normal quantile plots can be used to detect publication bias. Funnel plots are the most commonly used and offer an easy visual way to ensure that published literature is evenly weighted. The x-axis represents the effect size, and the y-axis represents the study size. A symmetrical, inverted funnel shape indicates that publication bias is unlikely, while an asymmetrical funnel indicates a relationship between treatment effect and study size, indicating either publication bias of small study effects.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 34
Incorrect
-
What is the accurate formula for determining the likelihood ratio of a negative test result?
Your Answer: Specificity / (sensitivity - 1)
Correct Answer: (1 - sensitivity) / specificity
Explanation:Clinical tests are used to determine the presence of absence of a disease of condition. To interpret test results, it is important to have a working knowledge of statistics used to describe them. Two by two tables are commonly used to calculate test statistics such as sensitivity and specificity. Sensitivity refers to the proportion of people with a condition that the test correctly identifies, while specificity refers to the proportion of people without a condition that the test correctly identifies. Accuracy tells us how closely a test measures to its true value, while predictive values help us understand the likelihood of having a disease based on a positive of negative test result. Likelihood ratios combine sensitivity and specificity into a single figure that can refine our estimation of the probability of a disease being present. Pre and post-test odds and probabilities can also be calculated to better understand the likelihood of having a disease before and after a test is carried out. Fagan’s nomogram is a useful tool for calculating post-test probabilities.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 35
Incorrect
-
A team of scientists embarked on a research project to determine if a new vaccine is effective in preventing a certain disease. They sought to satisfy the criteria outlined by Hill's guidelines for establishing causality.
What is the primary criterion among Hill's guidelines for establishing causality?Your Answer: Consistency
Correct Answer: Temporality
Explanation:The most crucial factor in Hill’s criteria for causation is temporality, of the temporal relationship between exposure and outcome. It is imperative that the exposure to a potential causal factor, such as factor ‘A’, always occurs before the onset of the disease. This criterion is the only absolute requirement for causation. The other criteria include the strength of the relationship, dose-response relationship, consistency, plausibility, consideration of alternative explanations, experimental evidence, specificity, and coherence.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 36
Incorrect
-
What is a true statement about searching in PubMed?
Your Answer: In PubMed, Boolean operators must be entered in lowercase letters
Correct Answer: Truncation is generally not a recommended search technique for PubMed
Explanation:Evidence-based medicine involves four basic steps: developing a focused clinical question, searching for the best evidence, critically appraising the evidence, and applying the evidence and evaluating the outcome. When developing a question, it is important to understand the difference between background and foreground questions. Background questions are general questions about conditions, illnesses, syndromes, and pathophysiology, while foreground questions are more often about issues of care. The PICO system is often used to define the components of a foreground question: patient group of interest, intervention of interest, comparison, and primary outcome.
When searching for evidence, it is important to have a basic understanding of the types of evidence and sources of information. Scientific literature is divided into two basic categories: primary (empirical research) and secondary (interpretation and analysis of primary sources). Unfiltered sources are large databases of articles that have not been pre-screened for quality, while filtered resources summarize and appraise evidence from several studies.
There are several databases and search engines that can be used to search for evidence, including Medline and PubMed, Embase, the Cochrane Library, PsycINFO, CINAHL, and OpenGrey. Boolean logic can be used to combine search terms in PubMed, and phrase searching and truncation can also be used. Medical Subject Headings (MeSH) are used by indexers to describe articles for MEDLINE records, and the MeSH Database is like a thesaurus that enables exploration of this vocabulary.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 37
Correct
-
How many people need to be treated with the new drug to prevent one case of Alzheimer's disease in individuals with a positive family history, based on the results of a randomised controlled trial with 1,000 people in group A taking the drug and 1,400 people in group B taking a placebo, where the Alzheimer's rate was 2% in group A and 4% in group B?
Your Answer: 50
Explanation:Measures of Effect in Clinical Studies
When conducting clinical studies, we often want to know the effect of treatments of exposures on health outcomes. Measures of effect are used in randomized controlled trials (RCTs) and include the odds ratio (of), risk ratio (RR), risk difference (RD), and number needed to treat (NNT). Dichotomous (binary) outcome data are common in clinical trials, where the outcome for each participant is one of two possibilities, such as dead of alive, of clinical improvement of no improvement.
To understand the difference between of and RR, it’s important to know the difference between risks and odds. Risk is a proportion that describes the probability of a health outcome occurring, while odds is a ratio that compares the probability of an event occurring to the probability of it not occurring. Absolute risk is the basic risk, while risk difference is the difference between the absolute risk of an event in the intervention group and the absolute risk in the control group. Relative risk is the ratio of risk in the intervention group to the risk in the control group.
The number needed to treat (NNT) is the number of patients who need to be treated for one to benefit. Odds are calculated by dividing the number of times an event happens by the number of times it does not happen. The odds ratio is the odds of an outcome given a particular exposure versus the odds of an outcome in the absence of the exposure. It is commonly used in case-control studies and can also be used in cross-sectional and cohort study designs. An odds ratio of 1 indicates no difference in risk between the two groups, while an odds ratio >1 indicates an increased risk and an odds ratio <1 indicates a reduced risk.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 38
Incorrect
-
You record the age of all of your students in your class. You notice that your data set is skewed. What method would you use to describe the typical age of your students?
Your Answer: Interquartile range
Correct Answer: Median
Explanation:When dealing with a data set that is quantitative and measured on a ratio scale, the mean is typically the preferred measure of central tendency. However, if the data is skewed, the median may be a better choice as it is less affected by the skewness of the data.
Measures of Central Tendency
Measures of central tendency are used in descriptive statistics to summarize the middle of typical value of a data set. There are three common measures of central tendency: the mean, median, and mode.
The median is the middle value in a data set that has been arranged in numerical order. It is not affected by outliers and is used for ordinal data. The mode is the most frequent value in a data set and is used for categorical data. The mean is calculated by adding all the values in a data set and dividing by the number of values. It is sensitive to outliers and is used for interval and ratio data.
The appropriate measure of central tendency depends on the measurement scale of the data. For nominal and categorical data, the mode is used. For ordinal data, the median of mode is used. For interval data with a normal distribution, the mean is preferable, but the median of mode can also be used. For interval data with skewed distribution, the median is used. For ratio data, the mean is preferable, but the median of mode can also be used for skewed data.
In addition to measures of central tendency, the range is also used to describe the spread of a data set. It is calculated by subtracting the smallest value from the largest value.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 39
Incorrect
-
What type of evidence is considered the most robust and reliable?
Your Answer: Randomised controlled trial
Correct Answer: Meta-analysis
Explanation:Levels and Grades of Evidence in Evidence-Based Medicine
To evaluate the quality of evidence on a subject of question, levels of grades are used. The traditional hierarchy approach places systematic reviews of randomized control trials at the top and case-series/report at the bottom. However, this approach is overly simplistic as certain research questions cannot be answered using RCTs. To address this, the Oxford Centre for Evidence-Based Medicine introduced their 2011 Levels of Evidence system, which separates the type of study questions and gives a hierarchy for each.
The grading approach to be aware of is the GRADE system, which classifies the quality of evidence as high, moderate, low, of very low. The process begins by formulating a study question and identifying specific outcomes. Outcomes are then graded as critical of important. The evidence is then gathered and criteria are used to grade the evidence, with the type of evidence being a significant factor. Evidence can be promoted of downgraded based on certain criteria, such as limitations to study quality, inconsistency, uncertainty about directness, imprecise of sparse data, and reporting bias. The GRADE system allows for the promotion of observational studies to high-quality evidence under the right circumstances.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 40
Incorrect
-
A cohort study of 10,000 elderly individuals aimed to determine whether regular exercise has an effect on cognitive decline. Half of the participants engaged in regular exercise while the other half did not.
What is a limitation of conducting a cohort study in this scenario?Your Answer: They can only provide information about one outcome
Correct Answer: When the outcome of interest is rare a very large sample size is needed
Explanation:Cohort studies involve following a group of individuals over a period of time to investigate whether exposure to a particular factor affects disease incidence. Although they are costly and time-consuming, they offer several benefits. For instance, they can examine rare exposure factors and are less prone to recall bias than case-control studies. Additionally, they can measure disease incidence and risk. Results are typically presented as the relative risk of developing the disease due to exposure to the factor.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 41
Correct
-
Which type of evidence is typically regarded as the most reliable according to traditional methods?
Your Answer: RCTs with non-definitive results
Explanation:Levels and Grades of Evidence in Evidence-Based Medicine
To evaluate the quality of evidence on a subject of question, levels of grades are used. The traditional hierarchy approach places systematic reviews of randomized control trials at the top and case-series/report at the bottom. However, this approach is overly simplistic as certain research questions cannot be answered using RCTs. To address this, the Oxford Centre for Evidence-Based Medicine introduced their 2011 Levels of Evidence system, which separates the type of study questions and gives a hierarchy for each.
The grading approach to be aware of is the GRADE system, which classifies the quality of evidence as high, moderate, low, of very low. The process begins by formulating a study question and identifying specific outcomes. Outcomes are then graded as critical of important. The evidence is then gathered and criteria are used to grade the evidence, with the type of evidence being a significant factor. Evidence can be promoted of downgraded based on certain criteria, such as limitations to study quality, inconsistency, uncertainty about directness, imprecise of sparse data, and reporting bias. The GRADE system allows for the promotion of observational studies to high-quality evidence under the right circumstances.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 42
Incorrect
-
What is the term used to describe a test that initially appears to measure what it is intended to measure?
Your Answer: Good internal validity
Correct Answer: Good face validity
Explanation:A test that seems to measure what it is intended to measure has strong face validity.
Validity in statistics refers to how accurately something measures what it claims to measure. There are two main types of validity: internal and external. Internal validity refers to the confidence we have in the cause and effect relationship in a study, while external validity refers to the degree to which the conclusions of a study can be applied to other people, places, and times. There are various threats to both internal and external validity, such as sampling, measurement instrument obtrusiveness, and reactive effects of setting. Additionally, there are several subtypes of validity, including face validity, content validity, criterion validity, and construct validity. Each subtype has its own specific focus and methods for testing validity.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 43
Incorrect
-
What resource is committed to offering complete articles of systematic reviews on the impacts of healthcare interventions?
Your Answer: CINAHL
Correct Answer: CDSR
Explanation:When faced with a question, it’s helpful to consider what the letters in the question might represent, even if you don’t know the answer right away. Don’t become overwhelmed and keep this strategy in mind.
Evidence-based medicine involves four basic steps: developing a focused clinical question, searching for the best evidence, critically appraising the evidence, and applying the evidence and evaluating the outcome. When developing a question, it is important to understand the difference between background and foreground questions. Background questions are general questions about conditions, illnesses, syndromes, and pathophysiology, while foreground questions are more often about issues of care. The PICO system is often used to define the components of a foreground question: patient group of interest, intervention of interest, comparison, and primary outcome.
When searching for evidence, it is important to have a basic understanding of the types of evidence and sources of information. Scientific literature is divided into two basic categories: primary (empirical research) and secondary (interpretation and analysis of primary sources). Unfiltered sources are large databases of articles that have not been pre-screened for quality, while filtered resources summarize and appraise evidence from several studies.
There are several databases and search engines that can be used to search for evidence, including Medline and PubMed, Embase, the Cochrane Library, PsycINFO, CINAHL, and OpenGrey. Boolean logic can be used to combine search terms in PubMed, and phrase searching and truncation can also be used. Medical Subject Headings (MeSH) are used by indexers to describe articles for MEDLINE records, and the MeSH Database is like a thesaurus that enables exploration of this vocabulary.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 44
Incorrect
-
What test would be the most effective in verifying the suitability of using a parametric test on a given dataset?
Your Answer: Kendall test
Correct Answer: Lilliefors test
Explanation:Normality Testing in Statistics
In statistics, parametric tests are based on the assumption that the data set follows a normal distribution. On the other hand, non-parametric tests do not require this assumption but are less powerful. To check if a distribution is normally distributed, there are several tests available, including the Kolmogorov-Smirnov (Goodness-of-Fit) Test, Jarque-Bera test, Wilk-Shapiro test, P-plot, and Q-plot. However, it is important to note that if a data set is not normally distributed, it may be possible to transform it to make it follow a normal distribution, such as by taking the logarithm of the values.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 45
Incorrect
-
What is the term used to describe the likelihood of correctly rejecting the null hypothesis when it is actually false?
Your Answer: Beta level
Correct Answer: Power of the test
Explanation:Understanding Hypothesis Testing in Statistics
In statistics, it is not feasible to investigate hypotheses on entire populations. Therefore, researchers take samples and use them to make estimates about the population they are drawn from. However, this leads to uncertainty as there is no guarantee that the sample taken will be truly representative of the population, resulting in potential errors. Statistical hypothesis testing is the process used to determine if claims from samples to populations can be made and with what certainty.
The null hypothesis (Ho) is the claim that there is no real difference between two groups, while the alternative hypothesis (H1 of Ha) suggests that any difference is due to some non-random chance. The alternative hypothesis can be one-tailed of two-tailed, depending on whether it seeks to establish a difference of a change in one direction.
Two types of errors may occur when testing the null hypothesis: Type I and Type II errors. Type I error occurs when the null hypothesis is rejected when it is true, while Type II error occurs when the null hypothesis is accepted when it is false. The power of a study is the probability of correctly rejecting the null hypothesis when it is false, and it can be increased by increasing the sample size.
P-values provide information on statistical significance and help researchers decide if study results have occurred due to chance. The p-value is the probability of obtaining a result that is as large of larger when in reality there is no difference between two groups. The cutoff for the p-value is called the significance level (alpha level), typically set at 0.05. If the p-value is less than the cutoff, the null hypothesis is rejected, and if it is greater or equal to the cut off, the null hypothesis is not rejected. However, the p-value does not indicate clinical significance, which may be too small to be meaningful.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 46
Incorrect
-
A study is being planned to investigate whether exposure to pesticides is a risk factor for Parkinson's disease. The researchers are considering conducting a case-control study instead of a cohort study. What is one advantage of using a case-control study design in this situation?
Your Answer: It is expensive to perform
Correct Answer: It is possible to study diseases that are rare
Explanation:The benefits of conducting a case-control study include its suitability for examining rare diseases, the ability to investigate a broad range of risk factors, no loss to follow-up, and its relatively low cost and quick turnaround time. The findings of such studies are typically presented as an odds ratio.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 47
Incorrect
-
Which of the following methods is most effective in eliminating of managing confounding factors?
Your Answer: Stratification
Correct Answer: Randomisation
Explanation:The most effective way to eliminate of manage potential confounding factors is to randomize a large enough sample size. This approach addresses all potential confounders, regardless of whether they were measured in the study design. Matching involves pairing individuals who received a treatment of intervention with non-treated individuals who have similar observable characteristics. Post-hoc methods, such as stratification, regression analysis, and analysis of variance, can be used to evaluate the impact of known or suspected confounders.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 48
Incorrect
-
In scientific research, what variable type has traditionally been used to record the age of study participants?
Your Answer: Categorical
Correct Answer: Binary
Explanation:Gender has traditionally been recorded as either male of female, creating a binary of dichotomous variable. Other categorical variables, such as eye color and ethnicity, can be grouped into two or more categories. Continuous variables, such as temperature, height, weight, and age, can be placed anywhere on a scale and have mathematical properties. Ordinal variables allow for ranking, but do not allow for direct mathematical comparisons between values.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 49
Incorrect
-
What is another term used to refer to Neyman bias?
Your Answer:
Correct Answer: Prevalence/incidence bias
Explanation:Neyman bias arises when a research study is examining a condition that is marked by either undetected cases of cases that result in early deaths, leading to the exclusion of such cases from the analysis.
Types of Bias in Statistics
Bias is a systematic error that can lead to incorrect conclusions. Confounding factors are variables that are associated with both the outcome and the exposure but have no causative role. Confounding can be addressed in the design and analysis stage of a study. The main method of controlling confounding in the analysis phase is stratification analysis. The main methods used in the design stage are matching, randomization, and restriction of participants.
There are two main types of bias: selection bias and information bias. Selection bias occurs when the selected sample is not a representative sample of the reference population. Disease spectrum bias, self-selection bias, participation bias, incidence-prevalence bias, exclusion bias, publication of dissemination bias, citation bias, and Berkson’s bias are all subtypes of selection bias. Information bias occurs when gathered information about exposure, outcome, of both is not correct and there was an error in measurement. Detection bias, recall bias, lead time bias, interviewer/observer bias, verification and work-up bias, Hawthorne effect, and ecological fallacy are all subtypes of information bias.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
-
Question 50
Incorrect
-
A team of researchers aim to explore the opinions of pediatricians who specialize in treating children with asthma. They begin by visiting a local pediatric clinic and speaking with a doctor who has expertise in this area. They then ask this doctor to suggest another pediatrician who specializes in treating children with asthma whom they could interview. They continue this process until they have spoken with all the recommended pediatricians.
Which sampling technique are they employing?Your Answer:
Correct Answer: Snowball
Explanation:Snowball sampling is a unique technique utilized in qualitative research when the desired sample trait is uncommon. In such cases, it can be challenging of expensive to locate suitable respondents. Snowball sampling involves existing subjects recruiting future subjects, which can help overcome these difficulties. For more information on this method, please refer to the additional resources provided.
-
This question is part of the following fields:
- Research Methods, Statistics, Critical Review And Evidence-Based Practice
-
00
Correct
00
Incorrect
00
:
00
:
00
Session Time
00
:
00
Average Question Time (
Mins)