Sunn Skepsis
Denne portalen er ment å gi deg som pasient råd om kvalitetskriterier for helseinformasjon og tilgang til forskningsbasert informasjon.
| 0 CommentsDancing statistics: correlation
A 4-minute film demonstrating the statistical concept of correlation through dance.
| 0 CommentsHow can you know if the spoon works?
Short, small group exercise on how to design a fair comparison using the "claim" that a spoon helps retain the bubbles in champagne.
| 0 CommentsDRUG TOO
James McCormick with another parody/spoof of the Cee Lo Green song ‘Forget You’ to prompt scepticism about many drug treatments.
| 0 CommentsCalling Bullshit Syllabus
Carl Bergstrom's and Jevin West's nice syllabus for 'Calling Bullshit'.
| 0 Comments‘Tricks to help you get the result you want from your study (S4BE)
Inspired by a chapter in Ben Goldacre’s ‘Bad Science’, medical student Sam Marks shows you how to fiddle research results.
| 0 CommentsIt’s just a phase
A resource explaining the differences between different trial phases.
| 0 CommentsStrictly Cochrane: a quickstep around research and systematic reviews
An interactive resource explaining how systematic and non-systematic reviews differ, and the importance of keeping reviews up to date.
| 0 CommentsThe Princess and the p-value
An interactive resource introducing reporting and interpretation of statistics in controlled trials.
| 0 CommentsExplaining the mission of the AllTrials Campaign (TED talk)
Half the clinical trials of medicines we use haven’t been published. Síle Lane shows how the AllTrials Campaign is addressing this scandal.
| 0 CommentsFast Stats to explain absolute risk, relative risk and Number Needed to Treat (NNT).
A 15-slide presentation on ‘Fast Stats’ to explain absolute risk, relative risk and Number Needed to Treat (NNT) prepared by PharmedOut.
| 0 CommentsUnsubstantiated and overstated claims of efficacy
A 32-slide presentation on misleading advertisements and FDA warnings prepared by PharmedOut.
| 0 CommentsCritical appraisal
University of New South Wales Medical Statistics Tutorial 4 addresses Critical Appraisal.
| 0 CommentsProbability and tests of statistical significance
University of New South Wales Medical Statistics Tutorial 6 addresses ‘Probability and tests of statistical significance’.
| 0 CommentsBias – the biggest enemy
University of New South Wales Medical Stats Online Tutorial 5 addresses ‘Bias - the biggest enemy’.
| 0 CommentsIntroduction to Evidence-Based Medicine
Bill Caley’s 26 slides with notes used as an ‘Introduction to Evidence-Based Medicine’.
| 0 CommentsApplying evidence to patients
A 27-minute talk on ‘Applying Evidence to Patients’, illustrated by 17 slides, with notes.
| 0 Comments2×2 tables and relative risk
A 10-min talk on ‘2x2 tables and Relative Risk’, illustrated by 14 slides, with notes.
| 0 CommentsAppraisal of evidence and interpretation of results
A 14-min talk on ‘Appraisal of the Evidence and Interpretation of the Results’, illustrated by 19 slides, with notes.
| 0 CommentsBasic principles of randomised trials, and validity
A 8-min talk on ‘Basic principles of Randomised Trials, and Validity’, illustrated by 15 slides, with notes.
| 0 CommentsDefining clinical questions
An 8-min talk on ‘Defining Clinical Questions’ illustrated by 10 slides, with notes.
| 0 CommentsA way to teach about systematic reviews
81 slides used by David Nunan (Centre for Evidence-Based Medicine, Oxford) to present ‘A way to teach about systematic reviews’.
| 0 CommentsApplying the evidence
Six key slides produced by the University of Western Australia on applying evidence in practice.
| 0 CommentsAppraising the evidence
Six key slides produced by the University of Western Australia to introduce critical appraisal.
| 0 CommentsTaking account of the play of chance
Differences in outcome events in treatment comparisons may reflect only the play of chance. Increased numbers of events reduces this problem
| 0 CommentsQuantifying uncertainty in treatment comparisons
Small studies in which few outcome events occur are usually not informative and the results are sometimes seriously misleading.
| 0 CommentsBringing it all together for the benefit of patients and the public
Improving reports of research and up-to-date systematic reviews of reliable studies are essential foundations of effective health care.
| 0 CommentsApplying the results of trials and systematic reviews to individual patients
Paul Glasziou uses 28 slides to address ‘Applying the results of trials and systematic reviews to individual patients’.
| 0 Comments10 Components of effective clinical epidemiology: How to get started
PDF & Podcast of 1-hr talk by Carl Heneghan (Centre for Evidence-Based Medicine, Oxford) on effective clinical epidemiology.
| 0 CommentsCritical appraisal of clinical trials
Slides developed by Amanda Burls for an interactive presentation covering the most important features of well controlled trials.
| 0 CommentsExplaining the unbiased creation of treatment comparison groups and blinded outcome assessment
A class were given coloured sweets and asked to design an experiment to find out whether red sweets helped children to think more quickly.
| 0 CommentsSystematic Reviews and Meta-analysis: Information Overload
None of us can keep up with the sheer volume of material published in medical journals each week.
| 0 CommentsCombining the Results from Clinical Trials
Chris Cates notes that emphasizing the results of patients in particular sub-groups in a trial can be misleading.
| 0 CommentsGenerationR – The importance of involving children and young people in research
3/3, 22-min video at the launch of GenerationR, a network of young people who advise researchers.
| 0 CommentsGeneration R – The importance of medical research in children and young people
2/3, 35-min video at the launch of GenerationR, a network of young people who advise researchers.
| 0 CommentsGeneration R – The need to reduce waste in clinical research involving children
1/3, 14-min video at the launch of GenerationR, a network of young people who advise researchers.
| 0 CommentsNo Power, No Evidence!
This blog explains that studies need sufficient statistical power to detect a difference between groups being compared.
| 0 CommentsBeginners guide to interpreting odds ratios, confidence intervals and p values
A tutorial on interpreting odds ratios, confidence intervals and p-values, with questions to test the reader’s knowledge of each concept.
| 0 CommentsSample Size matters even more than you think
This blog explains why adequate sample sizes are important, and discusses research showing that sample size may affect effect size.
| 0 CommentsWhat is it with Odds and Risk?
This blog explains odds ratios and relative risks, and provides the formulae for calculating both measures.
| 0 CommentsPreclinical animal studies: bad experiments cost lives
This blog notes that few therapies that treat disease in animals successfully translate into effective treatments for humans.
| 0 CommentsSurrogate Endpoints in EBM: What are the benefits and dangers?
What are surrogate outcomes, their pros and cons, and why you should be cautious in extrapolating from them to clinical decisions.
| 0 CommentsThe Systematic Review
This blog explains what a systematic review is, the steps involved in carrying one out, and how the review should be structured.
| 0 CommentsThe Mean: Simply Average?
This blog explains ‘the mean’ as a measure of average; describes how to calculate it; and flags up some caveats.
| 0 CommentsPublication Bias: An Editorial Problem?
A blog challenging the idea that publication bias mainly occurs at editorial level, after research has been submitted for publication.
| 0 CommentsThe Bias of Language
Publication of research findings in a particular language may be prompted by the nature and direction of the results.
| 0 CommentsDefining Bias
This blog explains what is meant by ‘bias’ in research, focusing particularly on attrition bias and detection bias.
| 0 CommentsBalancing Benefits and harms
A blog explaining what is meant by ‘benefits’ and ‘harms’ in the context of healthcare interventions, and the importance of balancing them.
| 0 CommentsData Analysis Methods
A discussion of 2 approaches to data analysis in trials - ‘As Treated’, and ‘Intention-to-Treat’ - and some of the pros and cons of each.
| 0 CommentsDefining Risk
This blog defines ‘risk’ in relation to health, and discusses some the difficulties in applying estimates of risk to a given individual.
| 0 CommentsTraditional Reviews vs. Systematic Reviews
This blog outlines 11 differences between systematic and traditional reviews, and why systematic reviews are preferable.
| 0 CommentsP Value in Plain English
Using simple terms and examples, this blog explains what p-values mean in the context of testing hypotheses in research.
| 0 CommentsCancer Screening Debate
This blog discusses problems that can be associated with cancer screening, including over-diagnosis and thus (unnecessary) over-treatment.
| 0 CommentsSurrogate endpoints: pitfalls of easier questions
A blog explaining what surrogate endpoints are and why they should be interpreted cautiously.
| 0 CommentsMisconceptions about screening
Screening should not be for everyone or all diseases. It should only be offered when it is likely to do good than harm.
| 0 CommentsMaking sense of randomized trials
A description of how clinical trials are constructed and analysed to ensure they provide fair comparisons of treatments.
| 0 CommentsRandomized Control Trials
1/2, 40-min lecture on randomized trials by Dr R Ramakrishnan (Lecture 25) for the Central Coordinated Bioethics Programme in India.
| 0 CommentsCompliance with protocol and follow-up in clinical trials
Denis Black’s 10-min, downloadable, PowerPoint presentation on compliance, follow up, and intention-to-treat analysis in clinical trials.
| 0 CommentsClinical Significance – CASP
To understand results of a trial it is important to understand the question it was asking.
| 0 CommentsStatistical Significance – CASP
In a well-conducted randomized trial, the groups being compared should differ from each other only by chance and by the treatment received.
| 0 CommentsP Values – CASP
Statistical significance is usually assessed by appeal to a p-value, a probability, which can take any value between 0 and 1 (certain).
| 0 CommentsMaking sense of results – CASP
This module introduces the key concepts required to make sense of statistical information presented in research papers.
| 0 CommentsScreening – CASP
This module on screening has been designed to help people evaluate screening programmes.
| 0 CommentsRandomised Control Trials – CASP
This module looks at the critical appraisal of randomised trials.
| 0 CommentsCommon Sources of Bias
Bias (the conscious or unconscious influencing of a study and its results) can occur in different ways and renders studies less dependable.
| 0 CommentsTamiflu: securing access to medical research data
A campaign by researchers has shown that Roche spun the research on Tamiflu to meet their commercial ends.
| 0 CommentsLos intervalos de confianza en investigación
¿Para qué sirven los intervalos de confianza en los estudios de investigación?
| 0 CommentsWatson en busca de la evidencia
Cómic acerca de conflictos de intereses y búsqueda de información.
| 0 CommentsThe need to compare like-with-like in treatment comparisons
Allocation bias results when trials fail to ensure that, apart from the treatments being compared, ‘like will be compared with like'.
| 0 CommentsWhy avoiding differences between treatments allocated and treatments received is important
Knowledge of which treatments have been received by which study participants can affect adherence to assigned treatments and result in bias.
| 0 CommentsThe need to avoid differences in the way treatment outcomes are assessed
Biased treatment outcome assessment can result if people know which participants have received which treatments.
| 0 CommentsAvoiding biased selection from the available evidence
Systematic reviews are used to identify, evaluate and summarize all the evidence relevant to addressing a particular question.
| 0 CommentsPreparing and maintaining systematic reviews of all the relevant evidence
Unbiased, up-to-date systematic reviews of all the relevant, reliable evidence are needed to inform practice and policy.
| 0 CommentsDealing with biased reporting of the available evidence
Biased reporting of research occurs when the direction or statistical significance of results influences how research is reported.
| 0 CommentsUsing the results of up-to-date systematic reviews of research
Trustworthy evidence from research is necessary, but not sufficient, to improve the quality of health care.
| 0 CommentsWhy treatment comparisons must be fair
Fair treatment comparisons avoid biases and reduce the effects of the play of chance.
| 0 CommentsIntroduction to JLL Explanatory Essays
Professionals sometimes harm patients by using inadequately evaluated treatments. Research addressing uncertainties can reduce this harm.
| 0 CommentsAvoiding biased treatment comparisons
Biases in tests of treatments are those factors that can lead to conclusions that are systematically different from the truth.
| 0 CommentsBias introduced after looking at study results
Biases can be introduced when knowledge of the results of studies influences analysis and reporting decisions.
| 0 CommentsReducing biases in judging unanticipated effects of treatments
As with anticipated effects of treatments, biases and the play of chance must be reduced in assessing suspected unanticipated effects.
| 0 CommentsRecognizing researcher/sponsor biases and fraud
The vested interests of researchers and organizations tend to be reflected in reports of treatment research in which they are involved.
| 0 CommentsVideo games and health improvement: a literature review of randomized controlled trials
This is a critical appraisal of a non-systematic review of randomized trials of video games for improving health.
| 0 CommentsWhy comparisons must address genuine uncertainties
Too much research is done when there are no genuine uncertainties about treatment effects. This is unethical, unscientific, and wasteful.
| 0 CommentsInterpreting 95% Confidence Intervals
Gilbert Welch’s 9-min video on how 95% confidence intervals relate to p values.
| 0 CommentsWhy treatment comparisons are essential
Formal comparisons are required to assess treatment effects and to take account of the natural course of health problems.
| 0 CommentsWhy treatment uncertainties should be addressed
Ignoring uncertainties about the effects of treatments has led to avoidable suffering and deaths.
| 0 CommentsWhat are systematic reviews?
A 3-min video by Jack Nunn and The Cochrane Consumers and Communication group for people unfamiliar with the concept of systematic reviews.
| 0 CommentsInteractive PowerPoint Presentation about Clinical Trials
An interactive Powerpoint presentation for people thinking about participating in a clinical trial or interested in learning about them.
| 0 CommentsIntroduction to clinical trials: lay-friendly video
This lay-friendly video introduction to clinical trials was created by the European Communication on Research Awareness Needs Project.
| 3 CommentsOverview of study designs
11 slides and a 4-min commentary overviewing study designs for clinical research. (from Univ Mass Med School).
| 1 CommentSystematic Reviews
3 slides and a 4-min commentary about systematic reviews and meta-analyses (from Univ Mass Med School).
| 0 CommentsRelative risk, absolute risk and number-needed-to-treat
8 slides and a 4-min commentary on relative risk, absolute risk, and number needed to treat (from Univ Mass Med School).
| 0 CommentsAttrition bias, publication bias, comparator bias and commercial bias
6 slides and a 3-min commentary on attrition bias, publication bias, comparator bias and commercial bias (from Univ Mass Med School).
| 0 CommentsP-values and the role of chance
Gilbert Welch’s 10-min video on p-values and assessing the likelihood that a difference between treatments is due to chance.
| 0 CommentsUnderstanding Confidence Intervals
A 4-min video explaining the concept of confidence intervals and how they are calculated, with helpful diagrams and examples.
| 0 CommentsUnderstanding Overdiagnosis bias
Gilbert Welch’s 14-min video discussing the risks of overdiagnosis bias and screening.
| 0 CommentsCalculating and interpreting absolute and relative change in an unwanted outcome after treatment
Gilbert Welch’s 6-min video explaining how to calculate and interpret absolute and relative change in an unwanted outcome.
| 0 CommentsUnderstanding lead-time bias
Gilbert Welch’s 10-min video explaining why survival ALWAYS rises following early detection -- even if no one is helped.
| 0 CommentsTesting Treatments Audio Book
The Testing Treatments Audiobook enables visitors to the TTi site to select whichever chapters in the book they would like to listen to.
| 0 CommentsDouble blind studies
A webpage discussing the importance of blinding trial participants and researchers to intervention allocation.
| 0 CommentsCEBM – Study Designs
A short article explaining the relative strengths and weaknesses of different types of study design for assessing treatment effects.
| 0 CommentsIntroduction to Critical Appraisal
30-slide introduction by Jason Curtis, to Critical Appraisal.
| 0 CommentsDISCERN online
A questionnaire providing a valid and reliable way of assessing the quality of written information on treatment choices.
| 0 CommentsMeans vs. Medians
Keith Bower’s 3-min video explaining how means (averages) and medians can be presented misleadingly.
| 0 CommentsRandomized Controlled Trial Protocols
A 1-hour videoed lecture explaining protocols for Randomized Control Trials (RCTs).
| 0 CommentsJohn Ioannidis, the scourge of sloppy science
A 8 min podcast interview with John Ioannidis explaining how research claims can be misleading.
| 0 CommentsScience Weekly Podcast – Ben Goldacre
A 1-hour audio interview with Ben Goldacre discussing misleading claims about research.
| 0 CommentsStatistical Significance and Practical Significance
Keith Bower’s 3-min video discussing the difference between Statistical Significance and Practical Significance.
| 0 CommentsSignals of overdiagnosis
Gilbert Welch’s 8-min video showing how population screening for disease leads to overdiagnosis.
| 0 CommentsRandomised Controlled Trials vs. Observational Studies
5-minute video explaining the difference between randomised trials and observational studies.
| 0 CommentsType I and Type II errors, and how statistical tests can be misleading
Gilbert Welch’s 12-min video explaining Type I and Type II errors, and how statistical tests can be misleading.
| 0 CommentsHow do you know which healthcare research you can trust?
A detailed guide to study design, with learning objectives, explaining some sources of bias in health studies.
| 0 CommentsCASP: making sense of evidence
The Critical Appraisal Skills Programme (CASP) website with resources for teaching critical appraisal.
| 0 CommentsSmart Health Choices: making sense of health advice
The Smart Health Choices e-book explains how to make informed health decisions.
| 0 CommentsMethodology of clinical trials
Eurordis training on the methodology of clinical trials for representatives of patients’ organisations.
| 0 CommentsCan measurements show if a treatment works?
An article discussing errors to avoid when testing treatments.
| 0 CommentsMore than average confusion about what mean means mean
Cartoon and blog about how averages can hide important variations in effects.
| 0 CommentsIn defence of systematic reviews of small trials
An article discussing the strengths and weaknesses of systematic reviews of small trials.
| 0 CommentsRelative or absolute measures of effects
Dr Chris Cates' article explaining absolute and relative effects of treatment effects.
| 0 CommentsEvidence from Randomised Trials and Systematic Reviews
Dr Chris Cates' article discussing control of bias in randomised trials and explaining systematic reviews.
| 0 CommentsThe perils and pitfalls of subgroup analysis
Dr Chris Cates' article demonstrating why subgroup analysis can be untrustworthy.
| 0 Comments