How We Score Studies

Our scoring system evaluates scientific research on four key dimensions, then adjusts for study type to give you a clear picture of evidence quality.

The Four Pillars of Quality

Every study is evaluated on four weighted components that capture different aspects of scientific rigor:

๐Ÿ”ฌ

Methodology

35% weight

Evaluates study design: randomization, blinding, control groups, sample size, and follow-up duration.

๐Ÿ“Š

Statistical Rigor

25% weight

Checks for p-values, effect sizes, confidence intervals, and pre-registration of the study.

๐Ÿ“‹

Reporting Transparency

20% weight

Assesses conflict of interest disclosure, data availability, and code sharing practices.

๐Ÿ“ฐ

Publication Quality

20% weight

Baseline score for peer-reviewed publication status and journal reputation factors.

What We Look For

Methodology Factors

Randomization
Up to 20 points

Full randomization earns maximum points; controlled but non-randomized designs earn partial credit

Blinding
Up to 15 points

Double-blind studies earn full points; single-blind earn 60%

Control Group
15 points

Presence of a proper control group for comparison

Sample Size
Up to 20 points

Uses exponential decay formulaโ€”larger samples score higher, with diminishing returns

Follow-up Duration
10 points

Studies with documented follow-up periods

Statistical Analysis Factors

Effect Size Reported
20 points

Clear reporting of the magnitude of findings

Confidence Intervals
15 points

Statistical uncertainty properly quantified

P-values Reported
15 points

Statistical significance testing included

Pre-registration
15 points

Study protocol registered before data collection

Transparency Factors

Conflict of Interest Disclosed
40 points

Authors clearly state funding sources and potential conflicts

Data Available
35 points

Raw data accessible for verification

Code Available
25 points

Analysis code shared for reproducibility

Study Type Adjustments

After calculating the base score, we apply a multiplier based on study type. This reflects that not all evidence is equally applicable to human health decisions.

1.0xHuman Studies

Clinical trials and observational studies in humans

0.9xNarrative Reviews

Literature reviews without systematic methodology

0.4xAnimal Studies (Oncology)

Cancer research in animal models

0.3xAnimal Studies (Neurology)

Brain and nervous system research in animals

0.3xAnimal Studies (General)

Other animal model research

0.2xIn Vitro (Molecular)

Cell and molecular level experiments

0.1xIn Vitro (General)

Lab dish experiments

0.1xOpinion/Editorial

Expert opinions without primary research

Why the difference? A perfectly designed mouse study might score 100 on methodology, but that doesn't mean the findings apply to humans. The multiplier helps you understand how directly the evidence relates to human health outcomes.

Interpreting Final Scores

70-100High Quality

Strong methodology, good statistical reporting, and transparent practices. Human studies in this range represent solid evidence.

40-69Moderate Quality

Some limitations in design or reporting. Evidence should be considered alongside other studies. Common for animal research.

20-39Low Quality

Significant methodological issues or very preliminary research. Should not be used alone to make health decisions.

0-19Very Low

Opinions, editorials, or studies with major flaws. Useful for generating hypotheses but not for drawing conclusions.

Our Commitment

๐ŸŽฏ

Objective Criteria

Every factor in our scoring system is based on established scientific quality indicators. No subjective judgments about whether we "like" the findings.

๐Ÿ”„

Consistent Application

The same algorithm scores every study. A nutrition study and a pharmacology study are evaluated by identical criteria.

๐Ÿ“–

Full Transparency

Every analyzed study shows the complete score breakdown. You can see exactly why a study received its score.