How We Work | Lead Stories

How we find claims and stories to fact check

Lead Stories looks for claims and stories to fact check using several tools and methods, here are the most important ones:

  • Our own patented Trendolizer™ engine
  • Google Trends
  • Meta's tool for Fact Checkers
  • ByteDance's tool for fact checkers & direct searches on TikTok
  • X Community Notes
  • Reader tips

Trendolizer

Lead Stories created the Trendolizer™ engine to monitor the internet and look for newly trending content. It can measure the engagement rate (likes, views, comments, retweets etc.) of links, images and videos appearing on various platforms (Facebook, Twitter, YouTube, TikTok...) and gives us an overview of what is currently going viral.

screenshot.jpg

(screenshot from our Trendolizer™ dashboard)

We use Trendolizer to monitor the internet for viral content in general but also look at popular content from specific sites that have posted false information in the past, along with viral content matching keywords and phrases associated with false narratives or conspiracy theories.

Google Trends

We use Google Trends daily to check which search queries related to fact checking phrases ("fake news", "fact check" etc.) are popular on Google. This usually indicates people are searching for answers about a particular story because they have doubts about the veracity.

Meta's tool for fact checkers

Lead Stories is part of Meta's Partnership with Third Party Fact Checkers and as such for some countries we get access to lists of content that has been flagged as potential misinformation (via a custom tool for fact checkers provided by Meta) . Note that Lead Stories is free to pick items from that list or to ignore them, Meta does not tell us which stories we should do.

ByteDance's tool for fact checkers & direct searches on TikTok

Lead Stories is part of Bytedance's Partnership with Third Party Fact Checkers and as such for some countries we get access to lists of content that has been flagged as potential misinformation (via a custom tool for fact checkers provided by ByteDance) . Note that for the purpose of publishing fact checks Lead Stories is free to pick items from that list or to ignore them, Bytedance does not tell us which stories we should do. We also perform direct searches of TikTok around current events and misinformation-prone topics to look for viral misinformation to fact check.

X Community Notes

Several members of our team are part of X's Community Notes program where X users can add notes with context to posts which then may or may not become visible after enough users upvote them. We sometimes use the queue of notes waiting for upvotes to get an early indication about newly trending misinformation to investigate.

Reader Tips

Readers can always contact us via our contact page if they have a tip about an online story that is potentially false and harmful. In practice we get very few tips like this but we would love to get more.


How we select claims and stories to fact check

On a typical day we will look for the most trending stories/videos/posts/... with these tools and pick the ones that meet following criteria:

  • They make a claim that can be checked (i.e. no pure opinions, predictions, vague statements or prophecies). Note we include claims that could be checked theoretically with access to the right evidence or witnesses but where this very hard or impossible in practice (quotes from closed meetings, insider information...).
  • They are harmful to someone. We interpret "harmful" quite broadly here, it can include anything where there is a risk of: negative health impacts, reputational damage, political decisions based on inaccurate information, minimizing or exaggerating real problems, emotional impact over fictional stories presented as real, fraudsters profiting from spreading false info, financial loss, distorted understanding of current events, people getting wrongly blamed for things, false evidence being used to support larger conspiracy theories...
  • They are (likely to go) viral or contain a claim that has gone viral in the past. Note that we don't use fixed numerical cut-off rates (i.e. it must have more than x likes or y views), we tend to look more at the relative ranking (i.e. "this is the most viewed video about this particular conspiracy" or "this is the most popular fact checking related Google search today") and we sometimes also look at the number of places a claim is appearing (hundreds of tweets with the same image and one image tweet with hundreds of retweets are both valid reasons to check out the image). To determine which items are "likely" to go viral we look at short term steep increases in engagement and the audience size of the source making or spreading the claim (an Instagram page with a million followers vs. an obscure blog read by almost nobody).
  • They are relevant to or affecting the audience of the website we publish the fact check on. We will generally leave the internal politics, hoaxes and harmful misinfo going viral in other languages/countries to the fact checkers in those countries unless there is a local angle or the same claims are also spreading in the country or countries covered by one of our websites.

Generally we prioritize stories related to current events and new and never before seen claims but we also tackle older claims sometimes, especially if they have a tendency to keep coming back periodically (think old hoax articles that keep getting reposted to new sites by financially motivated spammers or fake quotes attributed to a politician that keep showing up whenever that person is in the news).

We do not take political slant into consideration when choosing what to check. We check claims that harm (or benefit) people, causes and groups all over the political spectrum.

We do sometimes check claims that originated on satire websites or that are satirical in nature but only in cases where it seems significant numbers of people believe it is real. You can see our Satire Policy here.

Reader tips are judged by the same criteria listed above: if we get sent claims that are uncheckable, harmless or not very widespread there is very little chance we will follow up on them. Conversely, a concrete, harmful and viral story will get our attention.

One of the aims of our fact checking is to help stop or slow down the spread of trending false information. This means we rarely publish fact checks to point out something is indeed true: if the truth is going viral by itself we are quite happy to leave it alone.


How we fact check claims and stories

Methodology

Each claim is different and not all of them require the same approach. Generally we first try to reframe the claim as a question ("Is it true that X?") and then our goal is to provide an answer to that question along with the reason and the available evidence.

There are questions we usually ask and try to answer in order to be able to say if a claim is true or not (although not all of them may apply in all situations):

  • Where did the claim originally come from and is there evidence for that orgin? If something first appeared in a fictional setting or in a satirical piece it is quite unlikely to be true. In some cases it is unfortunately not possible to find the original source.
  • If the source is known: what evidence does it cite and does it support the claim? (If an article says "New study says X", does the study actually say X?)
  • Are there logical inconsistencies or unwarranted leaps and assumptions about the evidence in the claim? ("Person X was once seen with person Y and there is a photo of it, that must mean they are both involved in an elaborate scheme!")
  • Are there alternate explanations for the evidence? ("Person X was arresting person Y in the photo", "The photo was digitally manipulated", "The photo does not actually show person X or Y"...)
  • If a source claims to have inside or confidential information, is it likely/possible they would have it? ("Would a random guy on the internet have access to Top Secret higly classified intel from the CIA? If so: how and why isn't such a security lapse not the a bigger story here?")
  • Is the evidence from trustworthy official or reputable scientific sources (or is there such evidence available)? We tend to look for official statistics with a known methodology that have been collected over longer periods of time and peer-reviewed scientific articles about double-blind studies with large sample sizes. Pre-prints, non reviewed studies, press releases with no link to a full study, survey results, studies with very small sample sizes, publications from special interest groups etc. have far less value as evidence or support for a claim.
  • Can the people named as making claims in the story (or named in the claims) be contacted for comment? This is unfortunately not always possible: there might not be any contact info, they might be using a pseudonym or they might be hard to reach. There may also be cases where there are safety or other legitimate concerns that preclude a contact attempt. Barring those exceptions we will try anyway if it is important for checking the truth of a claim. If we receive no reply within a reasonable timeframe and in cases where we obtained sufficient evidence through other sources we reserve the right to publish the fact check anyway, leaving the option open to include the reply later.
  • Are there experts or local sources who can confirm or disprove the claim? When quoting experts or sources we prefer named ones with disclosed affiliations and qualifications and we prefer not just to take their word for it but to look for additional verifiable evidence where possible.
  • Does technical analysis of the source and/or evidence show proof of tampering, editing or manipulation?
  • Does the claim or the evidence match with the record? ("Does this quote really occur in this recorded interview?", "Was this photo really taken on date X in location Y?")
  • Does the claim mix things up or cause confusion, intentional or not? (Almost any judge, cop or prison guard in the world can be said to have had "associations with known criminals" and every almost every fireman "has been at the scene of a suspicious fire carrying an axe", that does not mean they are evil or something nefarious was going on.)
  • Can we find evidence from primary sources (preferably more than one, where possible/relevant)? I.e. instead of an article about a document or a published interview with a person, we prefer to have the document itself or to get comment from the person directly where possible.

Our goal is to serve our readers by publishing an article reliably proving or disproving a claim as quickly as possible. This means we tend to publish as soon as we have collected enough information to confidently make the call but we sometimes continue updating the story while we collect even more supporting evidence or when sources get back to us (more experts/sources calling back, FOIA requests being answered...).

All our articles are reviewed by at least three people before publication (the writer and at least two editors).

In the (fortunately rare) case where our initial assessment was wrong or when newer evidence requires changes to the conclusion we will update the story in accordance with our Corrections Policy.

Note that in some cases it is impossible to prove or disprove a claim, simply because no good evidence is available. In such cases we will point out there is no evidence we could find for or against the claim and that the person or website that made the original claim is doing so without providing any proof. We will also explain how and where we looked for evidence.

Headlines, thumbnail captions, article structure & our (lack of) rating scale

All our articles generally follow the same structure to allow for easy scanning and to avoid accidentally spreading false information.

We know many people will only ever see our headlines and thumbnail images on social media and that the vast majority of people who see a social media post will not click through to read the full article. Of those who do, many only read the first paragraph.

So we aim to put as much information in the headline, thumbnail image and first paragraph as we can.

Headline

Generally we try to have the claim we are fact checking in the headline, but with the addition of the words "Fact Check:" and a capitalized word to negate the meaning (when the claim is not true).

For example: "This Is An Apple" becomes "Fact Check: This Is NOT An Apple".

If an original viral article or video we are fact checking has the claim in the headline we will try to stick as close as possible to the wording of that headline to make it easier to find our story when people look up that headline through search engines.

If the original headline does not contain the claim being checked we will stick to the claim in our own headline. For example if the original headline reads "World Shocked By Person X" and the article then talks about something X did not actually do our headline would be "Fact Check: X Did NOT Do Y" instead of "Fact Check: World NOT Shocked By Person X".

Note that we also don't reflexively negate every headline or claim because proving the opposite is sometimes equally impossible or the negated headline has a different meaning than what we intend to say. If the claim is "X Killed Y" and there simply is no evidence either way that X did it or not, it would not be correct to say "X Did NOT Kill Y". In such cases we would say something like "NO Evidence X Killed Y".

And when a headline reads "X Causes Y According to Scientists" and it turns out the scientists did not actually say that, our headline would likely be something like "Scientists Did NOT Say X Causes Y" rather than "X Does NOT Cause Y According To Scientists".

Sometimes multiple negations might be needed to fully explain what is going on: "Fact Check: X NOT Arrested For Doing Y" would still imply X did do Y (but he got away with it) so if X is innocent the headline would be "Fact Check: X NOT Arrested; Did NOT Do Y".

The goal of our headlines is to immediately tell our readers the conclusion of the fact check so that even if they don't read the rest of the article they have not been exposed to false or misleading information. That is also the reason why we don't use question headlines. If someone only sees "Fact Check: Did X Kill Y?" they might still get it stuck in the back of their mind that maybe X really was the murderer of Y or that questions are being asked about it.

Thumbnail captions

Our stories all contain a screenshot of the original post/article/website/tweet/video that made the claim (or at least a representative example of it). This screenshot will have a caption superimposed over it that can contain a maximum of fourteen characters. This captioned image is also the thumbnail image when our stories get shared on social media.

Since the spring of 2020 we ask our writers not to put a summary of the *conclusion* of the fact check in the caption anymore because that information should already be in the headline. So you shouldn't see things like this on recent stories:

  • False
  • Misleading
  • Not True
  • Fake
  • Hoax
  • ...

Instead we ask our writers to boil down and summarize the reason for the conclusion there in one to three words.

Why is there a problem with the claim? Why is X NOT Y?

Some good examples of what the caption can be include:

  • "Old Numbers"
  • "Photoshop"
  • "Missing Info"
  • "Still Alive"
  • "Scam"
  • "Just A Joke"
  • "Slowed Video"
  • ...

Here is an example for a hypothetical post with the headline "Fact Check: This Is NOT An Apple":

Why is this not an apple? Because it is a banana!

Article structure

Our articles all have the same basic structure:

  • A headline that starts with "Fact Check:" and which immediately contains the conclusion of the fact check ("X Did NOT Do Y").
  • Introductory paragraph that starts with the claim as a question, directly followed by the answer and a two or three sentence summary of the reason for that answer. ("Did X Do Y? No, that's not true. Video shows Y was at a different place when X happened and local police confirmed this.")
  • A link to the oldest version of the claim or to a representative example of a post making the claim, followed by a short excerpt or a screenshot. ("The claim appeared in an article on site X, published on date Y, titled Z. The story opened...")
  • The detailed fact check, listing all the evidence and reasons. Our goal is to provide links to all relevant info so that readers can replicate the fact check for themselves ("Here is an embedded security video... Police sources say... We spoke to...").
  • If the fact check is about a claim by a site or page that we have repeatedly fact checked in the past or about a topic that we have written about before we may also include more information about it + links to earlier fact checks ("We wrote about reallyreliabletotallynotfakesite.com before, it is run by spammers who...").

There are also some key principles we ask our writers to adhere to:

  • Always provide links to the original story or source so our readers can verify we are properly representing what is said there.
  • All links to information that is important for (dis)proving the claim should be backed up on archive.org and backup links should be provided in the story. I.e. if a story links to a report it should be linked to like this: "as you can read in this report (archived here)".
  • Stick to the facts that are important for (dis)proving the claim and avoid giving too much background info that is not relevant for doing that, especially on deeply partisan issues or with privacy-sensitive data.
    • When checking if a number in the gun control debate (or about terminating pregnancies or about nuclear power or any controversial subject) is really x% or not, there is no need to include quotes from people or groups saying why that is a good or bad thing or how much it *should* be or how much it has historically been and why that was better or worse. People who want to know if something is good or bad can read about that on other websites or in opinion sections. People who want to know the latest news or the full history about a topic can check Google News or Wikipedia. Our site is there for people who just want to know if something is true or not.
    • When a claim involves privacy-sensitive information (names, photos, contact information...), especially about minors, care must be taken to only include information that is needed to prove or disprove the veracity of the claim and to obscure privacy-sensitive information that is not relevant.

No standardized rating scale

Lead Stories does not use a standardized rating scale or score to rate claims on our website. When we investigate a false or misleading claim we simply aim to explain what exactly is wrong with it and why. It would only be a source of confusion to also attempt to add a score, rating, number or a standardized description to say precisely how wrong it is. What is the difference between "half wrong", "partly real" or "mixed"? Is a claim that is 72% false really that much worse than one that is 28% true? Are five "TruthPoints" really better than four and what does that one point difference mean?

Most of our readers don't care or don't have the time to worry about distinctions like that.

They just want to know if that story they saw online is true or not. So we do our best to tell them.