When does it sound reasonable to say: that’s true, but I don’t want to amplify a toxic person

Thesis and practical problem

Rejecting a claim because of who states it or where it comes from is commonly described as the genetic fallacy. The intuition is simple: the origin of an idea, by itself, does not determine whether it is true or false. In real discussions, this pattern shows up frequently and not only as a logical mistake: it can act as an accelerant of polarization, because it replaces “what reasons and evidence are there?” with “which side does it come from?”.

A particularly influential variant today is the appeal to “not giving a platform to harmful voices”. In the abstract, that idea can have a defensible use as a governance criterion for a space (for example, to prevent harassment or incitement to violence). But it often functions as a pretext: it is used to harm the enemy, deny any point of agreement, and avoid the social cost of admitting that the other side can be right. When this happens, the effect can be paradoxical: in the name of “making the world better”, the mechanism that intensifies polarization is reinforced.

The thesis of this text is that, if the goal is constructive debate and less polarization, it is usually more effective to separate lanes (truth, evaluation of the speaker, and forum governance) and, additionally, to make visible verifiable points of agreement even when they come from “the other side”. This does not require absolving anyone; it requires not turning identity into a substitute for evidence.

 

What the genetic fallacy is (and what it is not)

The genetic fallacy consists in treating origin as refutation: “this comes from X, therefore it is false”, or “Y said it, therefore it does not deserve consideration”. The problem is not that origin is always irrelevant; it is that it is used as a verdict when, at most, it should be a preliminary signal.

Distinguish this from reasonable practices:

  • Using origin as an initial heuristic: if a source has a consistent history of manipulation, lowering prior trust can be prudent.
  • Demanding traceability: asking for data, method, replication, a chain of references.
  • Considering incentives and conflicts of interest: origin can indicate reasons to apply greater scrutiny.

None of these practices refutes content by itself. They only adjust initial confidence and verification effort.

 

Why it is so tempting

Avoiding it looks easy in the abstract. In practice, when we know things about a speaker that disgust us—because of conduct, values, or track record—the reaction can be almost automatic. A kind of “affective contamination” occurs: emotional rejection or contempt sticks to the content and ends up functioning, without us noticing, as a substitute for evidence.

That impulse is not mysterious. It is a cognitive and social shortcut:

  • It reduces effort: labeling is cheaper than analysis.
  • It protects identity: it avoids admitting the rival might be right about something.
  • It closes uncomfortable conversations without engaging specifics.
  • It strengthens group cohesion: “if it comes from them, reject it”.

This helps explain why the fallacy persists: it is not only a logical error; it is also a tool of tribal dynamics.

 

Advantages for whoever uses it

  • Saves time and energy: labeling is usually cheaper than analysis.
  • Persuades aligned audiences easily: it confirms loyalties.
  • Controls the frame: the focus shifts from content to reputation or identity.
  • Protects self-esteem and identity: it reduces the risk of conceding points to the other.
  • “Vaccinates” the group: it predisposes people to reject anything from the rival.

The problem is that these advantages tend to be short-term and, accumulated over time, can degrade collective conversation.

 

The “don’t amplify” pretext and the hijacking of the truth lane

Here is the main friction. In the abstract, it sounds reasonable to say: “even if it is true, I don’t want to amplify a toxic person”. But in many discussions that phrase operates less as prudent risk management for a space and more as a license to harm an adversary: deny legitimacy, block any agreement, and socially punish the recognition of the other side’s correct points.

When “don’t amplify” becomes an informal factional rule, something like this typically happens:

  • The status of the speaker (character, virtue, conduct) is conflated with the truth of the claim.
  • Verifiable agreement is penalized: “if you accept it, you are betraying the group”.
  • Polarization intensifies: each side becomes unable to learn from the other even in what is checkable.

In that usage, “don’t amplify” is a practical version of the genetic fallacy: the debate stops being “is it true?” and becomes “does it deserve to exist in our space?”, without explicit, symmetric, auditable criteria.

If the goal is to reduce polarization, the effect is usually the opposite: silencing the other side, even when it is right, tends to reinforce factional logic. And that logic turns truth into group property: “it is only acceptable if one of us says it”.

 

Three lanes: truth, evaluation of the speaker, and forum governance

To avoid confusion, it helps to separate three lanes:

  • Epistemic lane: is the claim true, false, or indeterminate given what is available?
  • Speaker lane (character/virtue and conduct): what assessment does the person deserve, their behavior, and their reliability as an agent?
  • Forum governance lane: what rules keep debate possible without turning into intimidation, harassment, or violence?

The third lane is the only place where “not giving a platform” can have a defensible meaning without degrading the epistemic lane. If the lanes are not separated, “moderation” and “refutation” get mixed, and disagreement becomes a war of legitimacy.

A formulation that preserves the separation would be:

“I do not want to interact with that person or I do not want them present in certain spaces for governance reasons. I can reproach their conduct or consider their character/virtue censurable. But that does not decide the truth of the claim, nor should it prevent recognizing verifiable correct points when they exist.”

 

The antidote to polarization: making “islands of agreement” visible

If the goal is constructive debate, it is often more effective to do the opposite of informal censorship: identify and make visible verifiable points of agreement when they exist. An “island of agreement” is a partial claim you can accept on evidence even if it comes from an adversary or from someone whose conduct you consider blameworthy.

Practicing this can have several plausible effects:

  • It reduces mutual caricature, because it forces recognition of complexity.
  • It weakens identity filtering (“it is only true if my side says it”).
  • It increases incentives to argue better: if agreement is possible, evidence becomes valuable.

Accepting an island of agreement does not imply absolution or whitewashing. It implies defending a norm: truth (or the best available approximation) should not depend on the identity of the messenger.

In terms of polarization, it is a bet: publicly recognizing correct points on the other side can lower temperature and open learning. It does not always work, but it is hard to imagine a less polarized public debate if recognizing correct points is socially forbidden.

 

Exceptions and the hard case: tolerance and the conditions of dialogue

Does this mean nobody should ever be restricted? Not necessarily. There are cases where the governance lane can justify limits, but they should be formulated narrowly and symmetrically so they do not become tribal pretexts.

A reasonable (though contestable) criterion is to restrict not because someone is “from the other side” or “disgusting”, but because they actively damage the conditions that make debate possible. Typical examples:

  • Incitement to violence.
  • Harassment, intimidation, doxxing.
  • Coordinated campaigns of targeted abuse.
  • Systematic sabotage of conversation through coercion or threat.

This connects with what is often summarized as the paradox of tolerance: unlimited tolerance toward those who seek to destroy the framework that enables tolerance can make that framework unviable. Even here, there are risks of abuse or capture. Safeguards help: proportionality, transparency, publishable rules, and the possibility of review.

Key point: even when a space restricts someone, this should not function as a refutation of their claims. If a claim is relevant and verifiable, it should be evaluable via independent paths (data, replication, additional sources) without turning messenger exclusion into “evidence” against the message.

 

Operational techniques to avoid the fallacy

  • Depersonalize: rewrite the thesis without the speaker’s name. If it sounds more reasonable, rejection may have been about origin.
  • Symmetry test: “if someone from my side said this, would I demand the same level of evidence?”.
  • Independent corroboration: find a verification path that does not depend on the speaker.
  • Partial steelman: state the best version of the argument before criticizing it; if you cannot, understanding or evidence may be missing.
  • Islands of agreement: identify an acceptable part and state it with conditions (“I accept X for these reasons, without implying Y”).
  • The prior rule: allow origin to adjust initial confidence, but require explicit evidence to settle the conclusion.

These techniques do not eliminate bias completely, but they reduce the chance that identity replaces evaluation.

 

AI, reputation, and traceability in an ocean of claims

An AI system could, in principle, be better positioned to avoid the genetic fallacy because it lacks visceral disgust in the human sense and can separate tasks: evaluate consistency, request evidence, cross-check references. But it would be unwise to assume “immunity”. It can reproduce similar shortcuts through:

  • Data bias: associations between groups and perceived reliability.
  • Design incentives: avoiding friction can push toward socially comfortable decisions.
  • Superficial correlations: learning “who tends to say what” as a proxy for truth.

The solution is not to idealize AI, but to require explicit procedures: distinguish “this lowers my initial confidence” from “this refutes the content”, show uncertainty, and prioritize traceability.

As content production grows (human and automated), origin will become an increasingly ambiguous signal. Traceability will matter more: data, methodology, replication, auditing. And so will cultural norms that reward recognizing correct points from the other side, because without that incentive verification can become politicized as well.

 

Practical recommendations

For individuals:

  • Separate lanes: truth, evaluation of the speaker, governance.
  • Practice islands of agreement: aim for at least one per discussion, even if small and conditional.
  • Make standards explicit: what evidence would change your mind.

For communities:

  • Penalize caricature; reward recognition of valid points from the rival.
  • Set explicit governance rules separate from ideological disagreement.
  • Prefer formats that require evidence and encourage concessions.

For platforms:

  • Moderation transparency: clear rules, symmetry, proportionality.
  • Designs that reduce incentives for viral humiliation and add friction to serious accusations.
  • Context and traceability mechanisms for verifiable claims.

 

Closing

The genetic fallacy is tempting because it is socially powerful: it saves effort, protects identity, and makes it easy to delegitimize an adversary. Its modern cousin, “don’t amplify”, can be a legitimate governance tool in narrow cases, but it often operates as a polarizing pretext: it blocks recognition of correct points from the other side and turns identity into a truth filter.

If the aim is less polarization, the antidote is not to pretend origin does not matter, but to put it in its place: origin can adjust initial confidence, evidence decides the conclusion, and governance protects the conditions of dialogue. Above all, it helps to normalize an uncomfortable but structurally valuable practice: precisely recognizing verifiable islands of agreement even when they come from people we do not like. Without that, conversation tends to degrade into exchanges of loyalty; with it, there is at least a real possibility of mutual learning.

Comments

Leave a Reply