Substack won't delete or defund Nazi content, they assert.

A detailed look at how content moderation issues are impacting Substack, a popular digital publishing platform, with a focus on a recent high-profile case.

Substack, a well-known platform for digital publishing, is currently grappling with dilemmas associated with content moderation. A scenario involving Hamish McKenzie, Substack's co-founder, recently illuminated these issues when he publicly disagreed with an author's removal. This incident has brought content moderation issues at Substack under sharper focus.

The conflict started when GoDaddy terminated web services to, a firearm enthusiast website, citing a policy violation. then switched to Epik, known for hosting controversial websites. This necessitated Substack to reconsider its standing arrangement with for promotional activities.

CVS, Rite Aid, Walgreens give cops medical records without warrants.
Related Article

The controversy took another turn when digital payment processor Stripe discontinued its services to for alleged policy violations. Subsequently, McKenzie stepped in to moderate and allowed to continue on Substack, despite Stripe's cessation of services.

Substack won

McKenzie's decision was based on his belief that Substack shouldn't remove any authors unless they violate Substack's content policies. He argued, given that Substack had not demonetized, it was not obliged to follow Stripe's decision.

However, critics pointed out an inconsistency with Substack's moderation principles. Previously, the platform had disallowed an author promoting Nazi ideologies, although this did not breach Substack's regulations. The critics argued for a need to standardize the moderation criteria and process.

The aforementioned Nazi promoter, a pseudonymous author known as Bronze Age Pervert, continued to thrive on Substack despite outrage. Many readers questioned why Substack allowed such content, indicating a contrasting approach to McKenzie's handling of

Substack initially responded to these concerns by stating that it wouldn't remove authors simply for their controversial views. This reflected the platform's commitment to free speech - a value it frequently highlights to differentiate itself from competitors.

However, enabling such free speech often results in problematic content that other platforms would censor. This has led Substack into a thorny territory, where it seems to be selectively moderating content while advocating for unrestricted expression.

Lawsuit claims Apple AirTags caused ruin & murders, as many join alleging they're favored tool for stalkers.
Related Article

The moderation issue at Substack runs deeper than just these controversies. There is an ongoing debate about where to draw a line between free speech and harmful, offensive content. This is a common struggle for many platforms, forcing them to constantly reassess their moderation strategies.

Substack's conflicting policies regarding content moderation reflect this struggle. As a platform, it plays a crucial role in controlling what is shared and read by millions worldwide. Any ambiguity in its policies can lead to inconsistent enforcement, as seen with and Bronze Age Pervert.

Substack's reluctance to take a decisive stand on moderation issues has led to a murky image of the platform. While it promotes itself as a haven for free speech, it appears to adopt a stricter stand when certain controversial issues emerge, leading to accusations of bias.

While many appreciate Substack's commitment to free speech, others believe that it should adopt a stricter moderation policy. Critics argue that providing a platform for potentially harmful or offensive content does more harm than good, especially given the platform's growing influence.

Responding to these arguments, McKenzie clarified that free speech at Substack doesn't mean avoiding moderation. Instead, Substack is building its character as a platform that doesn't compel compliance with conventional wisdom or popular opinion.

Yet, the concern remains that, without clear guidelines, the decisions on content removal rest with the company's founders. This can potentially lead to inconsistency and perceived bias in content moderation, which could harm Substack's reputation in the long run.

Substack acknowledges the complexity of the moderation issues it faces. It recognizes the fine balance between censorship and facilitating a robust exchange of ideas. The platform continues stepping up efforts to refine its moderation practices while maintaining its commitment to open discourse.

Ultimately, Substack's approach to content moderation will continue to evolve as it reacts to new controversies and absorbs criticism. However, the need for clear, consistent decision-making remains as paramount as ever.

As things stand, Substack has a tough journey ahead in managing its image while championing free speech. It must delicately balance both roles to maintain user trust and ensure the platform remains a credible source of content.

Clear guidelines for content moderation and effective enforcement mechanisms are vital for platforms like Substack. They not only ensure a safe space for users but also help to uphold free speech without enabling harmful narratives.

Substack's struggle to define and enforce its moderation policies encapsulates the broader debate about digital speech regulation. As this continues, users and observers will keenly follow how Substack navigates these grey areas while preserving its ethos of free speech.