Practical training using DefenderNet tools and real-world moderation scenarios.
After completing this module you will understand the following key concepts.
Moderators often encounter the same kinds of harmful behavior, but different communities may describe or interpret those behaviors in different ways. That can make moderation less consistent and make serious harm easier to misunderstand, downplay, or miss.
Shared language helps reduce confusion. It gives moderators a more consistent way to describe harmful behavior, supports clearer decision-making, and makes it easier for teams to stay aligned when reviewing incidents.
This becomes especially important when serious harms are involved. Without a shared way to classify behavior, one moderator or server may recognize a situation as severe, while another may misread it, minimize it, or respond too late.
DefenderNet's 20 Violation Categories is one example of a structured system that helps Discord and Minecraft communities classify different harms more clearly, including the most serious ones, while still allowing each community to decide its own rules and consequences.
By using standardized categories, DefenderNet helps communities classify and respond to harmful behavior more consistently. This makes it easier to act early, share signals across communities, and reduce repeat harm.
CSAM (Child Sexual Abuse Material) is explicit and severe, covering any visual or descriptive material that depicts child sexual abuse. This includes photos, videos, audio, live streams, computer-generated or illustrated depictions that appear realistic, or even textual content such as stories of child sexual abuse and exploitation. Possession, sharing, or linking to such content is illegal. Because this type of material is a serious form of harm, any case involving CSAM must be treated as a severe violation and escalated immediately to the platform and the appropriate authorities.
Having a standardized category for CSAM helps teams respond faster, reduce ambiguity, and handle serious cases more consistently.
One of the most important skills for moderators is knowing how to recognize and classify harmful behavior clearly and consistently. This makes enforcement easier, consistency stronger, data more reliable, and better protection across every community.
A behavior-first approach means paying attention to what someone is doing, not just waiting for the worst possible outcome before taking action.
This means moderators and communities do not have to wait until explicit material is shared before recognizing and reporting serious harm.
For example, any behavior intended to sexually manipulate, exploit, or harm a child — including grooming, coercive communication, the creation or editing of images, the promotion of harmful stereotypes, or the formation of inappropriate relationships with underage users — can be flagged even when no explicit content is shared.
This approach recognizes that the intent and behavior itself are harmful and allows community and moderation teams to act at the earliest stages.
The next modules will explore two important parts of addressing harmful behavior in your community: how to spot early warning signs, and how to respond when harmful behavior is identified.