LET'S LEARN ABOUT ML RISKS
We have developed our own machine learning models to detect risk more accurately in your cloud environments. They are available to all customers and can be enabled by reaching out to support@managedmethods.com.
Self Harm ML:
The self harm ML risk is a continually improving way to flag content that indicates self harm. We trained it using massive data sets from various sources and are continually re-training it using both new material that we've found and content submitted by our customers using the "False Positive" reporting feature with the platform. Because of this, we are able to keep honing the accuracy of what is considered a risk and reduce the overall number of false positives. See THIS guide page for details on submitting false positives.
Image Risk ML:
Toxicity ML:
- Identity Attacks: Attacks involving race, sexuality, gender, or other identities.
- Insults: Content aiming an insult at an individual.
- Obscene: Content containing sexually explicit, or vulgar language including profanity.
- Threats: content containing threatening language.
See THIS guide page for details on submitting false positives.
Comments
0 comments
Please sign in to leave a comment.