
Image Moderation API
Pricing
Pay per usage

Image Moderation API
Uses advanced AI models to analyze and classify user-generated content in real time. It detects harmful or inappropriate content, providing category-level flags and confidence scores to help you enforce community guidelines and keep your platform safe.
0.0 (0)
Pricing
Pay per usage
0
Total users
2
Monthly users
2
Runs succeeded
>99%
Last modified
18 days ago
🖼️ AI Image Moderation Actor
This Apify Actor leverages Sentinel Moderation's AI-powered API to analyze and flag images containing inappropriate, unsafe, or policy-violating content. It can detect NSFW material, violence, graphic content, and more — helping you maintain a safe and compliant platform.
📥 Input Schema
The actor expects the following JSON input:
1{ 2 "apiKey": "sample-api-key", 3 "image": "https://example.com/path-to-image.jpg" 4}
apiKey
(string, required): Your API key from SentinelModeration.com.image
(string, required): A publicly accessible image URL to analyze.
📤 Output
The actor returns a moderation result in the following structure:
1[ 2 { 3 "flagged": false, 4 "categories": { 5 "harassment": false, 6 "harassment/threatening": false, 7 "sexual": false, 8 "hate": false, 9 "hate/threatening": false, 10 "illicit": false, 11 "illicit/violent": false, 12 "self-harm/intent": false, 13 "self-harm/instructions": false, 14 "self-harm": false, 15 "sexual/minors": false, 16 "violence": false, 17 "violence/graphic": false 18 }, 19 "category_scores": { 20 "harassment": 0.000048, 21 "harassment/threatening": 0.0000066, 22 "sexual": 0.000039, 23 "hate": 0.0000142, 24 "hate/threatening": 0.0000008, 25 "illicit": 0.000022, 26 "illicit/violent": 0.000019, 27 "self-harm/intent": 0.0000011, 28 "self-harm/instructions": 0.0000010, 29 "self-harm": 0.0000020, 30 "sexual/minors": 0.000010, 31 "violence": 0.000016, 32 "violence/graphic": 0.0000056 33 }, 34 "error": null 35 } 36]
flagged
:true
if any category crosses a moderation threshold.categories
: A true/false map indicating which categories were flagged.category_scores
: Confidence scores (0.0 to 1.0) for each category.error
: Present only in test mode or if no valid API key is provided.
🧠 Categories Detected
The image is scanned for content under a wide range of moderation labels:
- Harassment / Threats
- Sexual content (including minors)
- Hate speech (including threats)
- Illicit activity
- Self-harm
- Violence / Graphic content
🔐 Getting an API Key
To receive real results, get your API key from Sentinel Moderation:
- Visit sentinelmoderation.com
- Sign up and generate your API key
- Use the key in the
apiKey
field of your input
✅ Example Use Cases
- Flagging NSFW content in profile photos or uploads
- Moderating image submissions on forums or marketplaces
- Pre-screening media in chat apps or social platforms
- Complying with platform-specific safety guidelines
Let me know if you also want to add support for uploading images directly instead of via URL!