🛡️ GIOMIND – CONTENT MODERATION POLICY

HOW WE KEEP THE PLATFORM SAFE

Last Updated: December 1, 2025

Effective Date: December 1, 2025

📋 INTRODUCTION

This Content Moderation Policy explains how GioMind moderates content to maintain a safe, supportive environment. It describes our moderation systems, processes, and the balance between automated and human review.

This policy supplements our Community Guidelines and Terms of Use.

TABLE OF CONTENTS

1. Purpose and Scope

2. What We Moderate

3. Moderation Systems Overview

4. Automated Content Filtering

5. AI Safety Measures

6. Human Review Process

7. Proactive vs. Reactive Moderation

8. User Reports

9. Content Removal and Actions

10. Transparency and Accountability

11. Limitations of Moderation

12. Your Role in Safety

13. Appeals and Feedback

14. Changes to This Policy

1. PURPOSE AND SCOPE

1.1 Why We Moderate Content

Content moderation exists to:

✓ Protect users from harmful content

✓ Maintain a safe and supportive environment

✓ Enforce Community Guidelines and Terms of Use

✓ Prevent illegal activity

✓ Reduce risks associated with AI-generated content

✓ Comply with legal requirements

1.2 What This Policy Covers

This policy applies to:

• AI-generated content (conversations with AI)

• User-generated content (journal entries if shared, future features)

• Any content created or transmitted through GioMind

• Content from all users, regardless of location

1.3 Legal Framework

Our moderation practices comply with:

• Digital Services Act (DSA) – European Union

• Online Safety Act – United Kingdom (as applicable)

• Section 230 protections – United States

• Other applicable local laws

2. WHAT WE MODERATE

2.1 Content Categories Subject to Moderation

We moderate content related to:

High Priority (Zero Tolerance):

🚨 Child Sexual Abuse Material (CSAM)

🚨 Terrorism and violent extremism

🚨 Credible threats of violence

🚨 Human trafficking and exploitation

🚨 Illegal weapons or explosives instructions

High Priority (Strict Enforcement):

⚠️ Hate speech and discrimination

⚠️ Targeted harassment and bullying

⚠️ Self-harm instructions or encouragement

⚠️ Dangerous health misinformation

⚠️ Sexual content and harassment

⚠️ Doxxing and privacy violations

Medium Priority:

• Spam and manipulation

• Misinformation (non-dangerous)

• Inappropriate AI use

• Platform abuse

• Copyright infringement

Low Priority (Context-Dependent):

• Off-topic content

• Oversharing personal information

• Minor guideline violations

2.2 Content We Do NOT Moderate

We generally do NOT moderate:

✓ Private journal entries (not shared publicly)

✓ Opinions or viewpoints (unless violating guidelines)

✓ Criticism of GioMind (constructive feedback is welcome)

✓ Discussions of controversial but legal topics

2.3 Context Matters

• Content is evaluated in context

• Educational or awareness content may be allowed

• Intent matters (malicious vs. accidental)

• We aim for nuanced, fair moderation

3. MODERATION SYSTEMS OVERVIEW

3.1 Multi-Layered Approach

GioMind uses a combination of:

1. Automated filtering (pre-AI and post-AI)

2. AI provider safety systems (OpenAI, Anthropic)

3. Proactive monitoring (sampling, pattern detection)

4. User reports (community flagging)

5. Human review (when needed)

3.2 Moderation Flow

```

User Input

[Pre-Filter] → Block obvious violations

AI Provider → Generate response + safety check

[Post-Filter] → Review AI output

Deliver to User

[Ongoing Monitoring] → Proactive review & user reports

[Human Review] → If flagged or reported

[Action] → Warning, content removal, account action

```

3.3 Speed vs. Accuracy Trade-off

• Automated systems are fast but imperfect

• Human review is accurate but slower

• We balance speed (to prevent harm) with fairness (to avoid false positives)

4. AUTOMATED CONTENT FILTERING

4.1 Pre-Input Filtering

Before content reaches AI, we filter for:

• Known harmful keywords or phrases

• Patterns indicative of prohibited requests

• Explicit sexual content indicators

• CSAM-related terms (zero tolerance)

• Violence and threat patterns

Actions:

• Block the request immediately

• Show user a warning message

• Log the attempt for monitoring

Example:

```

User: [Attempts prohibited request]

System: ⚠️ This request cannot be processed as it

violates our Community Guidelines.

```

4.2 Post-Output Filtering

After AI generates a response, we filter for:

• Harmful content in AI responses

• Medical advice or diagnoses (shouldn't occur but double-check)

• Inappropriate suggestions

• Content that slipped through AI provider filters

Actions:

• Block the response from being shown

• Show generic safe response instead

• Log for review and system improvement

Example:

```

AI generates inappropriate content

System blocks it

User sees: "I apologize, but I cannot provide that

type of information. Let me help with

something else."

```

4.3 Keyword and Pattern Detection

We use:

• Keyword lists (regularly updated)

• Regular expressions (pattern matching)

• Semantic analysis (meaning detection)

• Machine learning classifiers

Limitations:

⚠️ Can produce false positives (blocking safe content)

⚠️ Can miss creative evasion attempts

⚠️ Context may be misunderstood

4.4 False Positive Handling

If legitimate content is blocked:

• You can report the issue

• We review and adjust filters

• We aim to minimize false positives while maintaining safety

5. AI SAFETY MEASURES

5.1 AI Provider Built-In Safety

OpenAI:

• Has safety filters for harmful content

• Refuses certain types of requests

• Monitors for policy violations

• Updates safety systems regularly

Anthropic:

• Constitutional AI approach (trained to be helpful, harmless, honest)

• Built-in safety guardrails

• Refuses harmful requests

We rely on these providers but add our own layers.

5.2 GioMind-Specific AI Instructions (Prompts)

We instruct AI to:

✓ Focus on general wellness, not medical advice

✓ Refuse to diagnose or treat conditions

✓ Not provide crisis support (direct to professionals)

✓ Avoid generating harmful content

✓ Be supportive and non-judgmental

✓ Stay within wellness topics

5.3 AI Limitations We Cannot Fully Control

Despite measures, AI may:

❌ Sometimes generate inappropriate content

❌ Not always follow instructions perfectly

❌ Respond to cleverly-worded harmful requests

❌ Make mistakes in judgment

This is why we have multiple layers of filtering.

5.4 Jailbreak Detection

We monitor for attempts to:

• Override AI safety instructions

• Use prompt injection techniques

• Manipulate AI into harmful outputs

Actions:

• Block jailbreak attempts

• Flag accounts for repeat attempts

• May result in account suspension

6. HUMAN REVIEW PROCESS

6.1 When Human Review Occurs

Human moderators review content when:

• Automated systems flag potentially violating content

• Users report content

• Proactive sampling for quality assurance

• AI generates uncertain or borderline content

• Appeals are submitted

• Legal requests require review

6.2 What Human Reviewers Do

Moderators:

✓ Evaluate flagged content in context

✓ Determine if Community Guidelines are violated

✓ Decide on appropriate enforcement action

✓ Provide feedback to improve automated systems

✓ Handle complex or nuanced cases

6.3 Human Review Limitations

We do NOT:

❌ Review all conversations in real-time

❌ Monitor every AI interaction manually

❌ Read all journal entries (private by default)

❌ Have 24/7 human moderation coverage

Practical Constraints:

• Limited human resources

• Privacy considerations

• Volume of content is too high for 100% human review

6.4 Reviewer Training and Support

Our moderators:

✓ Are trained on Community Guidelines

✓ Use consistent evaluation criteria

✓ Follow internal moderation playbooks

✓ Have access to wellness support (to prevent burnout from reviewing harmful content)

6.5 Privacy Protections in Review

• Reviewers see only content necessary for evaluation

• Personal data is minimized where possible

• Reviewers are bound by confidentiality

• Data is handled according to Privacy Policy

7. PROACTIVE VS. REACTIVE MODERATION

7.1 Proactive Moderation

We proactively:

✓ Sample conversations randomly for quality checks

✓ Use automated systems to detect patterns

✓ Monitor for emerging threats or trends

✓ Review high-risk content categories more closely

✓ Update filters based on new risks

Purpose:

• Catch violations before they're reported

• Improve automated systems

• Identify systemic issues

7.2 Reactive Moderation

We reactively:

✓ Respond to user reports

✓ Investigate flagged content

✓ Address specific complaints

✓ Handle appeals

Purpose:

• Empower community to help moderate

• Address issues users care about

• Provide recourse for victims

7.3 Balance

• Proactive: Prevents harms before they spread

• Reactive: Respects privacy and community input

• We balance both approaches

8. USER REPORTS

8.1 Importance of User Reports

• You are our first line of defense

• You see content we may miss

• Community reporting helps us scale moderation

8.2 How to Report

See Community Guidelines Section 11 for full reporting instructions.

Quick Summary:

📧 Email: giomind.app@gmail.com

Subject: "Report Violation"

Or use in-app reporting (if available).

8.3 What Happens After You Report

1. Receipt: We receive and log your report

2. Triage: We assess urgency and priority

3. Review: Human moderator reviews the content

4. Action: We take appropriate action (if violation confirmed)

5. Feedback: You may receive confirmation (depending on case)

Timeline:

• High-priority (CSAM, violence): Immediate (within hours)

• Medium-priority: 24-48 hours

• Low-priority: Within 7 days

8.4 Report Quality Matters

Good Reports:

✓ Clear description of violation

✓ Specific location/time

✓ Screenshots or evidence

✓ Honest and accurate

Poor Reports:

❌ Vague or unclear

❌ False or malicious

❌ Frivolous or spam reports

8.5 False Reporting

• Abuse of reporting system is prohibited

• Repeated false reports may result in account action

• Report honestly and in good faith

9. CONTENT REMOVAL AND ACTIONS

9.1 Possible Actions

When violations are confirmed, we may:

Content-Level Actions:

• Remove or hide content

• Add warning labels

• Limit content visibility

Account-Level Actions:

• Issue warning

• Temporarily suspend account

• Permanently ban account

• Restrict specific features

Other Actions:

• Report to law enforcement (if illegal)

• Report to AI providers (for their records)

• Update filters to prevent similar content

9.2 Action Criteria

We consider:

• Severity of violation

• User's history

• Intent (malicious vs. accidental)

• Impact on others

• Legal requirements

9.3 Communication of Actions

You will be notified if:

• Your content is removed

• Your account receives a warning or suspension

• Action is taken on your report (sometimes)

Notification includes:

• What action was taken

• Reason for action

• How to appeal (if applicable)

9.4 Transparency in Enforcement

• We strive to explain our decisions

• Some actions cannot be disclosed (legal/privacy reasons)

• We publish general moderation statistics (when feasible)

10. TRANSPARENCY AND ACCOUNTABILITY

10.1 Transparency Report (Future)

We may publish periodic transparency reports including:

• Number of reports received

• Types of violations

• Actions taken

• Appeal outcomes

• System improvements

10.2 External Audit and Research

• We may engage external auditors to review moderation practices

• We support independent research on platform safety

• We are open to feedback and improvement

10.3 Public Engagement

• We welcome feedback on moderation policies

• We may consult users or experts on policy updates

• We are committed to ongoing improvement

10.4 Accountability to Regulators

• We comply with legal reporting requirements (e.g., DSA)

• We cooperate with law enforcement when legally required

• We respond to regulatory inquiries

11. LIMITATIONS OF MODERATION

⚠️ IMPORTANT – MODERATION IS NOT PERFECT ⚠️

11.1 We Cannot Catch Everything

Despite our efforts:

❌ Some harmful content may slip through

❌ Automated systems have false negatives

❌ Clever evasion attempts may succeed

❌ Human review cannot cover all content

❌ Context may be misunderstood

11.2 False Positives and False Negatives

False Positives:

• Safe content incorrectly flagged/removed

• We work to minimize these

• You can appeal

False Negatives:

• Harmful content incorrectly allowed

• We work to reduce these

• Report when you see them

11.3 No Guarantee of Safety

⚠️ We do NOT guarantee a completely safe or harm-free environment

⚠️ You may encounter content that disturbs or upsets you

⚠️ You use GioMind at your own risk

11.4 Technical Limitations

• AI is imperfect and evolving

• Moderation technology has limitations

• New types of harm emerge constantly

• We are always playing catch-up

11.5 Your Responsibility

• You are responsible for your own safety

• Use judgment when engaging with content

• Report violations when you see them

• Stop using features that cause distress

12. YOUR ROLE IN SAFETY

🤝 WE NEED YOUR HELP 🤝

12.1 Be a Responsible User

✓ Follow Community Guidelines

✓ Don't create harmful content

✓ Don't abuse AI features

✓ Respect others

12.2 Report Violations

✓ Report harmful content when you see it

✓ Report AI malfunctions or inappropriate responses

✓ Report bugs or security issues

12.3 Provide Feedback

✓ Tell us how we can improve

✓ Share ideas for better safety features

✓ Help us understand user needs

12.4 Take Care of Yourself

✓ Stop using GioMind if it causes distress

✓ Seek professional help if needed

✓ Don't rely on GioMind for safety-critical needs

13. APPEALS AND FEEDBACK

13.1 Appeal Process

If you disagree with a moderation decision:

• See Community Guidelines Section 12 for appeals process

• Email: giomind.app@gmail.com

• Subject: "Appeal – [Account Email]"

13.2 Feedback on Moderation

For general feedback:

📧 Email: giomind.app@gmail.com

Subject: "Moderation Feedback"

We review and consider all feedback.

14. CHANGES TO THIS POLICY

14.1 Updates

We may update this Content Moderation Policy to:

• Reflect new moderation technologies

• Address emerging threats

• Comply with legal requirements

• Improve clarity or effectiveness

14.2 Notice

• Updated policy will be posted in the app

• Material changes will be communicated via email or in-app notification

• Continued use constitutes acceptance

📋 SUMMARY – HOW WE MODERATE

Systems:

• Automated filters (pre and post AI)

• AI provider safety systems

• Human review (when needed)

• User reports

What We Moderate:

• Illegal content (zero tolerance)

• Harmful content (hate, violence, harassment)

• Dangerous misinformation

• Inappropriate AI use

• Platform abuse

Limitations:

• Cannot catch everything

• False positives and negatives occur

• No guarantee of complete safety

Your Role:

• Follow guidelines

• Report violations

• Provide feedback

Appeals:

• You can appeal moderation decisions

• Email giomind.app@gmail.com

🛡️ OUR COMMITMENT

We are committed to:

✓ Maintaining a safe platform

✓ Transparent moderation practices

✓ Continuous improvement

✓ Balancing safety and user freedom

✓ Accountability to users and regulators

Thank you for helping us keep GioMind safe and supportive.

📧 CONTACT

Questions or concerns about content moderation:

Email: giomind.app@gmail.com

Subject: "Content Moderation Question"

Last Updated: December 1, 2025

© 2025 GioMind. All Rights Reserved.