1. Age Requirements
Networkli is a professional networking platform intended for users 18 years and older. We do not knowingly collect or maintain information from persons under 18 years of age. If we become aware that a user is under 18, their account will be immediately terminated.
2. Child Sexual Abuse Material (CSAM) Prevention
Networkli maintains a zero-tolerance policy for child sexual abuse material (CSAM):
- Proactive Monitoring: All user-generated content, including profiles, photos, and messages, is monitored using automated systems and manual review processes
- Content Detection: Advanced detection systems flag potentially inappropriate content for immediate review
- Swift Action: Any account found sharing, storing, or distributing CSAM is immediately terminated
- Law Enforcement Cooperation: We fully cooperate with law enforcement agencies and the National Center for Missing & Exploited Children (NCMEC)
- Mandatory Reporting: We report all instances of CSAM to appropriate authorities as required by law
3. User Safety Features
Networkli provides multiple safety features to protect our community:
- In-App Reporting: Users can report concerning behavior or content directly within the app
- Block Functionality: Users can block other users to prevent unwanted contact
- Profile Moderation: All profile information and photos are reviewed before being made public
- Message Monitoring: Our systems monitor messages for policy violations
- Age Verification: We employ measures to ensure users meet our age requirements
4. Reporting Mechanisms
If you have concerns about child safety on Networkli, please contact us immediately:
- Email: dthadd13@gmail.com
- In-App: Use the "Report" feature available on all profiles and messages
- Response Time: We respond to all child safety reports within 24 hours
For emergencies involving immediate danger to a child:
Please contact local law enforcement or:
- National Center for Missing & Exploited Children: 1-800-843-5678
- CyberTipline: www.cybertipline.org
5. Content Moderation Practices
Our content moderation system includes:
- Automated Filtering: AI-powered systems detect and flag potentially harmful content
- Human Review: Trained moderators review flagged content and user reports
- Community Guidelines: Clear guidelines that prohibit any content harmful to minors
- Regular Audits: Ongoing review of our safety systems and processes
- User Education: Resources to help users identify and report concerning behavior
6. Privacy Protection for Minors
In the unlikely event that we identify a user under 18:
- Immediate account suspension
- Deletion of all personal information
- Notification to parent or guardian (where possible)
- Prevention of future account creation
7. Legal Compliance
Networkli complies with all relevant child safety laws and regulations, including:
- COPPA (Children's Online Privacy Protection Act)
- FOSTA-SESTA (Allow States and Victims to Fight Online Sex Trafficking Act)
- Regional and national child protection regulations
- International child safety standards
- Ongoing compliance with evolving legislation
8. Staff Training
All Networkli team members receive training on:
- Identifying and reporting CSAM
- Child safety best practices
- Legal obligations and reporting requirements
- Trauma-informed response procedures
9. Continuous Improvement
We are committed to maintaining and improving our child safety practices:
- Regular safety audits
- Updates to detection technology
- Collaboration with child safety organizations
- Transparency in our safety efforts
10. Contact Information
Child Safety Contact: dthadd13@gmail.com
General Inquiries: support@networkli.co
