1. SCOPE AND PURPOSE

The purpose of this AI Policy is to establish a comprehensive governance framework for the development, deployment, operation, monitoring, and maintenance of Artificial Intelligence systems developed by Imentiv Inc. This Policy is designed to ensure that all AI capabilities are utilized in a transparent, ethical, and accountable manner, safeguarding users from unintended harms, discriminatory practices, privacy violations, or inaccurate outputs. It reinforces Imentiv AI’s commitment to building and maintaining trustworthy AI solutions aligned with globally recognized standards and best practices and ethical guidelines for reliable AI. Furthermore, the Policy ensures full adherence to applicable legal and regulatory requirements such as the GDPR, India’s Digital Personal Data Protection (DPDP) Act 2023, the CCPA, and other regional data protection laws. By clearly outlining the responsibilities, expectations, and conduct required from all users, administrators, partners, and employees interacting with the platform, this AI Policy supports a culture of accountability, safety, fairness, and continuous improvement across all AI-driven activities.

This Policy governs all activities, systems, data interactions, and stakeholders involved in the development, deployment, access, and use of Imentiv AI and its associated SaaS products. This policy shall be read in conjunction with the Terms of Use, Data Processing Agreement, Privacy Policy Imentiv for Imentiv AI Products and Solutions.

2. IMENTIV AI SYSTEMS

This Policy covers all AI-related technologies, components, and subsystems that power the Imentiv AI platform, including but not limited to:

  • AI Models & Algorithms: All proprietary, custom-trained, fine-tuned, or third-party algorithms used for prediction, recommendation, classification, emotional analytics, generative tasks, or automated decision-making.
  • Machine Learning Pipelines: End-to-end ML workflows involved in data ingestion, feature engineering, model training, validation, optimization, monitoring, and lifecycle management.
  • Generative AI Capabilities: Text, image, video, speech, or multimodal generative features embedded into the SaaS product and related tools.
  • AI-Enabled Automation Tools: Rule-based engines, orchestration modules, automation scripts, workflow systems, and AI-driven triggers.
  • Emotional & Behavioural Analytics Modules: Components that analyse sentiment, emotion, engagement, tone, or behavioural patterns using AI or ML techniques.
  • Predictive Intelligence Systems: Forecasting engines, scoring models, risk analysers, and real-time prediction components.
  • Security, Monitoring & Audit Systems: AI systems used for anomaly detection, fraud monitoring, performance analytics, fairness evaluation, and compliance auditing.

3. USERS & CUSTOMERS

This Policy applies to all categories of individuals or entities that interact with Imentiv AI products or solutions, directly or indirectly, including:

  • Registered Users: Individuals who create accounts, use the web portal, mobile apps, dashboards, analytics tools, or AI-powered features.
  • Enterprise & Organizational Clients: Businesses and institutions using Imentiv AI for internal operations, integrations, automation, analytics, or customer services.
  • Developers & API Consumers: Individuals or organizations accessing Imentiv AI through APIs, SDKs, developer dashboards, or integration interfaces.
  • System Administrators & Account Managers: Users who manage organizational settings, permissions, security controls, and computational resources.
  • End Users Impacted by AI Outputs: Any individual whose data or experience is affected by automated decision-making, predictions, or AI-generated content

All such users are expected to comply with this Policy and ensure responsible use of Imentiv AI services.

A. Imentiv AI Personnel

This Policy applies to all individuals responsible for the handling, development, review, testing, deployment, or governance of AI systems, including:

  • Employees & Full-Time Staff: Developers, data engineers, data scientists, product managers, QA teams, and operations personnel.
  • Consultants & Contracted Professionals: External specialists involved in design, development, audits, research, or advisory work related to AI systems.
  • Third-Party Vendors & Affiliates: Service providers contributing to cloud infrastructure, model hosting, data preprocessing, annotation, or support services.
  • Security & Compliance Personnel: Teams responsible for ensuring adherence to ethical, legal, regulatory, and internal compliance requirements.
  • Support & Customer Success Teams: Personnel interacting with customers and managing AI-related queries, issues, or escalations.

All personnel must adhere to ethical AI development practices, privacy standards, and security protocols outlined in this Policy.

B. Third-Party Integrations

This Policy extends to all external systems, datasets, and technologies that interact with Imentiv AI, including:

  • External AI Components & Pre-Trained Models: Third-party or open-source models integrated into the platform or used for fine-tuning or reference.
  • Cloud Infrastructure & Hosting Services: Compute, storage, networking, security, or monitoring systems provided by cloud service partners.
  • Analytics & Data Processing Frameworks: External libraries, tools, and engines used for data analysis, visualization, reporting, or model enhancement.
  • API-Based Third-Party Services: External APIs utilized for authentication, enrichment, communication, or data verification.
  • Integration Partners & App Ecosystems: Any platform or software integrated with Imentiv AI through connectors, plugins, or API bridges.

All third-party systems must comply with this Policy, undergo risk evaluation, and adhere to required security and data protection standards before integration.

4. GUIDING PRINCIPLES

Imentiv AI is committed to building responsible, safe, and trustworthy AI systems. All products, technologies, and workflows adhere to globally recognized principles of ethical AI development and deployment. These principles guide how we design, train, validate, monitor, and operate our AI systems.

4.1 Transparency

Imentiv AI ensures that users understand when and how AI impacts their interactions with the platform.

  • Clear AI Disclosure: Users will be explicitly informed whenever content, insights, or recommendations are generated by AI rather than human contributors.
  • Interface-Level Transparency: UI/UX elements will reflect when AI is involved in decision-making, automated scoring, emotional analytics, or generative processing.
  • Model Documentation: Imentiv AI may publish model cards, documentation, architecture notes, permitted use-cases, limitations, and known risks to improve clarity and understanding.
  • Traceability & Explainability: Where required and technically feasible, users may receive explanations or reason summaries behind AI-generated outputs to support accountability and informed decision-making.
  • Legal Compliance in Transparency: The system will avoid “black-box” operations in areas where transparency is legally or contractually mandated.

4.2 Fairness & Non-Discrimination

Imentiv AI is committed to ensuring that AI systems operate equitably across demographic groups and do not propagate harmful or discriminatory biases.

  • Bias Testing & Validation: Models may undergo structured fairness evaluations to detect bias patterns during training, testing, and post-deployment monitoring.
  • No Harmful Processing of Protected Attributes: Sensitive attributes such as race, caste, religion, political beliefs, health data, disability, sexual orientation, and similar identifiers will not be intentionally used for decision-making unless strictly necessary, explicitly consented, and legally permitted.
  • Ongoing Dataset Improvements: Training datasets and model weights are periodically updated to reduce systemic biases, improve representation, and enhance fairness across contexts.
  • User Feedback Mechanisms: Channels will be provided for users to report perceived biases, unfair treatment, or discriminatory outputs, which will be investigated and addressed promptly.
  • Ethical Dataset Sourcing: All data used for model development is sourced ethically, lawfully, and without exploitation.
  • Customer Compliance Requirement: Customers and users are prohibited from using the platform to perform discriminatory profiling, exclusionary targeting, or harmful decision-making.

4.3 Privacy & Data Protection

Respect for user privacy is central to all AI operations at Imentiv AI.

  • Strict Privacy Controls: All personal and sensitive data processed within AI systems input data, generated output, and training data is handled per the highest privacy standards.
  • Purpose Limitation: Data is used only for disclosed, legitimate, and authorized purposes consistent with the Privacy Policy and Data Processing Agreement (DPA).
  • Data Minimization & Anonymization: Where feasible, personal data is anonymized, pseudonymized, masked, or aggregated to minimize exposure and protect user identity.
  • User Ownership: Users retain control over their data. Processing and model usage occur in accordance with consent, contractual obligations, and regulatory requirements.
  • No Covert Profiling: The platform does not create covert, behavioral, or psychological profiles of users unless explicitly requested by the user or mandated for the core service functionality.
  • Compliance with Global Privacy Laws: Data processing aligns with GDPR, DPDP Act (India), CCPA/CPRA, and other applicable privacy regulations.

4.4 Security

Imentiv AI implements robust technical and organizational controls to safeguard AI systems, data, and users.

  • Security-by-Design: All AI models and datasets are developed, deployed, and maintained following industry-standard security principles.
  • Comprehensive Security Controls may include:
    • encryption at rest and in transit
    • strong access controls with RBAC/ABAC
    • session-level and platform-level audit logs
    • secure isolation of model training and inference environments
    • protections against prompt injection, model corruption, and adversarial attacks
    • data validation and sanitization layers to prevent data poisoning
  • Regular Security Assessments: Vulnerability scanning, penetration testing, threat modelling shall be conducted as needed.
  • Incident Management: Any security incident involving AI systems is promptly investigated and remediated under the company’s incident response process

4.5 Human Oversight

Imentiv AI strives to ensure that humans remain accountable for decisions supported or influenced by AI systems.

  • Human-in-the-Loop (HITL): The platform supports workflows where human operators review, validate, or override AI-generated insights.
  • Human-in-Command: For mission-critical or sensitive operations humans must retain final control and decision-making authority.
  • Override Capability: Users and administrators may correct, disable, or challenge AI-generated outputs wherever appropriate.
  • Continuous Monitoring: Human observers evaluate model performance and intervene when anomalies, inaccuracies, or risks arise.

4.6 Reliability & Safety

Imentiv AI systems are designed with the intent to perform consistently, accurately, and within the boundaries of their intended use.

  • Model Validation: All models are validated for accuracy, stability, robustness, and relevance before deployment.
  • Monitoring & Drift Detection: Imentiv AI monitors its AI Systems for data drift, concept drift, accuracy decline, or deviations in prediction patterns over time.
  • Fail-Safe Mechanisms: Imentiv AI may restrict, modify, or deactivate AI features that pose risks, show degraded performance, or behave unexpectedly.
  • Stress & Load Testing: AI systems undergo reliability testing to ensure stable performance under varying real-world conditions.
  • Safety Filters: Mechanisms are in place to prevent the generation of harmful, offensive, or unsafe outputs.

4.7 Ethical Use

Imentiv AI promotes responsible, lawful, and socially beneficial uses of AI technologies.

  • No Harm Principle: Users must not use AI systems to harm, exploit, mislead, or manipulate individuals or communities.
  • No Impersonation: AI-generated content may not be used to impersonate real individuals, institutions, or entities without lawful authorization and explicit consent.
  • Prohibited Misuse: The platform must not be used for surveillance, harassment, deepfakes, misinformation, disinformation, political manipulation, or any fraudulent activity.
  • Social Responsibility: AI usage should contribute positively to individuals, organizations, and society, avoiding practices that undermine trust, dignity, or well-being.
  • Compliance with Laws: All AI activities must adhere to applicable global, regional, and sector-specific regulations and ethical frameworks.

5. ACCEPTABLE USE OF AI FEATURES

Users may access and utilize Imentiv AI’s capabilities solely for lawful, legitimate, and authorized business purposes consistent with this Policy, the Terms of Service, Privacy policy and any applicable agreements. Acceptable use is limited to activities that enhance operational efficiency, improve productivity, support data-driven decision-making, and responsibly integrate AI insights into organizational workflows.

5.1. Users may leverage Imentiv AI's analytical and prediction engines for insights that support strategic and operational objectives, including:

  • Sentiment & Emotional Analysis: Understanding user emotions, tone, sentiment shifts, and engagement patterns in text, voice, or communication data, for permitted business scenarios such as customer service improvement or feedback evaluation.
  • Behavioural Insights: Generating patterns, correlations, behavioural segmentation, or usage insights to identify trends, optimize user journeys, or understand customer lifecycle behaviour within ethical and lawful boundaries.
  • Predictive Analytics & Trend Modelling: Forecasting demand, predicting churn, identifying risk signals, analysing future patterns, or supporting decision frameworks using validated predictive models and business-approved datasets.
  • Intelligence Reporting: Producing high-level summaries, anomaly detection alerts, operational metrics, or aggregate analytics dashboards for organizational monitoring and decision-making.

All analytical outputs must be interpreted responsibly, with human review where appropriate.

5.2. Users may employ Imentiv AI’s automation and productivity features to streamline processes, reduce manual effort, and enhance workflow efficiency, including:

  • Automation Workflows: Implementing AI-driven triggers, workflow automations, and rule-based or ML-based task execution for internal operations, customer support, or content processing.
  • Recommendation Systems: Personalizing content, suggesting next actions, optimizing configurations, recommending resources, and supporting decision-making based on contextual analysis.
  • Data Processing, Categorization & Summarization: Automating repetitive tasks such as data extraction, classification, enrichment, normalization, summarization, clustering, or entity recognition.
  • Contextual Understanding Tools: Using AI systems that interpret intent, extract meaning, classify information, or understand domain-specific context to support search, retrieval, conversational tasks, or knowledge discovery.
  • Content Generation for Internal Use: Drafting internal documents, templates, reports, or operational content where AI-generated material serves as an assistive tool and is subject to human oversight.

5.3 Integration via API, SDKs & Webhooks- Organizations may integrate Imentiv AI into their approved systems, applications, and workflows through provided APIs, SDKs, and automation interfaces:

  • Embedding AI Insights into Authorized Applications: Integrating analytics, emotional intelligence indicators, summaries, and recommendations into websites, apps, CRMs, or internal platforms, provided they comply with contractual use limitations.
  • Developing Organizational Dashboards: Creating analytics dashboards, BI tools, or internal monitoring systems that visualize AI metrics, predictions, or operational data.
  • Enhancing Customer-Facing or Internal Workflows: Using AI to improve service delivery, personalize experiences, streamline internal processes, or augment employee tasks subject to proper data governance and user consent requirements.
  • Secure Automation: Triggering notifications, actions, or workflow events from AI outputs using secure webhook connections and authentication controls

All integrations must follow Imentiv AI’s API documentation, rate limits, security practices, and data handling standards.

5.4 All acceptable uses of Imentiv AI must consistently remain:

  • Lawful: Compliant with applicable laws, regulations, and industry requirements regarding data protection, privacy, security, and automated decision-making.
  • Ethical: Respectful of individuals, communities, and organizations, avoiding manipulative, discriminatory, invasive, or harmful practices.
  • Aligned with Documented Capabilities: Used only within the functional boundaries described in product documentation, model capability notes, and limitations.
  • Non-Exploitative: Ensuring AI outputs are not misused to exert undue influence, mislead stakeholders, or exploit vulnerable groups.
  • Consistent with Contractual Terms: Adhering to the Master Service Agreement (MSA), Terms of Use, Privacy Policy, Data Processing Agreement (DPA), and all other binding contractual obligations.

6. PROHIBITED USE

To ensure safe, lawful, and ethical operation of the platform, users must not use Imentiv AI for any purpose that violates this Policy, applicable laws, regulatory requirements, or ethical standards. The following activities are strictly prohibited. Violation of these terms may result in immediate suspension, access restrictions, termination of services, and reporting to relevant authorities where required.

6.1 Harmful, Illegal, or Deceptive Purposes: Users are strictly prohibited from using Imentiv AI in any manner that causes or attempts to cause harm, unlawful activity, or deceptive practices, including:

  • Deepfake or Synthetic Identity Creation: Generating hyper-realistic synthetic content (audio, video, image, or text) designed to impersonate real individuals without explicit, verifiable authorization.
  • Non-Consensual Impersonation: Creating content, responses, or communication that falsely represents an individual, organization, public authority, or institution.
  • Unauthorized Surveillance or Monitoring: Using AI features to observe, track, or infer personal behaviours, locations, or activities without legal authority, consent, or legitimate purpose.
  • Disinformation, Manipulation, or Influence Operations: Creating or distributing misleading, false, politically manipulative, or socially harmful content, including propaganda, fake news, or electoral interference.
  • Facilitating Illegal Activities: Using the platform for fraud, scams, harassment, cyberattacks, identity theft, or any unlawful purpose.

6.2 Discrimination, Harassment, or Abuse: The platform must not be used to create or facilitate abusive, discriminatory, or harmful behaviour, including:

  • Generating Defamatory, Hateful, or Obscene Content: Producing content that is abusive, hateful, harassing, violent, sexually explicit, defamatory, or intended to degrade or harm individuals or groups.
  • Sensitive Attribute Profiling: Inferring, predicting, or profiling individuals based on protected or sensitive characteristics such as race, caste, ethnicity, religion, political beliefs, union membership, sexual orientation, gender identity, genetic data, health information, or disability status.
  • Targeted Abuse or Harassment: Using AI outputs to intimidate, threaten, shame, or harass individuals.
  • Exploitation of Vulnerable Groups: Using AI to manipulate or exploit minors, elderly individuals, or vulnerable populations.

6.3 Unauthorized or Unethical Data Processing: Users must not upload, process, or submit data to Imentiv AI in violation of privacy rights, intellectual property rights, or legal and ethical standards, including:

  • Processing Data Without Consent or Legal Rights: Uploading personal, confidential, proprietary, or copyrighted data without the necessary permissions, lawful basis, or authorization.
  • Processing Sensitive or High-Risk Data Without Approval: Including, but not limited to, biometric identifiers, medical records, financial data, or health information, unless explicitly permitted under contract and compliant with applicable laws.
  • Reconstruction or Reidentification Attempts: Attempting to de-anonymize, re-identify, or derive personal identities from aggregated or anonymized outputs.
  • Collection of Data for Covert Purposes: Using AI features to harvest or scrape data from systems, individuals, or third-party services without authorization.

6.4 Model Tampering, System Abuse & Security Violations: Users must not manipulate, compromise, or abuse the platform’s models, infrastructure, or security mechanisms, including:

  • Reverse Engineering or Unauthorized Extraction: Attempting to extract model parameters, architecture details, weights, training data, or proprietary algorithms.
  • Prompt Injection, Model Corruption, or Adversarial Attacks: Attempting to exploit or manipulate AI behaviour using malicious prompts, adversarial inputs, or vulnerability probing.
  • Using AI Outputs to Train External Models: Reusing, exporting, or feeding Imentiv AI outputs into external training pipelines or models without explicit written permission.
  • System Overloading or Unapproved Stress-Testing: Intentionally sending excessive requests, load-testing, or attempting to disrupt service performance outside permitted usage limits, including rate-limit abuse.
  • Circumvention of Security Controls: Attempting to bypass authentication, access controls, usage quotas, safety mechanisms, or content filters.

6.5 High-Risk or Regulated Use Cases (Unless Explicitly Approved): Users must not deploy, rely on, or integrate Imentiv AI for high-risk applications without express written consent from Imentiv AI and compliance with relevant regulatory requirements. Prohibited high-risk scenarios include:

  • Medical or Clinical Decision-Making: Diagnosis, treatment recommendations, or health-critical decision systems.
  • Legal, Regulatory, or Compliance Determinations: Interpreting laws, offering legal advice, drafting binding interpretations, or substituting licensed legal professionals.
  • Financial Credit, Lending, or Eligibility Decisions: Automated scoring or decisions affecting creditworthiness, loan approvals, insurance eligibility, or financial risk evaluations.
  • Safety-Critical Industrial Controls: Systems involving autonomous machinery, robotics, transportation, utilities, energy grid management, or physical safety operations.
  • Emergency or Life-Critical Scenarios: Situations where AI errors could result in injury, loss of life, or severe operational disruptions.
  • Employment and Recruitment Screening: Use of AI to analyze candidates during recruitment (e.g., video interviews) or to monitor employee emotional states. Users operating in the European Economic Area (EEA) or other jurisdictions with specific AI regulation must obtain explicit approval from Imentiv AI Legal to ensure compliance with 'High Risk' AI system requirements before deploying for these purposes.

6.6 Enforcement & Remedial Actions- Imentiv AI reserves the right to:

  • monitor usage for compliance,
  • immediately restrict or block access for suspected violations,
  • suspend or terminate user accounts,
  • remove or disable problematic content,
  • report illegal activities to law enforcement or regulatory authorities, and
  • pursue legal remedies for damages or breach of contract.

Users are responsible for ensuring ongoing compliance with all prohibited-use restrictions.

7. DATA GOVERNANCE & AI TRAINING

Imentiv AI implements structured, transparent, and secure data governance practices to ensure that all data used for AI operations including training, inference, monitoring, and evaluation remains lawful, ethical, secure, and compliant with contractual and regulatory obligations.

7.1 Data Sources: Imentiv AI ensures that all data used for model development and enhancement is ethically sourced and legally authorized.

  • Lawful & Ethical Acquisition: Training and evaluation datasets are obtained only from legal, licensed, and ethically compliant sources, including publicly available data, commercial data providers, open-source datasets, or consent-based contributions.
  • Customer Data Restrictions: Customer proprietary data including content, metadata, logs, or uploaded files is never used for training or fine-tuning models unless the customer provides explicit written consent through a signed agreement or opt-in mechanism.
  • No Unauthorized Scraping or Harvesting: The platform does not engage in unauthorized scraping of third-party services, social networks, or websites to obtain training data.
  • Dataset Quality & Documentation: All datasets undergo review for accuracy, relevance, completeness, and lawful usage rights. Where applicable, dataset documentation (datasheets or lineage records) is maintained.
  • Bias, Safety & Appropriateness Checks: Training sources are assessed for harmful, biased, or inappropriate content that could compromise model performance or ethical standards.

7.2 Data Minimization: Imentiv AI adheres to the principle of collecting and processing only the data necessary for intended AI functions.

  • Purpose Limitation: Data is processed strictly for the purposes disclosed in product documentation, privacy policies, contracts, and user consent forms.
  • Minimal Scope Collection: AI models use only the minimal data required to achieve valid technical or operational outcomes avoiding unnecessary data attributes or excessive scope.
  • Controlled Feature Engineering: Only features relevant to model objectives are used, and sensitive attributes are excluded unless explicitly required, legally permitted, and authorized by the user.
  • Periodic Review of Data Needs: Imentiv AI periodically evaluates whether any dataset or attribute remains necessary and removes or masks data that is no longer required.

7.3 Anonymization, Privacy Protection & Security: Imentiv AI employs strong privacy and security controls across all stages of the AI lifecycle.

  • Anonymization & Pseudonymization: Wherever possible, personal data is anonymized, pseudonymized, aggregated, or masked to reduce identification risks before being used for model training or analytics.
  • Customer Data Isolation: Customer data is isolated by tenant and environment ensuring that one customer’s data cannot be accessed by another. Multi-tenant safeguards prevent data leakage or cross-customer model influence.
  • Secure Storage & Access Controls: Based on the security policies, AI datasets use techniques such as encryption at rest and in transit, role-based access controls, zero-trust network policies, and audit logging.
  • Data Integrity & Validation: Input data is validated and sanitized to prevent data poisoning, injection attacks, or ingestion of malicious training content.
  • Compliance with Privacy Laws: All data processing complies with GDPR, DPDP Act (India), CCPA/CPRA, LGPD, and other applicable privacy regulations.

7.4 Retention, Storage Lifecycle & Deletion: Imentiv AI maintains clear procedures for how long AI data is stored and how it is securely disposed of.

  • Defined Retention Schedules: AI logs, intermediate datasets, training/evaluation versions, and model artifacts are retained only for the durations defined in internal retention policies or regulatory requirements.
  • Purpose-Based Retention: Data may only be retained for as long as necessary to fulfil the purpose for which it was collected, including security audits, performance monitoring, regulatory compliance, or contractual obligations.
  • Secure Deletion: When data reaches end-of-life, it is deleted or anonymized using secure and irreversible methods that comply with international data destruction standards.
  • User-Initiated Deletion Requests: Users may request deletion of personal data in accordance with applicable privacy laws and contractual terms. Upon verification, Imentiv AI will process deletion requests across training, backup, and inference systems (unless legal obligations require retention).
  • Model Retraining Considerations: When customer data must be deleted, related training artifacts or model variants influenced by that data will be evaluated for retraining, removal, or replacement as required.

8. USER RESPONSIBILITIES

All users of the Imentiv AI platform are expected to adhere to the following responsibilities to ensure safe, lawful, and ethical use of the system:

8.1 Lawful and Ethical Use: Users must ensure that all activities performed on the platform comply with applicable laws, contractual obligations, and ethical standards.

Users are responsible for ensuring that their use of Imentiv AI does not infringe upon the rights, privacy, or safety of any individual or entity.

8.2 Verification of AI Outputs: Users must independently review, validate, and confirm the accuracy, relevance, and appropriateness of AI-generated outputs before relying on them for any decision-making. AI outputs should not be treated as definitive facts; users must apply human judgment and domain expertise.

8.3 Consent and Authorization for Data: Users are responsible for ensuring that all data uploaded, shared, or processed through the platform has been lawfully obtained.

Users must secure the necessary permissions, consents, licenses, and rights before submitting any personal, sensitive, proprietary, or third-party data. Specifically, where the data includes facial geometry, voice prints, or other inputs defined as 'Biometric Identifiers' or 'Biometric Information' under applicable laws (including but not limited to the Illinois Biometric Information Privacy Act or GDPR), the User warrants that they have obtained the required written releases and informed consent from all data subjects prior to processing such data through Imentiv AI. Users may not upload data that violates intellectual property, confidentiality agreements, or privacy obligations.

8.4 Account Security and Access Control: Users must safeguard their login credentials, API keys, access tokens, and other authentication mechanisms. Sharing credentials with unauthorized individuals, intentionally or negligently, is prohibited. Users must immediately notify Imentiv AI of any suspected unauthorized access, credential compromise, or account misuse.

8.5 Responsible Use in High-Risk Contexts: Users must not rely solely on AI outputs for decisions that may impact legal rights, financial outcomes, medical assessments, safety-critical operations, or any other high-risk domain. All such decisions must involve appropriate human oversight, professional judgment, and independent verification.

8.6 Compliance with Industry-Specific Regulations: Users are responsible for understanding and complying with all regulatory requirements applicable to their business, sector, or jurisdiction (e.g., financial compliance, data protection laws, healthcare regulations, government standards). Where required, users must implement additional safeguards, audits, or procedural controls before using AI-generated content or outputs.

8.7 Prohibited Sharing and Redistribution: Users may not redistribute, republish, or commercialize AI outputs in ways that violate licensing terms or intellectual property rights. Users must ensure that sharing outputs with third parties does not expose confidential data or violate privacy laws.

8.8 Reporting Misuse and System Concerns: Users should promptly report any suspected misuse, security vulnerabilities, harmful outputs, or ethical concerns encountered while using the platform.

Cooperation with investigations or remediation efforts by Imentiv AI is required.

9. MODEL PERFORMANCE, LIMITATIONS & DISCLAIMERS

Imentiv AI aims to provide reliable and high-quality AI outputs; however, users must understand the inherent limitations of machine-learning technologies. The following disclaimers govern all use of the platform:

9.1 No Guarantee of Accuracy: AI-generated outputs may contain inaccuracies, omissions, outdated information, or factual inconsistencies. Users acknowledge that the Imentiv AI platform analyzes behavioral signals (such as facial muscle movements, vocal intonation, and text sentiment) which serve as proxies for emotion. These outputs represent an inference of emotional expression and are not a definitive medical, psychological, or factual determination of a human subject's internal emotional state.

The platform does not warrant that outputs will always be correct, complete, reliable, or suitable for a specific user purpose. Predictions, classifications, and insights generated by the system are probabilistic and may not reflect real-world conditions.

9.2 Non-Deterministic Behaviour: AI systems can produce different responses to similar or repeated inputs due to their probabilistic nature, ongoing model updates, system optimization, or contextual variations. Users should not expect consistent or identical outputs across sessions or use cases.

9.3 Not a Substitute for Professional Judgment: The platform is not designed or intended to replace professional expertise, certified advice, or regulated decision-making frameworks. AI outputs must not be treated as authoritative guidance for:

  • Medical or clinical diagnoses, treatment recommendations, or patient risk assessments
  • Legal interpretation, compliance advice, contract evaluation, or regulatory analysis
  • Financial, credit, or lending decisions, including underwriting or scoring
  • Safety-critical operations, including engineering decisions, industrial controls, or emergency responses
  • Users must apply independent human judgment and seek qualified professional advice where required.

9.4 Model and Data Limitations: The system may reflect biases present in training data or external sources. Outputs may not consider real-time events unless explicitly stated. Performance may vary across languages, domains, and complexity of user queries.

9.5 User Reliance: Users rely on outputs, insights, and recommendations at their own discretion and risk. Imentiv AI is not liable for decisions, actions, or outcomes resulting from user reliance on AI-generated content. Users are responsible for verifying critical information before acting on or disseminating AI outputs.

9.6 Service Availability & Performance Variability: Model performance may be affected by network conditions, system load, scheduled maintenance, updates, or third-party dependencies. Imentiv AI does not guarantee uninterrupted service, real-time responsiveness, or specific performance levels unless otherwise contractually agreed.

10. THIRD-PARTY AI TOOLS & EXTERNAL SERVICE PROVIDERS

Imentiv AI may integrate or rely on third-party AI models, cloud infrastructure, APIs, data processors, and other external service components to deliver certain features and system functionalities. To ensure responsible and compliant integration, the following principles apply:

10.1. Vendor Security, Privacy & Compliance Due Diligence: All third-party service providers undergo security, privacy, and compliance assessments before integration. Evaluations consider the vendor’s certifications, data protection standards, regulatory adherence (e.g., GDPR, SOC 2, ISO 27001), and operational resilience. Imentiv AI maintains contractual safeguards to ensure that third-party tools adhere to equivalent or higher standards of security and ethical conduct.

10.2 Data Handling by Third Parties: Third-party providers may process limited, purpose-specific data required for performance, model inference, optimization, or infrastructure operation. Data shared externally is minimized to what is strictly necessary, following principles of data minimization and privacy-by-design. Sensitive or confidential customer data is only shared where legally permissible and contractually protected. Any transfer of personal data to external providers occurs in accordance with applicable data protection laws, cross-border transfer mechanisms, and contractual controls.

10.3 User Obligations Regarding Third-Party Services: Users acknowledge that some platform features depend on external tools and that their use may be subject to additional terms, privacy policies, or licensing conditions imposed by those providers. By using applicable features, users agree to comply with such third-party terms to the extent they apply. Users are responsible for reviewing any third-party terms that affect their organization’s regulatory or operational environment.

10.4 Impact of Third-Party Limitations or Failures: Imentiv AI is not responsible for service disruptions, latency, degraded performance, or inaccuracies arising from third-party outages, model limitations, or operational failures. Availability of certain features may change based on vendor updates, API lifecycle changes, depreciation schedules, or unforeseen external issues. Imentiv AI may replace, modify, or terminate third-party integrations at its discretion to maintain service quality, compliance, or security.

10.5 Continuous Monitoring & Governance: Integrated third-party components are periodically reviewed for compliance, reliability, and risk exposure. Any provider failing to meet required standards may be removed or replaced. Users will be informed of material changes to third-party integrations that significantly impact data processing or system behaviour.

11. MONITORING, AUDIT & CONTINUOUS IMPROVEMENT

Imentiv AI maintains an ongoing commitment to ensuring that all AI systems models, algorithms, datasets, and related components operate ethically, safely, and in accordance with legal, regulatory, and organizational standards. The platform uses a combination of automated and manual oversight to maintain system integrity and continually enhance performance.

11.1 Continuous Monitoring of AI Models: AI model behaviour is continuously monitored for accuracy, fairness, stability, explainability, and adherence to intended use-cases. Monitoring includes tracking false positives/negatives, model degradation, unpredictable behaviour, and bias indicators. Imentiv AI performs regular performance benchmarking and validation to ensure models continue to function as designed.

11.2 Automated Detection of Anomalies and Misuse: Automated systems are deployed to identify unusual activity, such as anomalous prompts, excessive API requests, or abnormal output patterns. Built-in alerts notify administrators of potential misuse, adversarial attacks, or suspicious user activity. Monitoring systems help detect:

  • model or data drift
  • prompt injection attempts
  • data poisoning attempts
  • misuse of features for unethical or prohibited purposes

Detected misuse may trigger automated safeguards, temporary feature suspension, or account restriction.

11.3 Periodic Ethical, Technical & Compliance Audits: Internal and external audits are conducted periodically to validate compliance with:

  • AI ethics principles
  • privacy and data protection laws
  • contractual commitments
  • security standards
  • transparency and fairness guidelines

Audit outcomes may result in corrective actions, model fine-tuning, policy revisions, or process improvements. All audit findings are logged and reviewed by designated compliance and AI governance teams.

11.4 Model Updates, Retraining & Lifecycle Management: Models are evaluated regularly for improvement opportunities, including retraining with updated or expanded datasets. Outdated, underperforming, or high-risk models may be:

  • deprecated
  • replaced with more secure or accurate versions
  • decommissioned entirely

When models are updated, Imentiv AI ensures:

  • backward compatibility where feasible
  • transparency about material changes to model behaviour
  • preservation of user experience and system stability

11.5 Continuous Improvement Framework: Feedback from users, audits, incident reports, and monitoring systems informs the platform’s continuous improvement initiatives. Imentiv AI invests in ongoing research, testing, and optimization to enhance accuracy, fairness, usability, and safety. Enhancements are deployed following secure development lifecycle (SDLC) practices including testing, validation, and impact assessment. Lessons learned from monitoring and incident investigations are incorporated into future product updates and governance policies.

12. REPORTING UNSAFE OR HARMFUL AI OUTPUT

Imentiv AI is committed to maintaining the safety, fairness, and reliability of its AI systems. To uphold this standard, users play a critical role in identifying and reporting issues that may compromise ethical or operational integrity. All users are required to promptly notify Imentiv AI of any unsafe, harmful, or suspicious AI behaviour.

12.1 Responsibilities to Report Issues: Users should immediately report any instances of:

  • Inaccurate, misleading, or unsafe AI outputs, especially where such outputs could affect business decisions or user rights.
  • Biased, discriminatory, or unfair model behaviour, including outputs that disproportionately affect individuals or groups.
  • Potential security concerns, such as data leakage, unauthorized access, or suspicious model behaviour indicating vulnerabilities.
  • Misuse or abuse of the platform, including attempts to circumvent safeguards, exploit loopholes, or generate prohibited content.
  • Unexpected model reactions or instability, such as hallucinations, erratic responses, or behaviour inconsistent with the system's intended use.

12.2 Reporting Channels: Users can report issues through the following official channels:

Data Protection Officer:

Name: Nabil Shabab

Email: support@imentiv.ai

In-app reporting (if available): Via the built-in “Report Issue” or “Flag Output” feature.

Customer Success / Account Manager: For enterprise clients with assigned contacts.

All reports should include sufficient detail such as timestamps, input samples, screenshots, or relevant logs (excluding sensitive personal information unless necessary and legally permissible) to allow timely investigation.

12.3 Investigation & Remediation Process: Upon receiving a report, Imentiv AI will: Acknowledge receipt of the report within a reasonable timeframe. Conduct an internal review to validate, categorize, and prioritize the issue. Determine whether the issue is:

  • a model limitation
  • a safety risk
  • a policy violation
  • a technical bug
  • a user-side integration or configuration error

Take appropriate corrective actions, which may include:

  • model refinement or retraining
  • adjusting safety filters
  • patching vulnerabilities
  • temporarily disabling affected features
  • issuing user guidance or clarifications

12.4 User Cooperation: Users agree to:

  • Provide necessary information to assist in diagnosis and remediation.
  • Refrain from further use of any feature known to be unsafe until Imentiv AI completes its review.
  • Follow any interim recommendations, safety notifications, or required updates communicated by Imentiv AI.

12.5 Commitment to Safety: Imentiv AI treats all safety-related reports as high priority and is committed to:

  • Transparent communication about confirmed issues and fixes where appropriate.
  • Implementing safeguards to prevent recurrence.
  • Continually improving detection, response, and mitigation processes.

13. ENFORCEMENT, VIOLATIONS & PENALTIES

Imentiv AI enforces this Policy to maintain platform integrity, ensure ethical use, and protect users, data subjects, and the public. Any violation of this Policy—whether intentional, negligent, or accidental may result in corrective or disciplinary actions proportionate to the severity, nature, and impact of the breach.

13.1 Identification of Violations: A violation may include, but is not limited to:

  • Misuse of AI features for harmful, illegal, or prohibited purposes.
  • Attempts to bypass safety filters, perform disallowed data processing, or exploit system vulnerabilities.
  • Unauthorized access, reverse engineering, scraping, or model extraction attempts.
  • Uploading unlawful, sensitive, or unpermitted data.
  • Repeated disregard for operational guidelines, safety warnings, or corrective instructions.
  • Violations may be detected through automated monitoring, audits, security logs, user reports, or compliance reviews.

13.2 Progressive Enforcement Actions: Depending on the severity, intent, recurrence, and potential harm, Imentiv AI may take one or more of the following actions:

  • Warning or Advisory Notice: Issued for minor, first-time, or unintentional violations.
    May include guidance, education, or required corrective steps.
  • Feature or Usage Restrictions: Temporary or permanent limitations on certain features, API calls, or model access. Imposed when a user’s behaviour poses moderate risk or indicates misuse of specific capabilities.
  • Account Suspension: Temporary suspension of access for repeated or serious violations. May occur if there is suspected malicious intent, security risk, or high-risk misuse. Access may be restored after investigation and remediation.
  • Account Termination or License Revocation: Applied for severe, intentional, or harmful violations. Includes permanent removal of account access and revocation of all licenses or entitlements. Imentiv AI may refuse future service to offenders.
  • Notification of Authorities: If a violation involves criminal activity, regulatory non-compliance, data breaches, or unlawful surveillance, Imentiv AI may report the incident to:
    • law enforcement
    • regulatory bodies
    • affected third parties (as required by law)

13.3 Financial, Legal & Contractual Consequences: Imentiv AI may seek compensation for damages, financial losses, or remediation costs arising from misuse. For enterprise clients, penalties may include contractual remedies, indemnification claims, or early termination of agreements. Violations may result in legal action where necessary to protect the platform, users, or data subjects.

13.4 Investigation Process: All suspected violations are subject to a formal internal investigation. Users may be required to provide additional information or logs (excluding sensitive personal data unless lawfully permissible). Enforcement decisions consider:

  • user intent
  • scale of impact
  • risk to individuals or systems
  • recurrence of similar incidents
  • cooperation during investigation

13.5 Right to Appeal: Users may dispute or appeal enforcement actions by contacting support@imentiv.ai. Appeals will be reviewed by the compliance or governance team, and outcomes communicated within a reasonable timeframe.

13.6 Zero-Tolerance for High-Risk Misuse: Imentiv AI maintains a strict zero-tolerance policy for:

  • malicious activities
  • exploitation of vulnerabilities
  • unlawful surveillance
  • discrimination using AI outputs
  • creation of deceptive or harmful content
  • Such actions may result in immediate termination and legal escalation.

14. POLICY UPDATES

Imentiv AI is committed to maintaining an AI Policy that evolves in step with advancements in technology, emerging regulatory frameworks, and best practices in ethics, safety, and data governance. To ensure that the Policy remains current, effective, and aligned with global standards, Imentiv AI may revise, update, or expand its provisions as needed.

14.1 Reasons for Policy Amendments: This Policy may be updated or amended to reflect:

  • Advancements in AI Technologies
    • Introduction of new AI features, models, or analytical capabilities.
    • Modifications to algorithms, ML pipelines, or training methodologies.
    • Updates required to address emerging risks, vulnerabilities, or industry-recommended practices.
  • Changes in Legal, Regulatory, or Industry Requirements
    • New or updated AI regulations (e.g., EU AI Act, NIST frameworks, global privacy laws).
    • Evolving contractual obligations, certification standards, or data protection rules.
    • Court rulings, enforcement actions, or public policy changes affecting AI governance.
  • Safety Enhancements & Ethical Considerations
    • Improved safeguards to mitigate bias, hallucination, adversarial attacks, or misuse.
    • Updated internal processes for monitoring, transparency, and user protection.
    • Lessons learned from audits, incident reports, or user feedback.
  • Business, Operational, or Infrastructure Changes
    • Adoption of new third-party vendors or cloud infrastructure providers.
    • Updates to internal governance structures, security frameworks, or service offerings.
    • Organizational policy refinements to support scalability, compliance, or performance.

14.2 Notification of Policy Changes: Users will be notified of material changes, such as updates that significantly alter data handling, user rights, obligations, or platform functionality. Notification may occur via:

  • in-platform alerts
  • email communication
  • dashboard announcements
  • policy changelogs accessible within the user portal

14.3 User Responsibility to Review Updates: Continued use of the platform after a Policy update constitutes acceptance of the revised terms. Users are responsible for reviewing updated policies to ensure ongoing compliance with their legal and organizational requirements. Enterprise clients may receive advance notice for changes that impact contractual obligations or integration workflows.

14.4 Version Control & Documentation: Imentiv AI maintains version history and timestamps for all Policy updates. Prior versions may be archived to support transparency, audit needs, or compliance inquiries. Users may request details of historical changes where necessary for regulatory or internal governance purposes.

14.4. Commitment to Transparency: Imentiv AI strives to ensure that all policy updates are communicated clearly, timely, and with sufficient context for users to understand their impact. The goal is to foster trust, accountability, and shared responsibility in the safe and ethical use of AI technologies.

15. COMPLIANCE AND ENFORCEMENT

We reserve the right to suspend access to our AI Solutions at any time if we determine, in good faith, that you have breached our usage policies or if the use of our service could risk or harm, our users, or anyone else. We reserve the right to update this policy as needed to address emerging issues and maintain a safe environment.

16. CONTACT AND OPTIONAL FEEDBACK

We encourage users to use our support channels to provide feedback on AI outputs, which helps us improve the AI’s performance and ensure responsible use. By using our AI chatbot, you commit to following these guidelines and contributing to a positive and respectful community for all users. If you have any questions or feedback about this policy, please contact our support team at support@imentiv.ai.