HockeyStack AI: Security, Privacy, and Responsible Use
At HockeyStack, we leverage AI to deliver go-to-market intelligence to the enterprise while maintaining the highest standards of security, privacy, and ethical responsibility. Our AI-powered services use OpenAI’s foundational models in conjunction with our proprietary analysis and agentic execution framework, ensuring transparency, security, and accuracy.
This document outlines our AI policies, data handling practices, and the measures in place to mitigate risks associated with AI technologies.
1. Data Processing and AI Training
How We Handle Data
All data processing is conducted through OpenAI’s secure infrastructure.
Customer data remains isolated and is not shared with third parties.
OpenAI does not use customer data to train their general AI models.
Data is never used to train models for other customers—each customer’s data remains completely separate.
Infrastructure & Environment
Our AI runs on the same secure environment as our existing product functionality.
No new external processing environments are introduced.
Customer data is not used to train or fine-tune vendor models.
Data Ownership & Control
Customers retain full ownership of any data submitted to AI.
Customers own all AI-generated outputs from their data.
AI functionality is opt-in, ensuring customer control.
2. AI Risk & Security Considerations
We recognize that AI introduces new security and ethical challenges. Below are the key risks associated with AI and how HockeyStack mitigates them.
Potential AI Risks
Lack of transparency: Understanding how AI generates output.
Data breaches or unauthorized access.
Unintended consequences of AI-driven decision-making.
AI hallucinations: Generation of misleading or false information.
Security Measures in Place
Regular AI evaluations and output testing.
Human-in-the-loop review for non-deterministic AI outputs.
Strict internal documentation on AI model testing.
Encrypted data transmission and storage.
Access controls and log monitoring.
3. Ethical AI Usage Guidelines
We are committed to the ethical use of AI and ensuring that AI-generated insights are fair, unbiased, and transparent.
Our AI Ethics Principles
Transparency: AI outputs must be explainable and traceable.
Fairness: AI models must be free from bias and designed for equitable outcomes.
Security: Customer data must remain protected at all times.
Human Oversight: AI is never the sole decision-maker in high-stakes business scenarios.
Preventing AI Misuse
We do not use AI to manipulate or push users toward specific decisions.
No third parties are granted access to AI-generated customer data.
We maintain easy and open feedback channels for AI-related concerns.
4. Addressing AI Hallucinations & False Outputs
While AI models are powerful, they are not infallible. To ensure the integrity of AI-generated insights, we have built mechanisms to identify and correct false outputs.
How We Handle AI Inaccuracies
All AI-generated outputs must be reviewable by a human (human-in-the-loop process).
Customers can report hallucinations or inaccuracies directly to us.
Logging & tracking AI inconsistencies allows for continuous model improvements.
5. AI Security & Adversarial Attack Prevention
We take active measures to prevent malicious attempts to exploit AI systems, such as data poisoning or model inversion.
Our Approach to AI Security
Data integrity monitoring to prevent adversarial poisoning.
Strict API key segmentation by product and feature to prevent cross-contamination.
Usage of authentication layers to block unauthorized access.
6. AI Compliance & Legal Considerations
We ensure that our AI aligns with legal, ethical, and regulatory requirements.
AI Compliance & Legal Framework
Risk and Conformity Assessments: Evaluating AI against security and privacy risks.
Responsible AI Use Policy: Ensuring compliance with ethical AI practices.
AI-Specific Legal Terms: Addressed within our Data Processing Agreement (DPA).
Customer Monitoring & Ownership
Customers have the ability to monitor AI-generated insights.
Customers can choose to disable AI functionality at any time.
7. HockeyStack AI Policy
Our AI policy outlines the best practices and principles that guide our use of AI.
HockeyStack AI Policy
Data Protection & Privacy: All AI interactions respect strict data privacy policies.
No AI Training on Customer Data: Customer data is never used to train or improve global AI models.
Transparency & Explainability: AI insights must be understandable and traceable.
Fair & Unbiased Models: We actively mitigate risks of AI bias.
Security-First Approach: AI implementations follow industry-leading security standards.
8. AI Sustainable Use Policy
AI should be used responsibly to support sustainable and ethical business practices.
HockeyStack AI Sustainable Use Policy
AI should be used to augment human decision-making, not replace it.
AI should be monitored continuously to ensure quality and accuracy.
AI should not be used for deceptive practices, including misleading data manipulation.
AI should follow a privacy-first approach, respecting user consent at all times.
9. Frequently Asked Questions (FAQ)
1. Does HockeyStack’s AI process customer data?
Yes, but customer data remains private, is not shared with third parties, and is never used to train global AI models.
2. Can customers opt out of AI functionality?
Yes, AI features are opt-in, and customers can choose to disable them at any time.
3. Does HockeyStack allow third-party access to AI-generated data?
No, all customer data and AI outputs are kept strictly private.
4. How does HockeyStack prevent AI hallucinations?
We implement human-in-the-loop verification, logging, and customer feedback mechanisms to catch and correct AI inaccuracies.
5. How does HockeyStack protect against AI security threats?
We use encrypted data transmission, strict access controls, authentication layers, and adversarial attack prevention strategies to secure AI functionality.
6. What legal agreements cover AI usage at HockeyStack?
Our Data Processing Agreement (DPA) includes specific AI-related terms to ensure compliance and security.
7. How can customers report AI-related issues?
Customers can report issues via our support channels or directly through security@hockeystack.com.
For any questions regarding AI security, compliance, or usage, please contact us at security@hockeystack.com.
Last updated