AI Explainability and Trust
This document explains how askEdgi uses AI, what data is processed, how results are generated, and what controls are in place to ensure trustworthy, compliant operations. It is designed for decision-makers responsible for evaluating AI systems within governance and risk frameworks.
How askEdgi Uses AI
askEdgi combines natural language processing with enterprise metadata to convert user questions into analytical operations. The system does not operate as a general-purpose chatbot. It is a constrained, governed assistant designed to work within predefined boundaries.
AI Model Architecture
askEdgi uses a hybrid architecture:
Governance Layer (Internal)
Operates within the OvalEdge platform. This layer interprets user questions, validates permissions, retrieves metadata, generates SQL queries, and applies access controls. It enforces governance policies before any data is processed.
Language Understanding Layer (External):
Uses large language models (OpenAI or Gemini) for natural language interpretation. This layer receives only metadata, summaries, and contextual information. It does not receive raw enterprise data.
What the AI Does Not Do
askEdgi does not learn from enterprise data
It does not train models on customer datasets
It does not retain conversation history for cross-user learning
It does not access data outside the user's permissions
It does not bypass governance controls
Data Flow and Processing
Understanding what data moves through the system is essential for risk assessment.
What Stays Internal
The following data never leaves the secure environment:
Raw enterprise data records
Personally identifiable information (PII)
Sensitive or regulated data values
Table contents (row-level values)
User credentials and authentication tokens
What Is Shared with External AI Providers
Only the following metadata and contextual information are sent to external language models:
Table names and column names
Column descriptions and data types
Semantic summaries and aggregated data profiles (example: "90% of values are between 0 and 100")
Business glossary definitions
Tag associations and stewardship details
User questions in natural language
No raw data values are included in these requests. The external model processes metadata to understand context and generate appropriate responses.
Data Residency by Edition
Public Edition:
Temporary workspaces operate in AWS-hosted environments. Uploaded files and analysis artifacts are stored in encrypted S3 buckets. Files are deleted after session completion per retention policies.
SaaS Editions:
All data remains within OvalEdge's secure AWS infrastructure. Enterprise connectors enforce data residency and access controls. No data is duplicated or moved outside governed boundaries.
On-Prem Edition:
All data remains within the customer's infrastructure. No external AI enrichments or cloud-based processing occur. Metadata analytics run entirely on local systems.
Access Control and Governance
askEdgi enforces access controls at multiple levels to ensure users only see data they are authorized to access.
Role-Based Access Control (RBAC)
Every query validates the user's role and permissions before execution. Users see only datasets and columns they have been granted access to via OvalEdge's catalog governance.
Validation points:
Before adding catalog objects to the workspace
Before executing queries on datasets
Before displaying metadata search results
Before allowing recipe execution
Data Masking
Masked columns remain masked during analysis. If a user does not have permission to view a column in its raw form, askEdgi will not expose it. Masking policies are enforced automatically without user intervention.
Workspace Isolation
Each user operates in a dedicated workspace. Uploaded files, analysis artifacts, and temporary datasets are not shared across users, even within the same organization. This prevents unintended data exposure and ensures audit trails remain clear.
Catalog Permissions
Only datasets with minimum Data Read access appear in the workspace selection modal. Non-permissioned datasets are hidden entirely. Users cannot circumvent catalog governance through askEdgi.
Explainability and Transparency
askEdgi delivers explainability through a layered explanation model that caters to both business and technical users.
Each analysis may include a high-level summary describing the outcome in plain language, followed by detailed technical explanations that outline transformations, filters, joins, and assumptions applied during execution.
Where applicable, executable artifacts such as SQL queries or Python logic are available to support full reproducibility and independent review.
askEdgi provides layered explanations:
Summary: Plain-language result description
Technical Detail: Transformations, filters, joins, assumptions
Reproducibility: Executable SQL, Python, or formula steps
View Generated SQL
Users can view the SQL queries generated by the system. This allows technical reviewers to verify logic, inspect joins, and confirm that filters are applied correctly.
Access method
Result blocks include a "View Code" option. Clicking it reveals the generated SQL for that step.
Edit Generated SQL
If the generated SQL is incorrect or incomplete, users can edit it before execution. This provides a safety mechanism for refining logic while maintaining transparency.
Access method
Result blocks include an "Edit Code" option. Users modify the SQL and re-run the query.
Insights and Summaries
askEdgi generates narrative summaries alongside tables and charts. These summaries explain key observations, trends, and anomalies detected in the data. The text is derived from computed results and metadata context, not from speculative reasoning.
Query Validation
Before executing a query, askEdgi validates:
Dataset availability in the workspace
Column name correctness
Syntax and structure of generated SQL
Compatibility with the target connector
If validation fails, the system prompts the user for clarification or correction.
Read-Only Operations
askEdgi performs read-only analysis. It does not write data back to source systems, modify catalog metadata, or trigger workflows that alter enterprise data. This prevents unintended changes and ensures all operations are safe for exploratory analysis.
Session Context Limits
Conversation context persists within a single session. When a user closes a thread or starts a new conversation, prior context is cleared. This prevents context leakage across unrelated analyses.
Audit Logging
All user actions are logged, including:
Queries submitted
Datasets accessed
Recipes executed
Workspace operations (uploads, removals)
Compute and token consumption
Logs are retained according to enterprise audit policies and support governance reviews.
Limitations and Failure Modes
No AI system is perfect. Understanding where askEdgi may fail helps set appropriate expectations.
Ambiguous Questions
If a question is vague or lacks necessary context, askEdgi may select the wrong dataset, misinterpret intent, or generate incomplete results. Users should provide clear, specific prompts for best results.
Mitigation: Pin datasets to focus analysis. Rephrase questions with explicit column names and filters.
Incorrect SQL Generation
The system may generate syntactically correct SQL that does not match user intent. This occurs when questions are ambiguous or when dataset relationships are not well-defined in metadata.
Mitigation: Review the generated SQL using the "View Code" option. Edit and re-run if necessary.
Dataset Structure Mismatch
Recipes created on one dataset may fail when executed on a different dataset with different column names or structures. This is not a system error but a compatibility issue.
Mitigation: Document expected dataset structures in recipe descriptions. Test recipes on representative data before sharing.
Compute and Performance Limits
Large datasets, complex joins, or resource-intensive queries may exceed workspace capacity. This results in timeouts or incomplete results.
Mitigation: Upgrade workspace containers (Public edition) or contact administrators to adjust resource allocations (Enterprise editions).
Compliance and Regulatory Alignment
askEdgi is designed to operate within regulated environments.
GDPR Compliance
Users retain full ownership and control of data
OvalEdge operates as a data processor, not a data controller
No customer data is used for AI model training
Data subject rights (access, deletion) are supported via platform-level controls
CCPA Compliance
askEdgi does not sell or share personal information
Data is used only for analysis within the platform
Users can request the deletion of uploaded files and workspace artifacts
ISO 27001 and SOC 2
askEdgi inherits compliance controls from AWS and OvalEdge platform architecture:
AES-256 encryption at rest
TLS 1.2+ encryption in transit
Regular vulnerability assessments and penetration testing
Continuous monitoring and audit logging
Industry-Specific Regulations
Organisations in healthcare, finance, or other regulated industries can deploy askEdgi in configurations that enforce additional controls:
On-Prem deployments for full data residency
Metadata-only editions to avoid data movement
Custom masking and access policies via catalog governance
Responsible AI Principles
askEdgi follows responsible AI practices to ensure ethical and accountable use.
Transparency
Users can inspect generated SQL, review recipe steps, and access audit logs. System behavior is documented in detail, and limitations are disclosed openly.
Accountability
askEdgi operates within OvalEdge's governance framework. Administrators control access, configure connectors, and enforce policies. Users are responsible for validating results and using insights appropriately.
Human Oversight
askEdgi assists decision-making but does not replace human judgment. Users review results, confirm accuracy, and apply business context before taking action.
No Autonomous Actions
askEdgi does not execute workflows, send alerts, or trigger downstream processes without explicit user instruction. All operations require human initiation.
Fairness and Bias
askEdgi processes data according to user-defined logic and catalog governance. It does not introduce bias through model training on customer data. However, biases present in source data or business rules are reflected in results. Organizations are responsible for reviewing data quality and governance policies.
Risk Mitigation Summary
Risk
Mitigation
Unauthorized data access
Role-based access controls are validated before every query
Data leakage to external AI
Only metadata and summaries sent to external models
Incorrect results
Generated SQL is reviewable and editable by users
Unintended data modification
All operations are read-only; no writes to source systems
Session data exposure
Workspace isolation ensures users cannot access others' data
Lack of auditability
Comprehensive logging of all queries, executions, and access
Non-compliance with regulations
Edition-specific controls support GDPR, CCPA, ISO, SOC 2
Model hallucination
RAG architecture grounds responses in enterprise metadata
Compute overruns
Spend limits and budget alerts enforce consumption controls
Governance Recommendations
Organizations deploying askEdgi should implement the following practices:
Define Access Policies:
Configure role-based access and data masking rules in the OvalEdge catalog before enabling askEdgi. This ensures users cannot bypass governance through the AI interface.
Establish Usage Policies:
Document acceptable use cases, prohibited actions, and escalation procedures. Communicate these policies to users during onboarding.
Monitor Usage:
Review audit logs regularly. Identify anomalous activity, excessive consumption, or policy violations. Use logs for compliance reporting and governance reviews.
Validate Outputs:
Encourage users to review generated SQL and cross-check results against established KPIs. Implement quality checks for high-impact analyses.
Control Marketplace Access:
In SaaS editions, decide whether to enable Public Marketplace recipe consumption. Policy-based controls allow administrators to restrict external recipe usage.
Test Before Production:
Conduct proof-of-concept deployments to validate behavior, performance, and governance enforcement. Use test data to simulate production scenarios.
Limit On-Prem Scope:
For On-Prem deployments, restrict askEdgi to metadata analytics only. Avoid enabling features that require external AI enrichments or cloud connectivity.
Summary
askEdgi operates as a governed AI assistant within the OvalEdge platform. It enforces access controls, respects data masking, and provides transparency through reviewable SQL and audit logs. Enterprise data remains secure. Only metadata is shared with external AI providers.
Organizations retain full control over data, policies, and usage. Deployment flexibility ensures askEdgi can operate within strict compliance and residency requirements.
Responsible use requires clear policies, regular monitoring, and user training. When deployed correctly, askEdgi accelerates insight generation while maintaining governance and trust.
Last updated
Was this helpful?

