Spotter 3 data handling
This article describes the changes in how ThoughtSpot handles your data in Spotter 3.
Security and governance
While previous versions of Spotter only process data, Spotter 3 is able to read user prompts, its responses, and even correct its own mistakes. This allows Spotter to access the same filtered data a user sees in ThoughtSpot, and provide deeper analytical insights, natural language summaries, and “Why” question analysis. To support Spotter 3 security, ThoughtSpot has created a robust data protection architecture to ensure that your organization’s security rules are Spotter’s security rules.
Data protection architecture
Spotter 3 utilizes a security mirroring architecture to ensure data integrity and privacy throughout the analytical process.
- Security parity
-
Spotter strictly honors existing data permissions. If a user cannot see a specific row or column in ThoughtSpot due to row-level or column-level security, that data is never sent to the Large Language Model (LLM).
- Minimalist transmission
-
Spotter never sends entire tables or databases to LLM providers. It only transmits the relevant, filtered query results necessary to answer the specific question that is asked.
- Stateless processing
-
While data context is sent to an LLM for real-time analysis, the LLMs provided by ThoughtSpot do not retain any data after processing.
- Encryption standards
-
All data in transit to the LLM is protected using HTTPS, and LLM-generated data is encrypted at rest within ThoughtSpot’s application databases.
Administrative controls
Administrators can manage data awareness and information persistence using the following controls:
| Control | Description | How to manage |
|---|---|---|
Spotter version |
An instance-level setting. Spotter 3 is disabled by default for existing instances. |
Admin > ThoughtSpot AI > Spotter version. |
LLM |
Your organization can provide their own Azure OpenAI keys. This ensures prompts and data remain within your organization’s self-managed cloud environment. |
For details, see Connect to your Large Language Model (LLM). |
Column exclusion |
Leverage the hidden column feature in data models to prevent specific sensitive fields from ever being shared with an LLM. |
For details, see Hide a column or define a synonym. |
Previous Chats |
An instance-level setting which allows each user’s Spotter conversation history to be accessible to them. |
Admin > ThoughtSpot AI > Spotter 3 capabilities > Enable chat history. |
Data pipeline |
Records all user interactions on the instance in the Spotter Conversations Liveboard. Enabled by default. |
To disable the data pipeline, contact ThoughtSpot Support. |
LLM data retention and storage
ThoughtSpot enforces strict retention and isolation protocols for all LLM-generated data such as summaries, insights, and logs.
- Previous Chats
-
Textual insights generated by Spotter are stored in the conversations for later consumption. These are retained for a default period of six months. Previous chats is an opt-in capability at the instance level. If you opt out, then data is persisted temporarily for up to six hours in the ThoughtSpot layer.
- Monitoring logs
-
Interaction logs are available in the Spotter Conversations Liveboard. Administrators can manually clean this storage or disable the pipeline entirely.
- Data sovereignty
-
ThoughtSpot utilizes region-specific models to honor local data residency laws, ensuring data is processed within the required jurisdiction (for example, EU-based models for EU customers).
- Tenant isolation
-
Data is logically isolated per tenant. For customers requiring physical isolation, Virtual Private ThoughtSpot (VPT) is available upon request.
LLM provider compliance
ThoughtSpot maintains strict contractual agreements with LLM sub-processors to protect enterprise data.
- Zero training policy
-
Your data is never used to train the foundation models of foundational LLM providers.
- Exclusive use
-
Your data is used exclusively to generate insights for your current session and is never co-mingled or shared with other entities.
Implementation considerations
- Standard search limitation
-
If a column is hidden to restrict it from being shared with the LLM, it is also hidden from standard ThoughtSpot Search.
- User experience
-
If sensitive columns are visualized in a Liveboard but hidden from Spotter, the LLM will not have access to those columns, which may result in a suboptimal experience for follow-up questions.