Spotter 3 data handling

This article describes the changes in how ThoughtSpot handles your data in Spotter 3.

Security and governance

While previous versions of Spotter only process data, Spotter 3 is able to read user prompts, its responses, and even correct its own mistakes. This allows Spotter to access the same filtered data a user sees in ThoughtSpot, and provide deeper analytical insights, natural language summaries, and “Why” question analysis. To support Spotter 3 security, ThoughtSpot has created a robust data protection architecture to ensure that your organization’s security rules are Spotter’s security rules.

Data protection architecture

Spotter 3 utilizes a security mirroring architecture to ensure data integrity and privacy throughout the analytical process.

Security parity

Spotter strictly honors existing data permissions. If a user cannot see a specific row or column in ThoughtSpot due to row-level or column-level security, that data is never sent to the Large Language Model (LLM).

Minimalist transmission

Spotter never sends entire tables or databases to LLM providers. It only transmits the relevant, filtered query results necessary to answer the specific question that is asked.

Stateless processing

While data context is sent to an LLM for real-time analysis, the LLMs provided by ThoughtSpot do not retain any data after processing.

Encryption standards

All data in transit to the LLM is protected using HTTPS, and LLM-generated data is encrypted at rest within ThoughtSpot’s application databases.

Administrative controls

Administrators can manage data awareness and information persistence using the following controls:

Control Description How to manage

Spotter version

An instance-level setting. Spotter 3 is disabled by default for existing instances.

Admin > ThoughtSpot AI > Spotter version.

LLM

Your organization can provide their own Azure OpenAI keys. This ensures prompts and data remain within your organization’s self-managed cloud environment.

For details, see Connect to your Large Language Model (LLM).

Column exclusion

Leverage the hidden column feature in data models to prevent specific sensitive fields from ever being shared with an LLM.

For details, see Hide a column or define a synonym.

Previous Chats

An instance-level setting which allows each user’s Spotter conversation history to be accessible to them.

Admin > ThoughtSpot AI > Spotter 3 capabilities > Enable chat history.

Data pipeline

Records all user interactions on the instance in the Spotter Conversations Liveboard. Enabled by default.

To disable the data pipeline, contact ThoughtSpot Support.

LLM data retention and storage

ThoughtSpot enforces strict retention and isolation protocols for all LLM-generated data such as summaries, insights, and logs.

Previous Chats

Textual insights generated by Spotter are stored in the conversations for later consumption. These are retained for a default period of six months. Previous chats is an opt-in capability at the instance level. If you opt out, then data is persisted temporarily for up to six hours in the ThoughtSpot layer.

Monitoring logs

Interaction logs are available in the Spotter Conversations Liveboard. Administrators can manually clean this storage or disable the pipeline entirely.

Data sovereignty

ThoughtSpot utilizes region-specific models to honor local data residency laws, ensuring data is processed within the required jurisdiction (for example, EU-based models for EU customers).

Tenant isolation

Data is logically isolated per tenant. For customers requiring physical isolation, Virtual Private ThoughtSpot (VPT) is available upon request.

LLM provider compliance

ThoughtSpot maintains strict contractual agreements with LLM sub-processors to protect enterprise data.

Zero training policy

Your data is never used to train the foundation models of foundational LLM providers.

Exclusive use

Your data is used exclusively to generate insights for your current session and is never co-mingled or shared with other entities.

Implementation considerations

Standard search limitation

If a column is hidden to restrict it from being shared with the LLM, it is also hidden from standard ThoughtSpot Search.

User experience

If sensitive columns are visualized in a Liveboard but hidden from Spotter, the LLM will not have access to those columns, which may result in a suboptimal experience for follow-up questions.


Was this page helpful?