Skip to main content

Working with Datasets

Datasets are the foundation of how Ask Monty works. Every time you ask a question, Monty creates focused datasets of feedback that are then used to generate your answer.

What is a Dataset?

A dataset is a filtered collection of feedback snippets that match specific criteria. Think of it as a saved search tailored to your question. Example:
Your question: "Summarize mobile app feedback from January"

Monty creates a dataset:
- Name: "Mobile app feedback from January"
- Filters: Source = Mobile, Date = Jan 1-31
- Snippets: 347 matching feedback items
dataset-filters

Dataset Components

Each dataset card displays a descriptive title (like “Checkout bugs • Last 60 days”), the total snippet count, and creation date. Below that, you’ll see all applied filters: time ranges, sources, categories, projects, tags, sentiment, text keywords, metadata fields, and customer segments. This transparency lets you verify that Monty understood your question correctly. Key actions are available on every dataset: you can export everything to CSV for external analysis, create a custom report around that specific dataset, or view the actual feedback snippets. If the filters aren’t quite right, rephrase your question to create a new dataset with better constraints.

Dataset Types

Monty creates different dataset types depending on what you’re asking about. Feedback datasets are the most common, pulling from your standard product feedback across app stores, email, chat platforms like Intercom or Zendesk, reviews, and custom sources. These appear whenever you ask about customer complaints, bugs, feature requests, or general feedback topics. Survey datasets load when you ask about NPS scores, CSAT results, or custom survey responses. They include both quantitative scores and open-ended text responses, letting you analyze ratings alongside qualitative feedback. survey dataset Interview datasets come from structured customer research, containing transcript snippets, interview goals, and participant information. Use these when you want to analyze insights from your customer interview programs. survey dataset External datasets pull from integrated sources like GitHub issues, Jira tickets, or other development tools. These include issue titles, status labels, dates, assignees, and links back to the original items. They’re particularly useful for connecting customer feedback to your team’s work. survey dataset Competitor datasets (if configured) focus on what customers say about competitors, featuring comparison mentions and references to competitor features. Web search datasets surface news coverage, online discourse, and public perception through web page excerpts and publication dates. survey dataset

Working with Multiple Datasets

Many questions require Monty to create and compare multiple datasets.

Comparison Queries

"Compare feature requests from 2024 vs 2023"

→ Dataset 1: Feature requests in 2023
→ Dataset 2: Feature requests in 2024
→ Comparative analysis using both datasets

Segment Analysis

"How does mobile feedback differ from web?"

→ Dataset 1: Mobile feedback
→ Dataset 2: Web feedback
→ Comparison summary
survey dataset

Complex Multi-Dataset Questions

"Compare iOS and Android performance feedback
to competitor mentions"

→ Dataset 1: iOS performance feedback
→ Dataset 2: Android performance feedback
→ Dataset 3: Competitor performance mentions
→ Three-way comparison

Dataset Actions

Exporting Datasets

Export any dataset to CSV for external analysis:
  1. Click the dataset menu (three dots)
  2. Select “Export to CSV”
  3. The exported CSV file will be saved to your browser’s downloads folder
The CSV includes:
  • Feedback content/title
  • Source and creation date
  • Categories, projects, tags
  • Sentiment scores
  • Custom metadata
  • Customer details (name, email, etc.)

Reusing Datasets

Within a conversation, Monty automatically reuses datasets when appropriate:
You: "Summarize checkout feedback from last month"
→ Creates "Checkout feedback • Last 30 days" dataset

You: "Show me specific examples"
→ Reuses the same dataset (no need to recreate)

You: "Now chart that by sentiment"
→ Still using the same dataset
This makes follow-up questions faster and ensures consistency.

Dataset Best Practices

Before diving into analysis, check that your dataset has sufficient data. While 500+ snippets enable robust insights and 100-500 provides good coverage, anything below 50 snippets might need broader filters. Dataset size directly affects the reliability of patterns Monty can identify. After Monty creates a dataset, take a moment to verify the filters match your intent. Check that date ranges are correct, the right sources are included, and categories align with what you asked for. If something looks off, rephrase your question to create a more accurate dataset. When comparing time periods, use equivalent ranges for fair comparison. Full months work better than rolling windows like “last 30 days” versus “previous 30 days,” which might include different day-of-week distributions that skew results. Remember that filters define what’s included and excluded. Filtering to “bugs” means you won’t see feature requests; filtering to “enterprise” excludes SMB feedback. Narrower isn’t always better—sometimes you need breadth to understand the full picture. For complex analysis beyond what Monty provides, export datasets to CSV and use Excel for pivot tables, your BI tools for custom visualizations, or statistical analysis software for deeper investigation.

Dataset Limitations

Time Range Limits

Some dataset types have restrictions:
  • External datasets (GitHub/Jira): Maximum 60 days
  • Web search: Shows recent results
  • Feedback datasets: No hard limit
If you request data outside these ranges, Monty will use the maximum available and let you know.

Data Availability

Datasets can only include what’s been ingested:
  • Feedback must be imported into the system
  • Surveys must be created and run
  • Integrations must be connected
  • Historical data depends on when ingestion started

Performance Considerations

Very large datasets (10,000+ snippets) may:
  • Take longer to process
  • Hit token limits in text responses
  • Be sampled for analysis
Monty will handle this automatically but mention if sampling occurs.

Troubleshooting

When a dataset returns zero results, the most common culprits are time ranges that are too narrow, sources that haven’t imported data, or misspelled categories and tags. Monty will suggest adjustments when this happens. Try expanding your time range, removing some constraints, or checking that your filters reference correctly spelled categories. If you get fewer results than expected, broaden your scope by extending the time period, using less specific categories, making keywords more general, or including additional sources. Sometimes filters that seem right are actually excluding relevant data. When a dataset’s filters don’t match your intent, rephrase your question with more explicit constraints. Be specific about the time range, sources, categories, and any other filters you want applied. Sometimes a small change in phrasing helps Monty better understand what you’re looking for.
Next: Surveys & Interviews →