4 Lightning Talks on Practical AI Workflows from Notion, 1Password, MotherDuck & Evidence
2025/04/10How Data Teams Are Using AI to Transform Their Workflows
Four data professionals from MotherDuck, Notion, 1Password, and Evidence shared practical approaches to integrating AI into their daily workflows, demonstrating how artificial intelligence is reshaping the modern data stack.
Using Cursor IDE for Rapid BI Development
Archie from Evidence demonstrated how Cursor, an AI-powered IDE built on VS Code, dramatically accelerates the development of data applications. Unlike traditional chat-based interfaces, Cursor provides comprehensive context about your entire codebase, enabling more accurate code generation.
The key advantages include:
- Automatic awareness of all files and dependencies in your project
- Integration with documentation (like Evidence docs) for enhanced context
- Real-time code generation with diff-style visualization
- Natural language commands for complex tasks
During the demonstration, Cursor successfully generated a complete deep-dive analytics page with multiple components on the first attempt, showcasing its ability to understand both the codebase structure and the specific requirements of BI tools.
Enriching CRM Data with LLMs in Snowflake
Nate from 1Password tackled a common go-to-market challenge: incomplete CRM data. Historically, teams would manually update Salesforce records, with team members updating 20 accounts each morning—a time-consuming and error-prone process.
Using Snowflake's LLM integration, Nate developed an automated approach to classify companies by industry:
Key Implementation Details:
- Model Selection: Llama models provided the best results for industry classification
- Prescriptive Boundaries: Defining 10-15 specific industries rather than letting the LLM choose freely
- Prompt Engineering: Including industry descriptions and definitions for accuracy
- Data Enrichment: Passing company names, domains, and notes to provide context
The solution achieved over 90% accuracy in returning single-word industry classifications, dramatically reducing manual data entry while improving data quality for territory planning and lead routing.
Automating Data Catalog Documentation at Notion
Evelyn from Notion addressed the perpetual challenge of maintaining data catalog documentation. Despite significant investments in data catalog tools, many organizations struggle with incomplete metadata, rendering these tools less effective.
The Documentation Generation Process:
- Context Gathering: Providing SQL definitions, upstream schemas, data types, and internal documentation
- Lineage Awareness: Using generated upstream descriptions to ensure consistency across tables
- Human Review: All AI-generated descriptions undergo review before publication
- Feedback Loop: Table owners can suggest improvements that the LLM incorporates
The system successfully generates table descriptions, column definitions, and example queries, though human oversight remains crucial—especially for nuanced details like date partitioning that could lead to expensive query mistakes.
Streamlining Data Pipeline Development with MCP
Mehdi from MotherDuck showcased how the Model Context Protocol (MCP) revolutionizes data pipeline development. Traditional data engineering involves slow feedback loops between writing, testing, and debugging code against actual data sources.
MCP enables LLMs to:
- Execute queries directly against data sources
- Validate schemas and data types in real-time
- Generate and test DBT models automatically
- Provision data directly in cloud warehouses
The demonstration showed an LLM independently:
- Querying S3 files to understand data structure
- Handling errors (like type mismatches) through iterative testing
- Creating validated DBT staging models
- Loading processed data into MotherDuck
This approach significantly reduces the traditional back-and-forth between code generation and testing, though it requires guidance to follow senior engineering best practices rather than brute-force solutions.
Common Challenges and Best Practices
Trust and Validation
All panelists emphasized the importance of skepticism when reviewing AI-generated outputs. Results often appear reasonable but may contain subtle errors that only domain expertise can catch. The recommendation: always implement human review processes, especially for production systems.
Model Selection Matters
Different models excel at different tasks. While GPT-4 might excel at product specification, Claude often performs better for code implementation. Mistral's Code Stral model specifically targets code generation without unnecessary markdown explanations. Teams should evaluate multiple models for their specific use cases.
Shifting Skill Requirements
AI tools are changing how data professionals spend their time:
- Less time writing boilerplate code: AI handles routine coding tasks
- More time reviewing and validating: Engineers become code reviewers rather than writers
- Focus on patterns and architecture: Understanding the "why" becomes more important than the "how"
- Reduced interruptions: Fewer requests to senior engineers for basic questions
The Junior Engineer Challenge
A surprising challenge emerged around supporting junior team members. As senior engineers become more self-sufficient with AI tools, they may inadvertently provide less mentorship. Teams need to actively ensure junior members receive adequate support and aren't just relying on AI without understanding fundamentals.
Key Takeaways for Implementation
- Start with Sandboxed Environments: Test AI workflows in controlled settings before production deployment
- Provide Rich Context: The quality of AI outputs directly correlates with the metadata and context provided
- Maintain Human Oversight: AI accelerates workflows but doesn't replace the need for expert validation
- Document AI Boundaries: Clearly define what AI should and shouldn't do in your workflows
- Iterate on Prompts: Invest time in crafting effective prompts rather than accepting first results
The consensus among panelists: AI tools are transforming data workflows by eliminating routine tasks and accelerating development cycles. However, success requires thoughtful implementation, continuous validation, and a clear understanding that these tools augment rather than replace human expertise. As data teams adopt these technologies, the focus shifts from manual execution to strategic thinking and quality assurance—ultimately enabling teams to deliver more value in less time.
CONTENT
- How Data Teams Are Using AI to Transform Their Workflows
- Using Cursor IDE for Rapid BI Development
- Enriching CRM Data with LLMs in Snowflake
- Automating Data Catalog Documentation at Notion
- Streamlining Data Pipeline Development with MCP
- Common Challenges and Best Practices
- Key Takeaways for Implementation
CONTENT
- How Data Teams Are Using AI to Transform Their Workflows
- Using Cursor IDE for Rapid BI Development
- Enriching CRM Data with LLMs in Snowflake
- Automating Data Catalog Documentation at Notion
- Streamlining Data Pipeline Development with MCP
- Common Challenges and Best Practices
- Key Takeaways for Implementation
Related Videos

2025-07-26
How to Efficiently Load Data into DuckLake with Estuary
Learn how DuckLake, MotherDuck, and Estuary enable fast, real-time data integration and analytics with modern open table formats, cloud data warehousing, and no-code streaming pipelines.
YouTube

20:44
2025-06-13
What can Postgres learn from DuckDB? (PGConf.dev 2025)
DuckDB an open source SQL analytics engine that is quickly growing in popularity. This begs the question: What can Postgres learn from DuckDB?
YouTube
Ecosystem
Talk

2025-06-12
pg_duckdb: Ducking awesome analytics in Postgres
Supercharge your Postgres analytics! This talk shows how the pg_duckdb extension accelerates your slowest queries instantly, often with zero code changes. Learn practical tips and how to use remote columnar storage for even more speed.
Talk
Sources