r/snowflake

MCP server for governed AI writeback to Snowflake
▲ 12 r/snowflake+1 crossposts

MCP server for governed AI writeback to Snowflake

Hi folks — I’m one of the builders behind Syntropic.

We just shipped an MCP server that lets AI agents help with controlled table edits for the kinds of Snowflake tables people already edit manually: control tables, mapping tables, budget/forecast tables, spreadsheet ingestion/uploads, etc.

If you wire an agent directly to Snowflake through a CLI today, that gets awkward pretty quickly for this kind of use case:

  • write access is broader than you usually want
  • schema validation is not the same as business-rule validation
  • query history shows what changed, but not the reason for the change
  • downstream workflows may not know an agent-driven edit just happened
  • rolling back a business edit cleanly can be difficult using time travel

What we built is a layer in front of selected tables where the agent reads/writes through a constrained interface instead of issuing raw warehouse writes.

That gives you things like:

  • scoped access only to the tables you expose
  • validation rules enforced on writes, before the data gets to the warehouse
  • required comment metadata on edits
  • versioning and rollback of changes
  • webhooks on every edit so you can trigger dbt / Airflow / Slack / whatever on each agent edit

The MCP App part

We also made the grid UI *render inside Claude chat as an MCP App*, so a user can ask Claude to show them the forecast for March, inspect rows, edit a few cells manually, and review validation errors inline

A few example workflows:

  • “Claude, Joe sent me a CSV — load it into the budget table”
  • “Who last changed this control table, and why?”
  • “Add a validation rule that SKUs in product_mapping must be 8 chars”
  • “Rollback the forecast adjustment from Monday”

MCP App demo: https://youtu.be/eWsu6m2P58M

Curious how others here are approaching this. Are you letting agents write to Snowflake tables at all right now?

u/jaredfromspacecamp — 12 hours ago

is there a better way to track schema changes without silently breaking downstream reports?

we have dbt models pushing schema changes to prod pretty regularly but downstream reports and bi dashboards keep breaking silently. no alerts, just find out when someone complains a week later.

current setup is basic git history + dbt docs but that doesn't catch when a column rename or type change nukes a join in some forgotten looker dashboard. tried adding pre deploy checks with sql fluff but its too static, misses runtime impacts.

our team is small, 4 data engs handling 50+ models across prod/staging. leadership wants zero breakage but manually reviewing every pr is killing us.

anyone got a lightweight way to track this like dbt macros that flag downstream deps, or some schema diff tool that pings slack on breaks open source preferred since budget sucks. What've you seen work at scale without turning into a full ci nightmare?

curious how others avoid this treadmill.

reddit.com
u/Ok_Abrocoma_6369 — 19 hours ago

When to use Streams vs Drlta load?

In what circumstance would you use Streams instead of simply cheking for rows where the timestamp (inserted at/updated at) changed?

Ehy are streams useful, can't you do the same wirh simple delta loading?

reddit.com
u/Sad_Veterinarian_630 — 11 hours ago

How to build a control plane to manage Snowflake Cortex Code Costs

Data engineers using Cortex Code require a single cost view that separates warehouse compute from AI token credits.

Snowflake now gives dedicated usage history for Cortex Code CLICortex Code in Snowsight, and AI SQL usage, but its not enough!

usage charts are super important but we also need guardrails and financial reporting. We've built them all and much more in SeemoreData but you can also build something sufficient by yourself

this blog explains a good methodology you can build on your own that helps you set the control plane for better cortex code management

https://seemoredata.io/blog/how-to-manage-snowflake-cortex-code-cost-without-slowing-engineers-down/

hope this helps the community!

this is the linkedin post (feel free to like and share) -> linkedin post to like and share :)
as always happy to connect directly -> lets connect directly

reddit.com
u/Spiritual-Kitchen-79 — 10 hours ago
▲ 9 r/snowflake+1 crossposts

Snowflake Summit 2026

Anyone attending this year's summit that would like to discuss code rewriting capabilities? Or any topics you are looking forward to?

reddit.com
🔥 Hot ▲ 74 r/snowflake

Couldn’t find a single diagram showing all the Snowflake ingestion paths while studying for the SnowPro Advanced Data Engineer Certification. So I made one.

u/Body-Beginning — 2 days ago

Cortex Code medium blog post

Just published Part 1 of my new series: How to Build an Enterprise-Grade Skill MD for Cortex Code

In this post, I break down the foundation for designing a production-ready Skill MD setup with an enterprise mindset — not just a quick demo. If you're working with Snowflake Cortex, AI-assisted development, or trying to make code generation workflows more structured and scalable, this might be useful.

Would love to hear your thoughts and feedback.

Read here: https://medium.com/@srivathsan.v91/how-to-build-an-enterprise-grade-skill-md-for-cortex-code-part-1-of-3-61fa09a34771

u/srivve — 8 hours ago

How to retrieve Secrets in Snowflake Notebook in Workspaces

Hello, does anyone have experience with accessing secrets in Snowflake Notebook in Workspaces?

I'm trying to access a username/password secret for a MySQL connection to migrate some data across into Snowflake. I have created and added the External Access Integration and I can see the Secret is loaded into the Service. However, the '_snowflake' and 'streamlit' modules do not seem to exist for Notebooks in Workspaces. I don't see any documentation regarding accessing secrets in this new Notebook editor.

I would greatly appreciate any help and if you can also point me to the corresponding documentation where this is explained. Thanks!

u/WesternTonight2736 — 1 day ago

Looking for Big Data Engineer SME

We are hiring for a Subject Expert Role in Data Engineer SME (Snowflake, AWS, DBT)

Experience Required : 3-5 years

Remote | Full Time | EST Working hours | Pay: ₹14-18 LPA
About the Opportunity
We are seeking a battle-tested Data Engineer SME with deep expertise in Snowflake, data modelling, SQL/Python, and a flair for creating EdTech video content and teaching.
What We're Looking For

  • 3-5 years in data engineering, with hands-on AWS, Snowflake, DBT & AWS Databricks experience
  • Expert in data modelling (star schema, dimensional, etc.)
  • Advanced SQL querying and optimization
  • Python mastery (Pandas, ETL scripting, PySpark)
  • Proficient in AWS ecosystem: S3, Glue, Lambda, Redshift, Airflow, dbt for pipelines
  • Comfortable shooting EdTech videos (tutorials, demos) and leading live teaching/workshops

Why Join Us?

  • Full-time remote— work-life balance
  • Blend engineering with teaching: Shape future data pros through content

Looking for Indian talent with immediate to 7 days notice period .Interested folks please DM.

reddit.com
u/Medium-Bar-9290 — 1 day ago

CoCo SDK

Hi!

I know Snowflake announced the expansion of Cortex Code today and they mentioned the SDK. This is going to be a huge unlock. Does anyone know where/how I can get access to this?

reddit.com
u/Perfect-Cricket6506 — 2 days ago
▲ 2 r/snowflake+1 crossposts

DataFrey: MCP server for Snowflake with text-to-SQL

I’m a data scientist and I find it hard to use Claude Code for SQL because of the lack of DB context. so I made yet another database MCP server! only Snowflake support for now.

I had to reconnect with nature after reading native Snowflake MCP setup docs so for my server I’ve made a nice CLI wizard to set up DB connection and install the Claude Code plugin: MCP + skill - you can ask it like `/db write dbt model to rank leads`.

It also has a `plan` tool for complex questions. when you ask a blurry question, it triggers a separate text-to-SQL agent that uses 1. (kinda) RAG for your schema (along with some values) that builds during DB connection (if you agree) 2. subagents to explore your data. 3. planning. This is what Snowflake Cortex is supposed to do, but when I try it, it never finds the right tables.

Database-as-MCP sounds like a security nightmare, but I put a lot of effort into making it safer. I’d appreciate any thoughts on the secure design. by default, CLI asks for select permissions on all schemas, not just information_schema. I’m convinced that it’s impossible to write good SQL without peeking into the data. maybe it's a hot take - share your thoughts!

Everything is free and hosted by me, but rate-limited. In the future, I want to charge for planning calls above the limit. I have a bunch of ideas on how to make a smarter text-to-SQL, so I want to keep this part closed-source. I’ll open-source more though - it’s just deployed as a monolith now.

docs.datafrey.ai
u/datafrey — 2 days ago

Snowflake Adaptive Warehouses are in public preview - my take

Snowflake Adaptive Warehouses are now in public preview. We tested out them out and think they are fantastic! (but dont offset all the engineering challenges)

You are more than welcome to read my latest blog about it --> https://seemoredata.io/blog/snowflake-adaptive-compute-warehouse-optimization/

That being said, although they solve (in some cases not all) the challenge of compute sizing...they don't completely remove the engineering problem. there are different decisions and different configurations that you still need to decide on and therefore the problem is not completely solved it just moved.

anyways, as always feel free to connect on linkedin --> https://www.linkedin.com/in/yanivleven/
if you feel like it give a like to the linkedin post about the blog --> https://www.linkedin.com/posts/yanivleven_snowflake-dataengineering-finops-share-7451981246001459200-4ONq?utm_source=share&utm_medium=member_desktop&rcm=ACoAAALvtzwB_CbAlsdiwFIwnfAr0dPMesH9I0M

Hope you enjoy the read :)

u/Spiritual-Kitchen-79 — 3 days ago

AMA: We benchmarked the new Adaptive Warehouses

Hey folks,

I'm the CTO at Espresso AI and rather than generate LLM slop content, we actually benchmarked what these new warehouses offer: https://espresso.ai/post/snowflake-adaptive-warehouse-benchmarks

Snowflake is pretty opaque with what's going on, and the benchmarks are a bit rough to try and get this out to folks quickly, but it seems like they deliver better performance (both throughput and latency), but not necessarily cost savings unless you were already paying a premium for performance.

This didn't make it into the post, but the oddest thing about them is that it seems like these warehouses do have a minimum billing period, but it is minute-aligned, i.e. if you issue a query at :55s and :10 seconds later, you get billed for 2 minutes.

Rather than rehash the entire post, I'll just leave that short blurb here and answer any questions folks might have about the benchmarks we ran.

- Alex Kouzemtchenko

u/kuza55 — 3 days ago
▲ 7 r/snowflake+1 crossposts

Snowflake Cortex (CoCo) CLI vs 10TB of Data. Here is what happened.

Most AI agents are tested on toy data (clean, verified datasets). Here is what happened when Cortex Code was hit with 55.8 billion rows:

  • The Win: It understands the Snowflake "secret menu" (Bloom filters, pruning).
  • The Surprise: It built a multi-channel dbt project without being told the connections.
  • The Difference: General LLMs know SQL syntax. CoCo knows the Snowflake platform.

If you’re just using AI for syntax, you’re missing the point. The value is in the native platform intelligence.

Read our full review here:
https://www.capitalone.com/software/blog/snowflake-cortex-code-cli/?utm_campaign=coco_ns&utm_source=reddit&utm_medium=social-organic

u/noasync — 3 days ago

Has anyone whitelist IP addresses for Azure services? Right now, Snowflake raises an error when Copilot Studio or PowerBI is making a request to it.

Hello,

I was wondering if anyone has an experience or familiar with whitelisting the Azure services (PowerApps, PowerBI, and Copilot Studio). Thanks to using Chrome's Inspect -> Networking, I found that this whitelisting is necessary.

I used this website for 'Azure Service tag' and downloaded a json file, I believe 'AzureConnectors' tag is the correct one but this seems to have over 40 different IP addresses that are associated with it.

Is anyone familiar with this by any chance? It reads that the ip addresses change time-to-time so do we have to set up a daily/manual job to update the entire ip addresses everytime in Snowflake?

u/pcgoesbeepboop — 3 days ago

CREATE OR ALTER AND INSERT OVERWRITE as a solution for keeping time travel data for tables with frequent schema changes

Hi,
So we had some workflows running with create or replace on tables. The tables schema change from time to time so everytime new data arrives i will just create or replace the tables also it's small data. Nevertheless it's a production workflow. To keep my code/sql as simple as possible and to improve the workflow so people can historic versions of the tables i changed this workflow to:
CREATE OR ALTER TABLE

INSERT AND OVERWRITE

What do you think anything against this solution? Just asking because this CREATE OR ALTER seems more of a use-case for initial bootstrapping of tables as i can see in snowflake doc.

reddit.com
u/BudgetSea4488 — 4 days ago

What should i learn now - AI like clause gonna replace our jobs?

have been working as snowflake developer and admin. where should i focus on to have job for next 10 years. previously worked on Data extract loading too.

will AI tools like clause automize snowflake jobs?

reddit.com
u/Peacencalm9 — 4 days ago

Jump from DE to Solutions Engineer?

Currently working as a Senior Data Engineer with 5 YoE getting a chance to work for Snowflake. The new role seems less technical and more on the sales side to work on POCs which i have some experience with. I am feeling underpaid at my current role and the new position at snowflake is a huge pay bump. While its something i would still like to work on i feel i will lose my technical skills if i take it up. Would an internal switch be possible after a few years in this case? Also is it safe in this market to be working in the sales division?

reddit.com
u/Ok_Illustrator_816 — 5 days ago

Snowflake Native Ingestion Methods Compared: When to Use Each

Hey folks,

Our team at Estuary works with customers all the time to enable their Snowflake pipelines, and we've seen the gamut of how these ingestion tradeoffs play out, from obliterating credits to underestimating the engineering effort needed to configure and monitor everything across pipelines. Based on this experience, we put together a comparison guide to evaluate each Snowflake ingestion method; how it works, what it costs, and when it makes sense for your use case.

Ultimately, we believe that a well-designed Snowflake data stack applies different ingestion methods for different data streams, and ensures the latency actually matches what the workflow needs. If you're looking for a framework to audit your existing Snowflake pipelines, or you're designing a new pipeline and feeling overwhelmed, we hope this guide helps you out!

estuary.dev
u/Which_Roof5176 — 4 days ago

Best data observability platform tools for data quality monitoring, lineage, and pipeline reliability.

We’re reviewing a bunch of vendors in the data reliability software space and trying to narrow things down. Quick thoughts so far:

monte carlo: strong enterprise presence, broad coverage across warehouses and bi tools, very polished but can feel heavy and expensive.

Bigeye: legacy stack support, decent anomaly detection, seems solid for teams not on modern data stack.

elementary: tbh stands out if you're running dbt, they seem to also have Python support. it’s deeply dbt native and seem easier to operationalize. great visibility into data health, freshness, and lineage without overwhelming onboarding. the AI agents look promising, setup is straightforward, and it feels more aligned with analytics engineering workflows instead of forcing a separate platform mindset.

anomalo: heavy focus on ml based anomaly detection, good for automated insights but may require tuning, and is very enterprise heavy.

metaplane: modern UI, focuses on column level monitoring and anomalies, decent balance between automation and control.

soda: flexible and developer friendly, works well if you want more hands on control.

great expectations: more framework than platform, powerful for custom validations but requires engineering effort to scale properly.

for teams that are dbt heavy and want something opinionated but not bloated, elementary feels less intrusive and more practical compared to some of the bigger enterprise suites.

curious what others would prioritize. Full automation and enterprise coverage, or tighter integration and lower operational overhead?

reddit.com
u/Ok_Abrocoma_6369 — 8 days ago