Dataops Role in the Automation of Automation

Dataops Role in the Automation of Automation

We provided in depth coverage of the concept of the Automation of Automation which we see accelerating Data Driven Business and Business Solutions to an entirely new level.  Besides Generative AI (GenAI) it really is the culmination of automations of software development over the last 20-30 years.  When I started this Data and Software game back in the nineties you had to develop pretty much everything.  Now I would say that 80-90% of software that we had to develop back then is now within libraries.  Most of which are open source.  

In addition to that coding has progressed and become incredibly easier by stable reliable software libraries and package managers, there has been the last 20 or so years of I hate to say it …. “Zapier” or “IFTTT” – If this then that solutions upon solutions.  There are hundreds of them now.  

I mean in the Snowflake ecosystem I have watched as far back as 2018 there were maybe 3-5 ELT type tools Stitch (Acquired by Talend), Fivetran (now Fivetran/HVR), and Matillion where really Stitch and Fivetran sort of pioneered the “Connect this to That … NO CODE solution”.  Now, there are at least 30 if not 50++ endless “Connect this to That” data solutions.  (actually I feel like there is 100+ but its a super fragmented market beyond Fivetran with Big Bucks Andresson Horowitz fueling the endless marketing on it.  We have even seen new upstarts like Omnata Data Superheroes come out with incredibly easy and low cost “Connect this to That Data Solutions”

Then to top off that, we had the Hightouches nad the Census AND many others come out with the “Data Activation” concepts which were actually I’m pretty sure built off of my original article around how to build a CDP on top of Snowflake.  When I wrote that it was really not thought to be possible and the tech wasn’t really there.  Now its totally possible and we see significant traction around “Don’t let your friends buy PURE CDPs” marketing and messaging.  (Which I typically agree with but hey – you have to evaluate what is the best business/technical solution for you.  What is even funnier though with this statement is that MANY of the CDP vendors I know have actually moved to having Snowflake as their backend.  

So with that fun start, let’s talk about Dataops Role in the Automation of Automation.  Quite a few years ago, Guy and Kent Graziano were telling me to check out this TrueDataOps stuff and I was like, wtf is this new term.  Like do we have to have 1 more new term.  I mean I still don’t think we need Data Mesh but that’s another story or “thought leadership argument”.

While I’m kind of slow to take on new concepts “sometimes” I now truly see how Dataops facilitates better Data Product Overall Solutions.  Without it, or without dataops.live at this point for our solutions, you can only go so far with your Automation of Automation of Quality and repeatable solutions 

Snowflake Solutions Center by Dataops.live

We wanted to cover one of the largest advances in overall Snowflake Dataops Technology by our partner, Dataops.live.

We have worked for many years with the Dataops.live Team and are incredibly excited to see them help Snowflake deploy by creating the overall Snowflake Solutions Center for powering hundreds and eventually thousands of Snowflake Industry and Solution demonstrations and proof of concepts.  This is huge step forward in leveraging innate Snowflake functionality of Zero-Copy Clone capabilities along with github repo type practices to truly enable Data Product building and testing automation.

Take a quick look at what Dataops is enabled through the new Snowflake Solutions Center.

Snowflake Solutions Center - It is really a Catalog of Solutions for Snowflake Sales Engineers

This is honestly amazing to see how much Snowflake Solutions and Features have grown over the years.  We believe this level of new Dataops automation really is what Snowflake needs to take it to another level of growth in the Data Cloud with quality and automation BUILT INTO Data Products.

Let's take a look at the Solutions Center (Powered by Dataops.live)

Snowflake Solutions Center Catalog of Demonstrations and Proof of Concepts

(This aligns with our viewpoint of the Automation of Automation that we have written extensively about)

The image of the Solutions Center below illustrates now how easily Snowflake Sales Engineers can select existing Demos or Proof of Concept Industry or Vertical based existing solutions.  What is REALLY important here is that this allows Snowflake Sales Engineers to have consistent deployment of solutions to prospects and customers.  It really "accelerates" the art of the possible and the automation of automation.

  • Search and Discover industry solutions - Within the image you will notice that now Sales Engineers can easily search on a catalog of Solutions
  • Deploy tailored solution instances for many different customers
  • HUGE!  Actually continuously monitor and validate solutions (Instead of the issues we would have with live version of the DCR 5, 5.5, 6.  Now you know what you are getting!)
  • In a way this is the Meta of Meta Data.
Snowflake Solutions Center
Snowflake Solutions Center

 

 

Solution Center Itemized "SOLUTION" Homepage and Description

The image below illustrates all the meta data related to a Solution.  This allows for repeatability with BOTH Code and Data for a quality tested and VALIDATED solution.

  • Each solution has both CUSTOMIZED and VALIDATED customer solution scripts
  • The SOLUTION itself can easily be created and tailored to the Snowflake Sales Engineers specific prospect or customer audience
  • This allows better understanding of the technical solutions available to BOTH Account Executives and Sales Engineers.
  • Each Solution is Repeatable and known and includes explanations and Step by Step Detailed Instructions so every level of Snowflake Sales Engineer can learn and demonstrate the unique solution.
Snowflake Solutions Center Homepage
Snowflake Solutions Center Homepage

 

Conclusion:

The Snowflake Solutions Center is honestly a Game Changer to us and we deploy the SAME TYPE of solutions for our customers with Dataops.live.
Get Started with a 15 minute review of how you can change your solutions to become truly Dataops Based Data Products!

We can show you how this new way of thinking and deploying data products will change your business and lives in certain industries!

  • Automate CICD predictably
  • Orchestrate your Data Solutions Data Products

Frank Slootman is retiring from Snowflake

Over the last few minutes, Snowflake communicated during their earnings call/report on 2024-02-28 at 2pm Pacific that our friend Frank Slootman would be retiring from Snowflake and that Sridhar Ramaswamy would be taking over as CEO.  The stock market immediately reacted either to that or the expectations around the earnings call by the stock going down by 20%.  Sridhar came from Snowflake's Neeva acquisition back in 2023.  We will be really interested in how he will lead the company going forward filling in for Big Frank Slootman's shoes.

We at Snowflake Solutions were deeply rooted in the Snowflake Community back in 2019 when Frank took over from Bob Muglia.  Overall, I think most people within the Snowflake Community including Snowflake Employees, Partners, and Customers would credit Frank for excellent overall leadership from 2019 until now taking Snowflake to the most successful Software IPO ever in September 2020.  Frank along with many of his team he brought over from his previous CEO jobs like Mike Scarpelli, John Sapone, and many others immediately went to work in 2019 working towards the Snowflake growith and positioning for a successful IPO.

We will add more later on our take of the Sloot Snowflake Era but we wish Frank good health and good luck in his retirement.  He was always good to us and we respect his decision on retirement.

The Ultimate Guide to Getting Started with Snowflake

Introduction:

Are you ready to revolutionize your data management and analytics? In this guide, we'll take you through a step-by-step process to help you get started with Snowflake, from watching live demos to participating in hands-on labs.

1. Watch a Live Demo: Unveiling the Power of Snowflake:

The first step in your Snowflake journey is to witness its capabilities in action. Head over to the Snowflake website and explore the "Resources" or "Learn" section. Keep an eye out for upcoming webinars or live demos conducted by Snowflake experts. These sessions provide a firsthand look at how Snowflake can transform your data management.

To get started, simply register for a live demo and mark your calendar. During the session, experts will showcase key features and functionalities, giving you a sneak peek into the world of Snowflake.

2. Try Snowflake for Free: Dive into the Cloud:

Ready to experience Snowflake hands-on? Navigate to the Snowflake website and find the "Free Trial" or "Sign Up for Free" button. The registration process is straightforward – provide basic information, set up a username, and create a secure password.

Once you're registered, log in to the Snowflake platform using your credentials. Congratulations! You now have access to a world-class data warehousing solution.

3. Explore the Interface: Navigating Snowflake with Ease:

With your account set up, take a moment to explore the Snowflake interface. Familiarize yourself with key components such as Worksheets, Object Browser, and Warehouses. Snowflake boasts a user-friendly design, and you can find detailed documentation on their website to guide you through each feature.

4. Load Sample Data: Hands-On Learning:

To truly grasp Snowflake's capabilities, consider loading sample data into the platform. Look for sample datasets provided by Snowflake within the documentation or the platform itself. This hands-on experience will give you a practical understanding of how data is managed and queried within Snowflake.

5. Participate in a Virtual Hands-On Lab: Guided Learning Experience:

Want a more guided approach to learning? Check if Snowflake offers virtual hands-on labs or workshops. These interactive sessions provide step-by-step guidance on key exercises, allowing you to deepen your understanding of Snowflake's functionalities.

Keep an eye on announcements on the Snowflake website, community forums, or other communication channels for information about upcoming virtual labs.

6. Join the Snowflake Community: Connect and Learn:

Snowflake's community is a treasure trove of knowledge and experience. Connect with other users and experts, ask questions, share your insights, and learn from real-world use cases. The Snowflake Community is an invaluable resource for expanding your Snowflake expertise.

7. Refer to Documentation and Tutorials: In-Depth Knowledge:

For in-depth understanding, explore Snowflake's comprehensive documentation and tutorials. The official documentation covers everything from specific features to best practices and troubleshooting tips. It's your go-to resource for mastering Snowflake at your own pace.

Introducing Snowflake Native Apps:

If you are looking to dive deep into the world of Snowflake Native Apps, we are pioneers and top experts in this field. We cover all you need to know about Native Apps in this section of our website.

Snowflake Cortex: Democratizing Generative AI for All

Introduction:

Generative AI is a powerful new technology that is transforming the way we work. But to fully capitalize on its potential, it's critical that everyone — not just those with AI expertise — is able to access and use it. Snowflake Cortex Generative AI is a new, fully managed service that makes it easy for anyone to use generative AI to analyze data and build AI applications. With Snowflake Cortex, you can:

  • Access industry-leading AI models, including large language models (LLMs) and vector search functionality.
  • Build custom LLM-powered apps in minutes.
  • Maintain flexibility and control over your data.

What is Generative AI?

Generative AI is a type of artificial intelligence that can create new content, such as text, images, and code. It is trained on massive datasets of existing data, which allows it to learn the patterns and relationships that exist in that data. Once trained, a generative AI model can be used to generate new content that is similar to the data it was trained on.

Snowflake Cortex Generative AI:

Snowflake Cortex democratizes generative AI by making it easy for anyone to use, regardless of their technical expertise. It does this by providing:

  • A set of serverless functions that can be used to access and use generative AI models with a single line of SQL or Python.
  • Pre-built user interfaces for popular generative AI tasks, such as text summarization, sentiment detection, and translation.
  • A secure and governed environment for developing and deploying generative AI applications.

Benefits of Using Snowflake Cortex:

There are many benefits to using Snowflake Cortex for generative AI, including:

  • Ease of use: Snowflake Cortex is designed to be easy to use, even for those with no prior experience with AI.
  • Scalability: Snowflake Cortex is a cloud-based service, so it can scale to meet the needs of any organization, regardless of size.
  • Security and governance: Snowflake Cortex provides a secure and governed environment for developing and deploying generative AI applications.
  • Cost-effectiveness: Snowflake Cortex is a cost-effective way to use generative AI, as it is priced based on usage.

Use Cases for Snowflake Cortex:

Snowflake Cortex can be used for a wide variety of use cases, including:

  • Data analysis: Snowflake Cortex can be used to analyze data and extract insights that would be difficult or impossible to obtain using traditional methods.
  • AI application development: Snowflake Cortex can be used to develop and deploy AI applications that can automate tasks, make predictions, and generate new content.
  • Customer service: Snowflake Cortex can be used to build customer service chatbots that can answer questions and resolve issues quickly and efficiently.
  • Sales and marketing: Snowflake Cortex can be used to build sales and marketing applications that can generate leads, qualify prospects, and predict customer churn.

Conclusion:

Snowflake Cortex is a powerful new tool that is democratizing generative AI by making it easy for anyone to use, regardless of their technical expertise. It offers a wide range of benefits, including ease of use, scalability, security and governance, and cost-effectiveness. Snowflake Cortex can be used for a wide variety of use cases, including data analysis, AI application development, customer service, and sales and marketing.

To learn more about Snowflake's Snowday announcements, check out our latest article on the past Snowday event of 2023.

To learn more about Snowflake Cortex and how you can use it to build with AI, please visit the Snowflake website.

Part 2: A Deep Dive into the Snowflake Native Apps Ecosystem

Introduction:

In Part 1 of our series, we introduced you to the world of Snowflake Native Apps, highlighting their transformative role in data analytics and management. Now, in Part 2, we'll take a closer look at Snowflake's Native App ecosystem, exploring the range of tools available and how they cater to different data roles and responsibilities. Let's dive in.

Exploring Snowflake Native Apps Portfolio:

These are a few of the Native apps currently accessible in the Snowflake Native Apps Marketplace. For a detailed overview of the available options, please visit this webpage.

  • Cost optimizer for Snowflake: Understanding your credit leakages by (NTT Data)
  • Test Automation for Snowflake: Automate the data quality monitoring by (NTT Data)
  • Dark Data Discovery: Identify and monetize your unused data by (NTT Data) 
  • Aero Health Check: Get instant cost & latency optimization recommendations by aero
  • Health Check for Snowflake: Summarized report for Snowflake’s consumption and performance by Flurry Insights
  • License Patrol: Machine learning model for software license utilization tracking by Elementum
  • Ops Center: Open, Free, Snowflake Warehouse Cost and Operations Management by Sundeck

Overview of Snowflake Native Apps

Snowflake's Native Apps are a family of specialized applications, each tailored to address specific data tasks and roles. Here are some key Native Apps that are essential for different data professionals:

1. Streamlit - Streamlit is a Python framework for building data apps. It is open source, easy to use, and provides a variety of features that make it a good choice for building Snowflake Native Apps. It is easy to use, flexible, performant, and integrates seamlessly with Snowflake. 

  • Ease of use: It provides a simple and intuitive API for building data apps.
  • Flexibility: It can be used to build a wide variety of data apps, from simple dashboards to complex machine learning applications.
  • Performance: They can handle large datasets and complex workloads with ease.
  • Integration with Snowflake: This makes it easy to build Snowflake Native Apps that can access and process Snowflake data.

2. Data Science: Snowflake's data science tools enable data scientists and AI practitioners to access and analyze data within Snowflake, making it easier to train machine learning models. With a secure and integrated environment, data scientists can streamline their workflows and collaborate effectively.

3. Data Sharing and Collaboration: Snowflake's native apps for data sharing and collaboration foster teamwork by simplifying data sharing both within and outside the organization. These apps ensure data accuracy and security while promoting efficient collaboration.

Data Science and AI Integration

1. Snowflake Data Science: This Native App allows data scientists to perform data science tasks directly within Snowflake. It supports popular machine learning libraries, making model development and deployment a seamless process. By leveraging Snowflake's data-sharing capabilities, data scientists can access, analyze, and train models using Snowflake data.

2. Snowflake's Data Marketplace: This data-sharing platform offers access to a vast array of external data sources, enabling data scientists to enrich their analyses and models with diverse datasets. The integration of these datasets into Snowflake's ecosystem accelerates the development of AI and machine learning models.

Data Sharing and Collaboration

Efficient data sharing and collaboration are pivotal in today's data-driven business landscape. Snowflake's Native Apps are equipped with features that facilitate these critical aspects:

1. Snowflake Data Exchange: Data Exchange is a marketplace for data providers and consumers. Data providers can publish datasets for others to access, while data consumers can easily integrate these datasets into their analytics processes. This platform encourages collaboration among organizations, opening the door to valuable data insights.

2. Snowflake's Secure Data Sharing: Secure Data Sharing enables organizations to securely share data with their partners, customers, and other stakeholders. With fine-grained access controls and robust security measures, data sharing becomes a streamlined and trustworthy process.

Introducing our own Native App: Snoptimizer

Unlock the power of Snowflake Native Apps with Snoptimizer, created by our visionary founder and 4x Snowflake Data Superhero, Frank Bell. With 1.6 years of dedicated experience in Native Apps, we're the pioneers you need to bring your ideas to life.

Sign up today for a personalized demo and let us transform your data world.

Don't miss out on the expertise that only Snoptimizer can offer. Visit our website to explore the endless possibilities. Your data deserves the best!

We are happy to help optimize your Snowflake account or help you build your own Snowflake Native App.

Conclusion:

Snowflake's Native Apps ecosystem is a testament to the platform's commitment to simplifying data management, analytics, and collaboration. The integrated development environment, specialized data science tools, and data-sharing capabilities cater to a wide range of data professionals, making Snowflake a versatile and powerful platform.

In Part 3 of our series, we will explore the future of Snowflake Native Apps, discussing emerging trends and predictions that are shaping the landscape of data analytics. Stay connected to discover how Snowflake is evolving to meet the data needs of today and tomorrow.


Part 3 will look ahead at the evolving landscape of data analytics and the trends shaping the future with Snowflake Native Apps.

Snowflake’s Snowday 2023

Introduction:

Snowflake Snowday is always a highly anticipated event in the data world. This year's event was no exception, with a slew of new product announcements and updates that will further extend the capabilities of the Data Cloud.

One of the biggest announcements at Snowday 2023 was the launch of Cortex, Snowflake's new generative AI offering. Cortex makes it possible to build and deploy AI models entirely within the Data Cloud, without the need for any specialized infrastructure. This is a major breakthrough for data professionals, as it makes AI more accessible and affordable for businesses of all sizes.

Snowflake also announced a number of updates to its Snowpark environment for traditional machine learning. Snowpark now supports a wider range of ML frameworks and languages, making it easier for data scientists to get started with ML on Snowflake.

In addition, Snowflake announced support for Iceberg Tables, a new open source format for data tables that is designed to be scalable and efficient. Iceberg Tables are a great option for storing and managing large datasets in the Data Cloud.

Finally, Snowflake announced updates to its Horizon data governance tool and a new Snowflake Notebook. These updates make it easier for data professionals to manage and govern their data in the Data Cloud.

Snowflake Snowday Announcements 2023:

Cortex:

Cortex is Snowflake's new generative AI offering. It makes it possible to build and deploy AI models entirely within the Data Cloud, without the need for any specialized infrastructure. Cortex provides a variety of features to help data scientists get started with generative AI, including:

  • A library of pre-trained generative AI models
  • A drag-and-drop interface for building and deploying AI models
  • A variety of tools for training and evaluating AI models
  • The ability to share and collaborate on AI models with other data scientists

Snowpark:

Snowpark is Snowflake's environment for traditional machine learning. It makes it possible to run ML workloads directly on Snowflake's data warehouse. Snowpark now supports a wider range of ML frameworks and languages, including:

  • TensorFlow
  • PyTorch
  • scikit-learn
  • R

Iceberg Tables:

Iceberg Tables are a new open-source format for data tables that is designed to be scalable and efficient. Iceberg Tables are a great option for storing and managing large datasets in the Data Cloud. Snowflake now supports Iceberg Tables, making it easier for customers to take advantage of this new format.

Horizon:

Horizon is Snowflake's data governance tool. It helps data professionals to manage and govern their data in the Data Cloud. Snowflake announced a number of updates to Horizon at Snowday 2023, including:

  • New features for managing data access and permissions
  • New features for auditing data usage
  • New features for managing data lineage

Snowflake Notebook:

Snowflake Notebook is a new interactive notebook environment for Snowflake. It makes it easy to explore and analyze data in the Data Cloud. Snowflake Notebook includes features such as:

  • Code completion
  • Syntax highlighting
  • Integration with Snowflake's data warehouse

Introducing Snoptimizer:

Are you curious to learn more about Snoptimizer, our newest Snowflake Native App.

Conclusion:

Snowday 2023 was a major event for the Data Cloud. Snowflake announced a number of new features and products that will further extend the capabilities of the Data Cloud. These announcements make it clear that Snowflake is committed to providing data professionals with the tools and resources they need to succeed.

Part 1: Unlocking the Power of Snowflake Native Apps

Introduction:

In today's fast-paced data-driven world, businesses are constantly seeking innovative solutions to streamline their data management, analytics, and decision-making processes. As data volumes grow and complexity increases, having the right tools becomes essential. Snowflake, a leading cloud data platform, recognized this need and introduced a game-changing solution: Snowflake Native Apps. In Part 1 of our series, we'll explore what Snowflake Native Apps are, their key features and advantages, and their real-world use cases.

What are Snowflake Native Apps?

At its core, Snowflake Native Apps are specialized applications designed to work seamlessly within the Snowflake platform. They are not just integrations or add-ons but fully integrated tools that leverage Snowflake's architecture and capabilities. These apps cater to specific data roles and responsibilities, making it easier for users to access, analyze, and visualize data.

Key Features and Advantages:

Snowflake Native Apps come with a range of features that set them apart:

1. Integration with Snowflake's Data Sharing: Native apps are tightly integrated with Snowflake's data sharing capabilities, allowing users to share data securely with internal and external parties. This promotes collaboration and data-driven decision-making.

2. Simplified User Experience: These apps provide an intuitive, user-friendly experience, reducing the learning curve for users. They offer a consistent interface and a familiar environment for data tasks.

3. Streamlined Data Access: Native apps allow users to access data stored in Snowflake without the need for complex data transfers or copies. This minimizes data movement and ensures data accuracy.

4. Enhanced Security and Compliance: Security is a priority in Snowflake Native Apps, with built-in security features and robust compliance options. Users can access data while adhering to data privacy and governance regulations.

5. Performance Optimization: These apps are optimized for performance, ensuring efficient data processing and analytics. Users can work with data at scale without compromising speed.

Real-World Use Cases:

To better understand the value of Snowflake Native Apps, let's explore a few real-world use cases:

1. Data Science and AI: Snowflake's native apps are a boon for data scientists and AI practitioners. With features like data sharing and advanced analytics, they can seamlessly access, analyze, and train machine learning models using Snowflake data. This integration accelerates the model development process and promotes data science collaboration.

2. Data Collaboration and Sharing: Organizations can use native apps to facilitate data collaboration among teams and partners. By sharing data directly these Apps, they can ensure data accuracy, security, and compliance while fostering cross-functional teamwork.

In conclusion, Snowflake Native Apps are a game-changing addition to the Snowflake platform, redefining how businesses interact with their data. These apps enhance data access, analysis, and sharing while maintaining robust security and performance. In the next part of our series, we'll take a deep dive into Snowflake's Native Apps portfolio to explore the range of tools available and how they cater to different data roles and responsibilities. 

Introducing our own Native App: Snoptimizer

Unlock the power of Snowflake Native Apps with Snoptimizer, created by our visionary founder and 4x Snowflake Data Superhero, Frank Bell. With 1.6 years of dedicated experience in Native Apps, we're the pioneers you need to bring your ideas to life.

Sign up today for a personalized demo and let us transform your data world.

Don't miss out on the expertise that only Snoptimizer can offer. Visit our website to explore the endless possibilities. Your data deserves the best!

We are happy to help optimize your Snowflake account or help you build your own Snowflake Native App.



Part 2 will delve into specific Native Apps and their use cases, and Part 3 will explore the evolving landscape and trends of data analytics with these powerful tools.

Snowflake’s Plan to Acquire Ponder – Python in the Cloud

Introduction:

In the ever-evolving landscape of cloud computing, Snowflake has once again made headlines with a significant move. The data warehousing giant recently announced its acquisition of Ponder, a company specializing in Python capabilities in the cloud. This strategic move is poised to have a profound impact on how organizations leverage data analytics and Python for their business operations. This article will cover everything around Snowflake's plan to acquire Ponder - Python in the Cloud.

The Power of Python in the Cloud:

Python has firmly established itself as a leading programming language for data analysis, machine learning, and artificial intelligence. Its simplicity, versatility, and rich ecosystem of libraries have made it a go-to choice for data professionals. Snowflake, the data warehousing company known for its cloud-based approach to data management, has recognized the growing importance of Python in modern data analytics.

Snowflake's Plan to Acquire Ponder:

Ponder, a company with a strong focus on enhancing Python capabilities in the cloud, caught Snowflake's attention. The acquisition is part of Snowflake's broader strategy to empower its customers with more robust data analytics tools. By integrating Ponder's expertise and technologies into the Snowflake platform, users will have access to enhanced Python capabilities for their data-driven tasks.

If you care to read the official release from Snowflake on this acquisition, here's the link to see it.

Key Benefits of the Acquisition:

  1. Seamless Integration: Snowflake's acquisition of Ponder aims to seamlessly integrate Python capabilities into its data warehousing platform. This integration will make it easier for data professionals to work with Python within Snowflake's ecosystem. Hence, allowing for a smoother, more efficient workflow.
  2. Efficient Data Analysis: Python users will benefit from the cloud's scalability and elasticity. This enables them to analyze and process massive datasets without the limitations of on-premises solutions.
  3. Collaboration and Sharing: The integration of Ponder's technology will facilitate better collaboration among data teams. Python-based workflows can be shared, modified, and scaled with ease, leading to improved teamwork and productivity.
  4. Security and Governance: Snowflake's robust security and governance features will extend to Python workloads. This will ensure that sensitive data remains protected, even when processed using Python.

Implications for Data Professionals:

This acquisition presents exciting opportunities for data professionals. With enhanced Python capabilities available within Snowflake, data scientists, analysts, and engineers can expect to work more efficiently, analyze larger datasets, and extract deeper insights from their data. Furthermore, the cloud-based environment simplifies data management and maintenance, reducing the operational burden.

What we can help you with:

Here, at ITS Snowflake Solutions, we are experts at building data-driven businesses in the Snowflake cloud. Our latest product, Snoptimizer, aims to optimize your Snowflake account and help your business thrive.

Conclusion:

Snowflake's acquisition of Ponder is a clear sign of the growing importance of Python in the world of cloud-based data analytics. This move empowers data professionals to harness the full potential of Python within Snowflake's robust platform. The combination of these two technologies promises to accelerate data-driven decision-making, boost productivity, and further establish Snowflake as a leader in the data warehousing and analytics space.

As the integration progresses, organizations are encouraged to explore the enhanced Python capabilities offered by Snowflake to take their data analytics to new heights in the cloud.

Shortest Snowflake Summit 2023 Recap

Introduction:

Similar to last year, I wanted to create a “shortest” recap of the Snowflake Summit 2023, including the key feature announcements and innovations.  This is exactly 2 weeks after Snowflake Summit has ended and I have digested the major changes.  Throughout July and August we will follow up with our view of the massive Data to Value improvements and capabilities being made.

 

Snowflake Summit 2023 Recap from a Snowflake Data Superhero:

If you were unable to attend the Snowflake Summit, or missed any part of the Snowflake Summit Opening Keynote, here is a recap of the most important feature announcements.

 

Top Announcements:

 

 1.Native Applications goes to Public Preview: 

I am slightly biased here because my teams have been working with Snowflake Native Apps since Feb/March 2022. We have been on the journey with Snowflake from early early Private Preview to now over the last 16 months or so.  We are super excited about the possibilities and potential of where this will go.  

 

2. Nvidia/Snowflake Partnership, Support of LLMs, and Snowpark Container Services (Private Preview):   

Nvidia and Snowflake are teaming up (because as Frank S. says… some people are trying to kill Snowflake Corp) and they will integrate Nvidia’s LLM framework into Snowflake. I’m also really looking forward to seeing how these Snowpark Container Services work.

 

3. Dynamic Tables (Public Preview):  

Many Snowflake customers including myself are really excited about this.  These allow new Data Set related key features beyond a similar concept like Materialized Views. With Dynamic Tables you can… have declarative data pipelines, dynamic SQL Support , user defined low latency freshness, automated incremental refreshes, and snapshot isolation.

 

4. Managed Iceberg Tables (Private Preview): 

“Managed iceberg tables” allows Snowflake Compute Resources to manage Iceberg data.  This really helps with easier management of iceberg format data and helps Snowflake compete for Data Lake or really large Data File type workloads. So Snowflake customers can manage their data lake catalog with Iceberg BUT still get huge value with better compute performance with Snowflake’s query engine reading the metadata that Iceberg provides.  In some ways this is a huge large file data to value play.  It enables what blob storage (S3, Azure, etc.) do best BUT then being able to utilize Snowflake’s compute means less DATA TRANSFORMATION and faster value from the data including dealing with standard data modifications like updates, deletes and inserts.

 

5. Snowpipe Streaming API (Public Preview): 

As someone that worked with and presented on the Kafka Streaming Connector back at Summit 2019 it is really great to see this advancement. Back then the connector was “ok”. It could handle certain levels of streaming workloads. 4 years later this streaming workload processing has gotten much much better.

 

Top Cost Governance and Control Changes:

As anyone who has read my blog over the past few years, I’m a huge advocate of the Snowflake pay for what you use is AWESOME but ONLY when tools like our Snoptimizer® Optimization tool is used or you really really setup all the cost guard rails correctly.  98% of accounts we help with Snoptimizer do not have all the optimizations set correctly.  Without continuous monitoring of costs (and for that matter performance and security – which we also offer unlike a lot of the other copycats).

1. Budgets (Public Preview): 

This “budget” cost control feature was actually announced back in June 2022.  We have been waiting for it for some time now.  It is good to see Snowflake finally delivering this functionality. Since we started as one of the top Snowflake Systems Integrators back in 2018 there has been ONLY Resource Monitors to have ANY control whatsoever with guardrail limit type functionality.  This has been a huge pain point for many customers for many years.  Now, with this budget feature, users can actually specify a budget and get much more granular details about their spending limits.

2. Warehouse Utilization (Private Preview): 

This is another great step forward for Snowflake customers looking to optimize their Snowflake warehouse utilization.  We already leverage meta data statistics that are available to do this within Snoptimizer® but we are limited by the level of detail we can gather. This will allow us to optimize workloads much better across Warehouses to get even higher Snowflake Cost Optimization for our customers.

 

My takeaways from Snowflake Summit 2023:

  • If you would like more content and my summaries are not enough details then you are in luck. Here are more details from my team on our top findings around Snowflake Summit 2023.
  • Snowpark Container Services allow Snowflake customers to now run any job, function or service — from 3rd party LLMs, to Hex Notebook to a C++ application to even a full database, like Pinecone, in users’ own accounts. It now supports GPUs.
  • Streamlit is getting a new faster and easier user interface to develop apps. It is an open-source Python-based framework compatible with major libraries like sci-kit-learn, PyTorch, and Pandas. It has Git integration for branching, merging, and version control.
  • Snowflake is leveraging two of its recent acquisitions — Applica and Neeva to provide a new Generative AI experience. The former acquisition has led to Document AI, an LLM that extracts contextual entities from unstructured data and queries unstructured data using natural language. The unstructured to structured data is persisted in Snowflake and vectorized. Not only can this data be queried in natural language, but it can also be used to retrain the LLM on private enterprise data. While most vendors are pursuing prompt engineering. Snowflake is following the retraining path.
  • Snowflake now provides full MLOps capabilities, including Model Registry, where models can be stored, version controlled, and deployed. They are also adding a feature store with compatibility with open-source Feast. It is also building LangChain integration.
  • Last year, Snowflake added support for Iceberg Tables. This year it brings the tables under its security, governance, and query optimizer umbrella. Iceberg table’s performance now matches the tables’ query latency in native format.
  • Snowflake is addressing the criticism of its high cost through several initiatives designed to make costs predictable and transparent. Snowflake Performance Index (SPI) — using ML functions, it analyzes query durations for stable workloads and automatically optimizes them. This has led to 15% improvement on customers’ usage costs.
  • Snowflake has invested hugely in building native data quality capabilities within its platform. Users can define quality check metrics to profile data and gather statistics on column value distributions, null values, etc. These metrics are written to time-series tables which helps build thresholds and detects anomalies from regular patterns.
  • Snowflake announced two new APIs to support the ML lifecycle:
  • ML Modeling API: The ML Modeling API includes interfaces for preprocessing data and training models. It is built on top of popular libraries like Scikit Learn and XGBoost, but seamlessly parallelizes data operations to run in a distributed manner on Snowpark. This means that data scientists can scale their modeling efforts beyond what they could fit in memory on a conventional compute instance.
  • MLOps API: The MLOps API is built to help streamline model deployments. The first release of the MLOps API includes a Model Registry to help track and version models as they are developed and promoted to production.
  • Improved Apache Iceberg integrations
  • GIT Integration: Native git integration to view, run, edit, and collaborate within Snowflake code that exists in git repos. Delivers seamless version control, CI/CD workflows, and better testing controls for pipelines, ML models, and applications.
  • Top-K Pruning Queries: Enable you to only retrieve the most relevant answers from a large result set by rank. Additional pruning features, help reduce the need to scan across entire data sets, thereby enabling faster searches. (SELECT ..FROM ..TABLE ORDER BY ABC LIMIT 10).
  • Warehouse Utilization: A single metric that gives customers visibility into actual warehouse utilization and can show idle capacity. This will help you better estimate the capacity and size of warehouses.
  • Geospatial Features: Geometry Data Type, switch spatial system using ST_Transformation, Invalid shape detection, many new functions for Geometry and Geography
  • Dynamic Tables
  • Amazon S3-compatible Storage
  • Passing References for Tables, Views, Functions, and Queries to a Stored Procedure — Preview

 

Marketplace Capacity Drawdown Program

Anomaly Detection: Flags metric values that differ from typical expectations.

Contribution Explorer: Helps you find dimensions and values that affect the metric in surprising ways.

 

What did happen to Unistore? 

 

UNISTORE. OLTP type support based on Snowflake’s Hybrid Table features: This was one of the biggest announcements by far. Snowflake now is entering a much larger part of data and application workloads by extending its capabilities beyond olap [big data. online analytical processing] into OLTP space which still is dominated by Oracle, SQL Server, mysql, postgresql, etc. This is a significant step that positions Snowflake as a comprehensive, integrated data cloud solution for all data and workloads.

This was from last year too – it’s great to see this move forward!  (even though..Streamlit speed is still a work in progress)

 

Application Development Disruption with Streamlit and Native Apps:

 

Low code data application development via Streamlit: The combination of this and the Native Application Framework allows Snowflake to disrupt the entire Application Development environment. I would watch closely for how this evolves. It’s still very early but this is super interesting.

Native Application Framework: I’ve been working with this tool for about three months and I find it to be a real game-changer. It empowers data professionals like us to create Data Apps, share them on a marketplace, and even monetize them. This technology is a significant step forward for Snowflake and its new branding.

Snowflake at a very high level (still) wants to:

Disrupt Data Analytics

Disrupt Data Collaboration

Disrupt Data Application Development

Snowflake Summit Guide by ITS

Snowflake Summit 2023
Practical Guide

ESSENTIAL INFO

1. Get your badge as early as possible. avoid the crowds that will be there tuesday.

2. Get a copy of the conference map if you haven’t already (attached below)

3. If you are interested in the vendors, then make a short list of who you want to visit. With 179 vendors this year along with Snowflake’s own set of booths make it hard to cover all of them.

4. Reserve your sessions early.  Already by a week ago many of the LLM ones were already full. (wow, sounds like AWS re:invent – such a wonderful experience of too many people)

5. IF you have a RESERVED session you actually don’t want to miss then get there 10 minutes early because supposedly at 5 minutes before they start giving away reserved seats.

KEYNOTES

June 26, 20235:00 PM – 5:45 PM PT
Generative AI’s Impact on Data Innovation in the Enterprise 


June 27, 20239:00 AM – 11:15 AM PT
Snowflake Summit Opening Keynote 


June 28, 20239:00 AM – 10:30 AM PT
Builder Keynote

COME & MEET FRANK BELL

WHY ATTEND & WHAT TO DO?

In this world of information overflow, it’s important to take a step back and ask yourself what you really desire to learn from the Snowflake Summit. 

Here are some some guideline questions to help you prioritize your interests. 

1. Do you want to learn?  If so, how do you learn?  

2. Do you want to learn more about vendor offerings?

3. Do you want to get certified? 

4. Do you want to sell? 

5. Do you want to party? 

6. Do you want to network 

EXTRA PRO-TIPS

1. If you learn visually,  you can check in your own time the conferences being recorded. 

 

2. There is not enough time for you to see more than ~24 sessions out of 431 over 3-4 days.  So do yourself a favor, ask yourself what really matters to you personally and where do you want to focus.

 

3. If you are not a partier or networker than skip the parties.  Also, pro-tip – remember going all out Sunday, Monday or Tuesday night can make the next day(s) hard on yourself and your health! 

WANT TO PARTY?

Here is an “unofficial” list of parties as well as if you can still register as of Saturday 2023-06-24.  Snowflake Summit Parties – 2023 Unofficial List

We will try to update this as we gather more information. 


SNOWFLAKE: BEYOND AI

Specific Industry Sessions (snowflake has made a big move into multiple industry verticals.  check out the latest in these 7 specific verticals at summit) Industry Sessions at Snowflake Summit:

BE AT THE RIGHT TIME & PLACE

Last year the conference was completely within Caesar’s Forum but this year (2023) the sessions are distributed between BOTH Caesar’s Forum and Caesar’s Palace which is about .7 mile walk

For the #datageeks,  session distribution between Forum and Palace

DISCOVER OUR PARTNERS!

Dataops.live

 
Visit us at booth #2253 to see a demo and learn how to super charge your data engineering teams and realize 10x data engineering productivity: mitigate risk, accelerate data products, reduce costs.
And don’t miss our session June 27th at Noon PDT with our CTO & Co-Founder Guy Adams titled “Build Your Snowpark-Powered Data Products and Data Applications w DataOps.live”.

Sigma Computing


Awarded Snowflake Business Intelligence. Partner of the Year 2023. Whether you spreadsheet or SQL—teams explore, analyze, & decide with data in Sigma. Whether attending one of our customer sessions, visiting us in our booth, or joining on of our flagship evening events, we hope you had a chance to learn more about what it means to data confidently.

Hightouch

 

Hightouch is a data integration platform that helps businesses connect and sync their customer data across various tools and systems. Whether you attended one of our customer sessions, visited us in our booth, or joined one of our flagship evening events, we hope you had a chance to learn more about what it means to confidently handle data. 

Coalesce


Visit the Coalesce booth 1300 in Basecamp West for an incredible amount of fun, mind-blowing demos, and an abundance of goodies.
Experience new feature demos on four TVs, witness mesmerizing magic tricks (some not even related to Coalesce), and grab your exclusive Data Transformer giveaways, including the hottest t-shirt at Summit.

 

Fivetran


Looking for a chance to see Fivetran in action and learn more about our data integration platform? Look no further than Booth #1700 in Basecamp West at the upcoming event!
Stop by our booth to get a live demo and see firsthand how easy Fivetran makes data movement into Snowflake — regardless of your data source.

 

Snowflake


We are one of the first original Partners of Snowflake.
Our founder, Frank Bell, is 1 of 72 Snowflake Data Superheroes worldwide. 

 
 

Data to Value – Part 2

Introduction:

Welcome to our part 2 Data to Value series. If you've read Part 1 of the Data to Value Series, you've learned about some of the trends happening within the data space industry as a whole.

In Part 2 of the Data to Value series, we'll explore additional trends to consider, as well as some of Snowflake's announcements in relation to Data to Value.

As a refresher on this series, we are making a fundamental point that data professionals and data users of all types need to be focused not just on creating, collecting, and transforming data. We need to make a cognizant effort to focus on and measure what is the true value that each set of data creates. Also, we need to measure, how fast we can get to that value if it provides any real business advantages. There is an argument to also alter the value of the data that is time-dependent since it loses value sometimes the older it is.

 

Data to Value Trends - Part 2:

 

8) - Growth of Fivetran and now Hightouch.

The growth and success of Fivetran and Stitch (now Talend) has been remarkable. There is now a significant surge in the popularity of automated data copy pipelines that work in the reverse direction, with a focus on Reverse ETL (Reverse Extraction Transformation and Load), much like our trusted partner, Hightouch. Our IT Strategists consulting firm became partners with Stitch, Fivetran, and Matillion in 2018.

At the Snowflake Partner Summit of the same year, I had the pleasure of sitting next to Jake Stein, one of the founders of Stitch, on the bus from San Francisco to Sonoma. We quickly became friends, and I was impressed by his entrepreneurial spirit. Jake has since moved on to a new startup, Common Paper, a structured contracts platform, after selling Stitch to Talend. At the same event, I also had the opportunity to meet George Frazier from Fivetran, who impressed me with his post comparing all the cloud databases back in 2018. At that time, such content was scarce.

 

9) - Resistance to “ease of use” and “cost reductions” is futile.

Part of me as a consultant at the time wanted to resist these “Automated EL Tools” EL (Extract and Load) vs ETL – (Extract, Transform, and Load) or ELT (Extract, Load, and then Transform within the database).  As I tested out Stitch and Fivetran though, I knew that resistance was futile. The ease of use of these tools and the reduction of development and maintenance costs cannot be overlooked. There was no way to stop the data market from embracing these easier-to-use data pipeline automation tools.

What was even more compelling is you can set up automated extract and load jobs within minutes or hours most of the time. This is unlike any of the previous ETL tools we have been using for decades which were mostly software installations. These installations took capacity planning, procurement, and all sorts of organizational business friction to even get started at all. With Fivetran and Hightouch, there is no engineering or developer expertise needed for almost all of the work. In some cases, it can be beneficial to have the expertise of data engineers and architects involved.

Overall, the concept is simple: connecting destinations and connectors to facilitate Fivetran. Destinations refer to databases or data stores. Connectors are sources of data, such as Zendesk, Salesforce, or one of the many other connectors in Fivetran. Fivetran and Hightouch are great examples of trends in data services and tools that really speed up the process of getting value from your data.

 

10) - Growth of Automated and Integrated Machine Learning Pipelines with Data.

Many companies, including Data Robot, Dataiku, H2O, and Sagemaker, are working to achieve this goal. However, this field appears to be in its early stages, with no single vendor having gained widespread adoption or mindshare. Currently, the market is fragmented, and it is difficult to predict which of these tools and vendors will succeed in the long run.

 

Snowflake's Announcements related to Data to Value

Snowflake is making significant investments and progress in the field of data analysis, with a focus on delivering value to its clients. Their recent announcements at the Snowflake Summit this year, as detailed in this source, highlight new features that are designed to enhance the Data to Value experience.

 

Snowflake recently announced its support of Hybrid Tables and the concept of Unistore.

This move is aimed at providing Online Transaction Processing (OLTP) to its customers. There has been great interest from customers in this concept, which allows for a single source of truth through web-based OLTP-type applications operating on Snowflake with Hybrid tables.

 

Announcements about Snowflake's Native Apps:

 

  • Integrating Streamlit into Snowflake.

If done correctly, this could be yet another game-changer in turning data into value.
Please note that these two items mentioned not only enable data to be processed more quickly, but also significantly reduce the cost and complexity of developing data apps and combining OLTP/OLAP applications. This removes many of the barriers that come with requiring expensive, full-stack development. Streamlit aims to simplify the development of data applications by removing the complexity of the front-end and middle-tier components. (After all, aren't most applications data-driven?) It is yet another low-code data development environment.)

  • Announcement of Snowpipe streamlining.

I found this particularly fascinating, as I had collaborated with Isaaic from Snowflake before the 2019 Summit using the original Kafka to Snowflake Connector. At Snowflake Summit 2019, I also gave a presentation on the topic. It was truly amazing to witness Snowflake refactor the old Kafka connector. As a result, there were significant improvements in speed and lower latency. This is yet another major victory for streamlining data to improve value, with an anticipated 10 times lower latency. The public preview is slated for later in 2022.

  • Announcement: Snowpark for Python and Snowpark in General

Snowflake has recently introduced a new technology called Snowpark. While the verdict is still out on this new technology, it represents a major attempt by Snowflake to provide ML pipeline data with increased speed. Snowflake is looking to integrate full data event processing and machine learning processes within Snowflake itself.

 

If Snowflake can execute this correctly, it will revolutionize how we approach data value. Additionally, it reduces the costs associated with deploying data applications.

 

Conclusion:

In part 2 of the "Data to Value" series, we explored additional trends in the data industry, including the growth of automated data copy pipelines and integrated machine learning pipelines. We also discuss Snowflake's recent announcements related to data analysis and delivering value to clients, including support for hybrid tables and native apps. The key takeaway is the importance of understanding the value of data and measuring the speed of going from data to business value.

Executives and others who prioritize strategic data initiatives should make use of Data to Value metrics. This helps us comprehend the actual value that stems from our data creation, collection, extraction, transformation, loading, and analytics. By doing so, we can make better investments in data initiatives for our organizations and ourselves. Ultimately, data can only generate genuine value if it is reliable and of confirmed quality.

Snowflake Plugin Available NOW on VSCode

Snowflake Plugin Available NOW on VSCode:

 

We have great news! Snowflake has released its own VSCode Plugin! It’s currently in Public Preview (PuPr) and you can download it from the Microsoft Visual Studio Extension marketplace. With the Snowflake Plugin, you will have access to some features such as:

 

  • Accounts and Sessions: the plugin allows you to connect to and easily switch between multiple Snowflake accounts. And (this is cool) you can share a single session between multiple open VSCode editor windows! Support for Single Sign On (SSO) is available.
  • Snowflake SQL Intellisense: autocomplete for object names, keywords, and built-in functions, with signature help for function calls. Links to documentation for keywords and built-in functions on hover.
  • Database Explorer: a treeview-based panel that lets you drill down into object definitions and details.
  • Query Execution: not just single statements, but multiple statement executions!
  • Query Results and History panel: View and sort query results and export results to CSV format. Review prior statement history and results, and copy/paste support on previous queries.

 

How to install the Snowflake plugin on VSCode:

 

  1. Launch VSCode and head over to the Extensions Marketplace tab

     2. Type in “Snowflake” and select the verified Snowflake extension (It should have the verification checkmark)

     3. Click on the Snowflake icon to log in. The extension will ask for your account’s URL however this part can be tricky. Instead of inputting the whole URL just add the part before .snowflakecomputing.com

 

For example, if your account URL is https://ja13154.east-us-2.azure.snowflakecomputing.com/, enter ja13154.east-us-2.azure in the Account Name/URL box.

 

                                 

 

    4. As a final step add your username and password and you are all set to go!

 

With these simple steps, you can now use the Snowflake Plugin on VSCode. If you want to learn about other new features on Snowflake, be sure to check out our blog for new updates.

Snowflake Snowday – Data to Value Superhero Summary

Snowflake Snowday  —  Summary

Snowflake's semiannual product announcement, Snowflake Snowday, took place on November 7, 2022, the same day as the end of Snowflake's Data Cloud World Tour (DCWT).

I attended 5 DCWT events across the globe in 2022. It was fascinating to see how much Snowflake has grown since the 2019 tour. Many improvements and new features are being added to the Snowflake Data Cloud. It's hard to keep up! These announcements should further improve Snowflake's ability to turn data into value.

Let's summarize the exciting Snowflake announcements from Snowday. The features we're most enthusiastic about that improve Data to Value are:

  • Snowflake's Python SDK (Snowpark) is now generally available.
  • Private data sharing significantly accelerates collaborative data work.
  • The Snowflake Kafka connector, dynamic tables, and Snowpipe streaming enable real-time data integration.
  • Streamlit integration simplifies dashboard and app development.

All of these features substantially improve Data to Value for organizations.

Snowflake Snowday Summary - Top Announcements

TOP announcement! – whoop whoop – SNOWPARK FOR PYTHON! (General Availability – GA)

I believe this was the announcement all Python data scientists were anticipating (including myself). Snowpark for Python now enables every Snowflake customer to develop and deploy Python-based apps, pipelines, and machine-learning models directly in Snowflake. In addition to Snowpark for Python being Generally Available to all Snowflake editions, these other Python-related announcements were made:

  • Snowpark Python UDFs for unstructured data (Private Preview)
  • Python Worksheets – The improved Snowsight worksheet now supports Python so you don't need an additional development environment. This simplifies getting started with Snowpark for Python development. (Private preview)

One Product. One Platform.

  • Snowflake’s major push is to make its platform increasingly easy to use for most or all of its customers’ data cloud needs.
  • Snowflake now offers Hybrid Tables for OLTP workloads and Snowpark. Snowflake is expanding its core platform to handle AI/ML and online transaction processing (OLTP) workloads. This significantly increases Snowflake’s total addressable market.
  • Snowflake acquired Streamlit earlier this year for a main reason. They aim to integrate Streamlit's data application frontend and backend. They also want to handle data application use cases.
  • Snowflake is investing heavily to evolve from primarily a data store to a data platform for building frontend and backend data applications. This includes web/data apps needing millisecond OLTP inserts or AI/ML workloads.

Additionally, Snowflake continually improves the core Snowflake Platform in the following ways:

The Cross-Cloud Snowgrid:

https://snowflakesolutions.net/wp-content/uploads/Snowday-Cross-Cloud-Snowgrid-1024x762.png

Replication Improvements and Snowgrid Updates:

These improvements and enhancements to Snowflake, the cross-cloud data platform, significantly boost performance and replication. If you're unfamiliar with Snowflake, we explain what Snowgrid is here.

  • Cross-Cloud Business Continuity – Stream & Task Replication (PUBLIC PREVIEW) – This enables seamless pipeline failover, which is fantastic. It takes replication beyond just accounts, databases, policies, and metadata.
  • Cross-Cloud Business Continuity – Replication GUI (PRIVATE PREVIEW). You can now more easily manage replication and failover from a single interface for global replication. It enables easy setup, management, and failover of an account.
  • Cross-Cloud Collaboration – Discovery Controls (PUBLIC PREVIEW)
  • Cross-Cloud Collaboration – Cross-Cloud Auto-Fulfillment (PUBLIC PREVIEW)
  • Cross-Cloud Collaboration – Provider Analytics (PUBLIC PREVIEW)
  • Cross-Cloud Governance – Tag-Based Masking (GA)
  • Cross-Cloud Governance – Masking and Row-Access Policies in Search Optimization (PRIVATE PREVIEW)
  • Replication Groups – Looking forward to updates on this as well. These can enable sharing and simple database replication in all editions.
  • The above are available in all editions EXCEPT:
  • Enterprise or higher needed for Failover/Failback (including Failover Groups)
  • Business Critical or higher needed for Client Redirect functionality

Performance Improvements on Snowflake Updates:

New performance improvements and performance transparency were announced were related to:

  • Query Acceleration (public preview): Speeds up search queries.
  • Search Optimization Enhancements (public preview): Improves search relevance and precision.
  • Join eliminations (GA): Removes unnecessary table joins.
  • Top results queries (GA): Returns the most relevant search results.
  • Cost Optimizations: Account usage details (private preview): Reduces search costs.
  • History views (in development): Provides search query history.
  • Programmatic query metrics (public preview): Offers API for search analytics. Available on all editions EXCEPT: ENTERPRISE OR HIGHER REQUIRED for Search Optimization and Query Acceleration

Data Listings and Cross-Cloud Updates

I’m thrilled about Snowflake’s announcement regarding Private Listings. Many of you know that Data Sharing, which I’ve been writing about for over 4 years, is one of my favorite Snowflake features. My latest article is “The Future of Data Collaboration.” Data Sharing is a game-changer for data professionals.

Snowflake’s announcement makes private data-sharing scenarios much easier to implement. Fulfilling different regional requirements is now simpler too (even 1-2 years ago, we had to write replication commands). I’ll provide more details on how this simplifies data sharing and collaboration. I was happy to see presenters use the Data to Value concepts in their announcement.

I appreciated Snowflake incorporating some of my Data to Value concepts, like “Time to value is significantly reduced for the consuming party.” Even better, this functionality is now available for ALL SNOWFLAKE EDITIONS.

Private Listings (Get a crisper-looking visual)

https://snowflakesolutions.net/wp-content/uploads/Snowday-Listings-Cross-Cloud-Improvements-300x190.png

Snowflake Data Governance Improvements

All Snowflake features enable native data governance and protection.

  • Tag-based Masking automatically applies designated policies to sensitive columns using tags.
  • Search Optimization now supports tables with masking and row access policies.
  • FedRAMP High for AWS Government (authorization pending). *Available ONLY on ENTERPRISE+ OR HIGHER

Building on Snowflake

New announcements related to:

  • Streamlit integration (PRIVATE PREVIEW in January 2023) – This integration will be exciting. The private preview can’t come soon enough.
  • Snowpark Optimization Warehouses (PUBLIC PREVIEW) – This was a smart move by Snowflake to support AI/ML Snowpark customers’ needs. Great to see it rolled out, allowing customers access to higher memory warehouses better suited for ML/AI training scale. Snowpark code can run on both warehouse types.
  • *Available for all Snowflake Editions

Streaming and Dynamic Table Announcements:

Conclusion:

Overall, I'm thrilled with where this is headed. These enhancements greatly improve Snowflake's streaming data integration, especially with Kafka. Now, Snowflake customers can get real-time data streams and transform data with low latency. When fully implemented, this will enable more cost-effective and high-performance data lake solutions.

If you missed Snowday and want to watch the recording, here's the link: https://www.snowflake.com/snowday/agenda/

We'll cover more updates from Snowday and Snowflake BUILD in depth this week in the Snowflake Solutions Community.

Data to Value – Part 1 – Snowflake Solutions

Introduction:

 

Welcome to our Frank’s Future of Data four-part series. In these articles, we will cover a few tips on how to get value out of your Snowflake data.

I spend a ton of time reviewing and evaluating all the ideas, concepts, and tools around data, data, and data. The “data concept” space has been exploding with an increase in many different concepts and ideas. There are so many new data "this" and data "that" tools as well so I wanted to bring data professionals and business leaders back to the core concept that matters around the creation, collection, and usage of data. Data to Value.

In layman’s terms, the main concept is that we need to remember that the entire point of collecting and using data is to create business, organizational, and/or individual value. This is the core principle that we should keep in mind when contemplating the value that data provides.

The truth is that while the technical details and jargon involved in creating and collecting data, as well as realizing its value, are important, many users find them overly complex.

For a moment, let's set aside the technical jargon that can be overused and misused, such as Data Warehouse, Data Lake, Data Mesh, and Data Observability. I've noticed that data experts and practitioners often have differing views on the latest concepts. These views can be influenced by their data education background and the types of technologies they were exposed to.

Therefore, I created these articles to prepare myself for taking advantage of new paradigms that Snowflake and other “Modern Data” Stack tools/clouds provide.

On Part 1 of the Data to Value series we will cover the Data to Value trends you need to be aware of.

 

Data to Value Trends:

 

In 2018, I had the opportunity to consult with some highly advanced and mature data engineering solutions. Some of these solutions were actively adopting Kafka/Confluent to achieve true "event-driven data processing". This represented a significant departure from the traditional batch processing that had been used in 98% of the implementations I had previously encountered. I found the idea of using continuous streams of data from different parts of the organization, delivered via Kafka topics, to be quite impressive. At the same time, these concepts and paradigm shifts were quite advanced and likely only accessible to very experienced data engineering teams.

1) – Non-stop push for faster speed of Data to Value.

Within our non-stop dominantly capitalist world, faster is better and often provides advantages to organizations, especially around improved value chains and concepts such as supply chains.  Businesses and organizations continuously look for any advantage they can get. I kinda hate linking to McKinsey for backup but here it goes. Their number 2 characteristic for the data-driven enterprise of 2025 is “Data is processed and delivered in real-time”.

 

2) – Data Sharing.

More and more Snowflake customers are realizing the massive advantage of data sharing allowing them to share “no-copy,” in-place data in near real-time.  Data Sharing is a massive competitive advantage if set up and used appropriately. You can securely provide or receive access to data sets and streams from your entire business or organization value chain which is also on Snowflake. This allows for access to data sets at reduced cost and risk due to the micro-partitioned zero-copy securely governed data access.

 

3) – Creating Data with the End in Mind.

When you think about using data for value and logically think through the creation and consumption life cycle then data professionals and organizations are realizing there are advantages to capturing data in formats that are ready for immediate processing.  If you design your data creation and capture as logs of data or other outputs that can be easily and immediately consumed you can gain faster data-to-value cycles creating competitive advantages with certain data streams and sets.

 

4) – Automated Data Applications.

I see some really big opportunities with Snowflake’s Native Applications and Streamlit integrated. Bottom-line, there is a need for consolidated “best-of-breed” data applications that can have a low-cost price point due to massive volumes of customers.

 

5) – Full Automated Data Copying Tools.

The growth of Fivetran and Stitch (Now Talend) has been amazing.  We now are also seeing huge growth in automated data copy pipelines going the other way like Hightouch.  At IT Strategists, we became a partner with Stitch, Fivetran, and Matillion back in 2018.

 

6) – Full Automation of Data Pipelines and more integrated ML and Data Pipelines.

With the introduction of a fully automated data object and pipeline service at Coalesce, we saw for the first time that data professionals improve Data to Value through fully automated data objects and pipelines. Some of our customers are referring to parts of Coalesce as a Terraform-like product for data engineering. What I see is a massive removal of data engineering friction similar to what Fivetran and Hightouch did but at a separate area of the data processing stack. We have become an early partner with Coalesce because we think it is similar to how we viewed Snowflake at the beginning of 2018. We view Coalesce as just making Snowflake even more amazing to use.

 

7) – The Data Mesh Concept(s) and Data Observability.

Love these concepts or hate them, they are taking hold within the overall data professionals’ brain trust. Zhamak Dehghani (previously at Thoughtworks) and ThoughtWorks from 2019 until now have succeeded in communicating to the market the concept of a Data Mesh.  Whereas, Barr Moses from Monte Carlo, has been beating the drum very hard on the concept of Data Observability. I’m highlighting these data concepts as trends that are aligned with improving Data to Value speed, quality, and accessibility.  There are many more data concepts besides these two.  Time will reveal which of these will gain mind and market share and which will go by the wayside.

 

Conclusion:

That is it for Frank’s Future of Data part 1 series article. In our second section, Part 2, we will continue exploring more trends that we should keep in mind, as well as exploring Snowflake’s announcements related to Data to Value.

Snowflake Data Clean Rooms

Introduction: What is a Data Clean Room?

 

In this article, I will explain what a Snowflake Data Clean Room is on the Snowflake Data Cloud.
Data clean rooms on Snowflake are a set of data-related technologies that facilitate double-blind joins of data. These technologies include Data Shares, Row Access Policies, and Secure User Defined Functions. The underlying Data Sharing technology is based on Micro-Partitions, which provide features like Data Sharing and Data Cloning.

Although the original concept of data clean rooms was developed for data exchanges in advertising, I believe the concept can be applied to many other areas where "controlled" and "governed" double-blind joins of data sets can create significant value. This approach enables companies and their partners to share data at an aggregated double-blind join level, without sharing personally identifiable information (PII).
On Snowflake, sharing data through secure views and tables using their Data Share technology is already straightforward. You can share double-blind join previously agreed upon identifiers.

 

Part 1: Data Clean Room Example Use Cases

We helped Snowflake pioneer this new offering a couple of years ago with our client VideoAmp which we brought over to the Snowflake Data Cloud. Our original article back in July 2020 shows how to analyze PII and PHI Data using the earlier Data Clean Room concepts. Fast forward 2 years and now Snowflake has dramatically improved the initial version and scope that we put together. These are just a few examples; there are many other potential use cases for Snowflake Data Clean Rooms.

 

Media/Advertising:

  • Addressing the challenge of the "end of cookies" in a meaningful way, Snowflake's Data Clean Rooms enable Advertisers to merge their first-party data and their publisher(s)’ viewership/exposure data, delivering more value for their marketing spend.
  • Collaborative Promotions. Conducting customer segment overlap analysis with a co-branding/co-marketing partner can reveal areas where customer segments and audiences are aligned.
  • Joint loyalty offerings and/or upsells can also be developed in partnership with aligned customer "interests".

 

Healthcare and Life Sciences:

  • There are some extremely valuable use cases where we can securely share patient data and patient outcomes across government, healthcare, and life sciences to hopefully make some huge leaps forward in healthcare and life.

 

Financial Services:

  • Combining data from multiple financial institutions to identify fraud or money laundering activities without sharing sensitive customer information.

 

Retail:

  • Combining customer data from different sources to create targeted marketing campaigns and promotions.

 

Government:

  • Sharing data across different government agencies to improve public services while protecting individual privacy.

 

Part 2: Looking for more information about Data Clean Rooms?

Here are some additional resources to help you learn more about Data Clean rooms and Data Collaboration.

 

Lastly, here’s an interview I provided on my view of the opportunities around Data Clean Rooms on Snowflake. I shared some insights gained from decades of experience working in data, including thoughts about the transformational impact that cloud-based data sharing, data collaboration, data marketplaces, and data clean rooms are having on companies and industries.

What’s Next in Data Collaboration & Why Data Clean Rooms Are Exciting: Insights From Frank Bell

 

Are you interested in how you can use a Snowflake Data Clean Room for your business? Contact Us Today.

Cost Governance on Snowflake in 2022

Introduction: What is Snowflake’s Cost Governance?

 

Snowflake cost governance refers to the process of managing and optimizing the costs associated with using the Snowflake cloud data platform. This involves monitoring and analyzing usage metrics to identify areas where costs can be reduced, as well as implementing strategies to control spending and prevent unexpected expenses. Snowflake offers various tools and features for cost governance, including resource groups, budgets, and usage views. However, some users may still choose to use third-party optimization tools to fully optimize their Snowflake accounts and save money.

 

Part 1: My Take on Snowflake’s Cost Governance

In this article, I'll explain what you can do to manage costs on Snowflake as of July 2022. Although Snowflake has made significant progress in this area, it's still recommended to use a comprehensive Snowflake cost optimization service like Snoptimizer™ or Nadalytics. This is due to the fact that Snowflake still generates most of its revenue from consumption-based services, and despite having impressive NPS scores, there are still many cost-related issues to be aware of. Before the Summit 2022 announcements, here's a brief overview of what was available.

 

Before Snowflake Summit 2022, Cost Governance in Snowflake was honestly pretty weak. It only had the following GUI and optimization tools:

 

  1. Daily Summary is available in Snowflake's Standard Classic Console. This provided very limited information and was available ONLY to very limited ROLES!
  2. Usage Views can be utilized in Snowsight - It shows more granularity of costs but there are problems with some default views and bugs. Again, by default, it is locked down to certain roles.
  3. Third-Party Optimization Tools can help you view your information and make sense of it. Some are:
    1. Nadlytics
    2. Snoptimizer™
  4. Third-Party "Reactive" Reporting Tools (from all the Snowflake Health Check Consulting Engagements I've done, this was the most common set of tools for Cost Governance on Snowflake).
    1. Sigma Computing Cost and Usage
    2. Looker Snowflake Cost and Usage
    3. Tableau Snowflake Cost and Usage
    4. Many other smaller fragmented brands with "reactive" reporting around costs. However, the problem with reactive reporting is that if something goes awry like a long-running query where there is NO Resource monitor OR the resource monitor is ONLY set to kick in when the query ends which by default could be 48 hours... If this happens $1000s or $10,000+ of dollars can be spent within a day easily with no true Data to Value provided!

After Snowflake Summit 2022, these major Cost Governance announcements were provided:

 

#1. A New Resource Groups concept announced where you can combine all sorts of Snowflake data objects to monitor their resource usage. [This is huge since Resource Monitors were pretty primitive]
#2. Concept of Budgets that you can track against. [both Resource Groups and Budgets are available in Private Preview in the next]
#3. More Usage Metrics are being made available as well for SnowPros like us to use or Monitoring tools. This is important since many enterprise businesses were looking for this.

Conclusion:

If you're interested in staying up-to-date with our latest updates, be sure to check our website regularly for more information. Looking to reduce costs and optimize your Snowflake Account to save money? Try our Snoptimizer™ Assessment for Free and see the results for yourself. We are confident that our assessment will provide you with valuable insights and recommendations to improve your Snowflake usage and help you save money in the process.

What is a Snowflake Data Superhero?

What is a Snowflake Data Superhero? 

 

Currently, a Snowflake Data Superhero (abbreviated as DSH) is a Snowflake product expert who is actively involved in the Snowflake community and is helping others learn more about Snowflake through blogs, videos, podcasts, articles, books, etc.

Snowflake states it chooses DSHs based on their positive influence on the overall Snowflake Community. Snowflake Data Superheroes get some decent DSH benefits as well, keep reading to learn more.

I'm Frank Bell, the founder of IT Strategists and Snowflake Solutions, and I'm also a Snowflake Data Superhero. In this article, I'd like to give you an overview of what a Snowflake Data Superhero is, what the program entails, and what are some of the benefits of being chosen as a DSH.

 

The Snowflake Data Superhero Program (Before Fall 2021)

 

For those of you new to Snowflake within the last few years, believe it or not, there was this really informal Data Superhero program for many years.  I don't even think there were an exact criteria list to be in it. Since I was a long-time Snowflake Advocate and one of the top Snowflake consulting and migration partners from 2018-2019 with IT Strategists (before we sold the consulting business), I was invited to be part of the informal program back in 2019.

Then those of us who had been involved with this informal program got this mysterious email and calendar invite in July 2021.  Invitation: Data Superhero Program Restructuring & Feedback @ Mon Jul 26, 2021 8am - 9am - Honestly, when I saw this and attended the session this sounded like it was going to be a pain in the ass having to validate our Snowflake expertise again within this new program. Especially for many of us in the Snowflake Advocate Old Guard. (There are probably around 40 of us I'd say who never decided to switch to be Snowflake employees of Snowflake Corporate to make a serious windfall as the largest software IPO in history (especially the Sloot and Speiser who became billionaires. Benoit did too but as I've stated before, Benoit, Thierry, and Marcin deserve some serious credit for the core Snowflake architecture. As an engineer you have to give them some respect.)

 

The Snowflake Data Superhero Program (2022)

 

This is a combination of my thoughts and the definitions from Snowflake.

Snowflake classifies Snowflake Data Superheroes (DSH) as an elite group of Snowflake experts! They also think the DSHs should be highly active in the overall Snowflake community. They share feedback with Snowflake product and engineering teams, receive VIP access to events, and their experiences are regularly highlighted on Snowflake Community channels. Most importantly, Data Superheroes are out in the community helping to educate others by sharing knowledge, tips, and best practices, both online and in person.

How does the Snowflake Corporation choose Snowflake Data Superheroes?

 

They mention that they look for the following key attributes:

 

  • You must overall be a Snowflake expert.
  • They look for Snowflake experts who create any type of content around the Snowflake Data Cloud (this could be any type of content from videos and podcasts to blogs and other written Snowflake publications.
  • They look for you to be an active member of the Data Hero community which is just the overall online community at snowflake.com.
  • They also want people who support other community members and provide feedback on the Snowflake product.
  • They want overall energetic and positive people

 

Overall, I would agree many of the 48 data superheroes for 2022 definitely meet all of the criteria above. This past year, since the program was new I also think it came down to that only certain people applied. (I think next year it will be less exclusive since the number of Snowflake experts is really growing from my view.  Back in 2018, there honestly was a handful of us. I would say less than 100 worldwide. Now there are most likely 200+ true Snowflake Data Cloud Experts outside of Snowflake Employees. Even though now, the product overall has grown so much that it becomes difficult for any normal or even superhero human to be able to cover all parts of Snowflake as an expert. The only way that I'm doing it (or trying to) is to employ many automated ML flows and Aflows I call them to organize all Snowflake publicly available content into this one knowledge repository of ITS Snowflake Solutions. I would also say that it comes down to your overall known prescience within the Snowflake Community and finally your geography. For whatever reason, I think Snowflake DSHs chosen by Snowflake for 2022 missed some really really strong Snowflake experts within the United States.

Also, I just want to add that even within the 48 Snowflake Data Superheroes, there are a few that just stand out as producing an insane amount of free community content.  I'm going to name them later after I run some analysis but there are about 10-15 people that just pump out the content non-stop!

 

What benefits do you get when you become a Snowflake Data Superhero?

 

Snowflake Data Superhero Benefits:

 

In 2022, they also provided all of these benefits:

 

  • A ticket to the Snowflake Summit - I have to say this was an awesome perk of being part of the program and while I disagree sometimes with Snowflake Corp decisions that are not customer or partner-focused, this was Snowflake Corporation actually doing something awesome, and really the right thing considering that of these 48 superheroes, most of us have HEAVILY contributed to Snowflake's success (no stock, no salary).  While employees and investors reap large financial gains from the Snowflake IPO, many of us basically helped the company grow significantly.
  • Snowflake Swag that is different (well, it was for a while, now others are buying the "kicks" or sneakers)
  • Early education on new Snowflake Features
  • Early access to new Snowflake Features (Private Preview)
  • Some limited opportunities to speak at events. (Let's face it, the bulk of speaking opportunities these days goes in this order:  Snowflake Employees, Snowflake Customers (the bigger the brand [or maybe the spend] the bigger the speaking opportunity), Snowflake Partners who pay significant amounts of money to be involved in any live speaking event, and finally external Snowflake experts, advocates, etc.
  • VIP access to events (we had our own Data Superhero area within Snowflake Summit)
  • Actual Product Feedback sessions with the Snowflake Product Managers

 

The only action that I can think of that really has been promised and not done so far in 2022 is providing every DSH with a test Snowflake Account with a certain number of credits.  Also, I do not think many of the DSHs have received their Data Superhero card. This was one of those benefits provided to maybe 10 or more of the DSHs back in 2019 or so.  Basically, anyone who was chosen to speak at Snowflake Build I believe is where some of it started.  I'm not 100% sure.

 

The Snowflake Data Superhero Program (2023)

 

How do I apply to be a Snowflake Data Superhero?
Here you go:  [even though for me the links are not working]
https://community.snowflake.com/s/dataheroes

 

Snowflake's Data Superhero Program Evolution

 

I will add some more content around this as I review how the 2023 program is going to work.  I will say I have been surprisingly pleased with the DSH Program overall this year in 2022.  It has provided those Snowflake Data Superheroes that are more involved with the program as a way to stand out within the Snowflake Community.

 

Snowflake's Data Superhero Program Internal Team

 

I also want to give a shout-out to the main team at Snowflake who works tirelessly to make an amazing Snowflake Data Superhero program. These individuals and more have been wonderful to work with this year:

  • Howard Lio
  • Leith Darawsheh
  • Elsa Mayer

There are many others too, from the product managers we meet with to other Snowflake engineers.

 

Other Snowflake Data Superhero Questions:

 

Here was the full list from Feb 2021.

Who are the Snowflake Data Superheroes?

https://medium.com/snowflake/introducing-the-2022-data-superheroes-ec78319fd000

 

Summary

 

I kept getting all of these questions about, hey - what is a Snowflake Data Hero?  What is a Snowflake Data Superhero?  How do I become a Snowflake Data Superhero?  What are the criteria for becoming one?

This article is my attempt to answer all of your Snowflake Data Superhero-related questions in one place. Coming from an actual Snowflake Data Superhero, I've been one for 3+ years in a row now. Hit me up in the comments or directly if you have any other questions.

Shortest Snowflake Summit 2022 Recap

Introduction:

 

Today’s article provides a recap of the Snowflake Summit 2022, including the key feature announcements and innovations. We highlight the major takeaways from the event and the outline of Snowflake's position as a full-stack business solution environment capable of creating business applications.

We also include a more in-depth discussion of Snowflake's seven pillars of innovation, which include all data, all workloads, global, self-managed, programmable, marketplace, and governed.

 

Snowflake Summit 2022 Recap from a Snowflake Data Superhero:

 

If you were unable to attend the Snowflake Summit, or missed any part of the Snowflake Summit Opening Keynote, here is a recap of the most important feature announcements.

Here are my top 20 announcements, mostly in chronological order of when they were announced. It was overwhelming to keep up with the number of announcements this week!

 

Cost Governance:

 

1. The concept of New Resource Groups has been announced. It allows you to combine all kinds of Snowflake data objects to monitor their resource usage. This is a huge improvement since Resource Monitors were previously quite primitive.

2. The concept of Budgets that you can track against. Resource Groups and Budgets coming into Private Preview in the next few weeks.

3. More Usage Metrics are being made available as well for SnowPros like us to use or Monitoring tools. This is important since many enterprise businesses were looking for this.

 

Replication Improvements on SnowGrid:

 

4. Account Level Object Replication: Snowflake previously allowed only data replication and not other account-type objects. However, now all objects that are not just data can supposedly be replicated as well.

5. Pipeline Replication and Pipeline Failover: Now, stages and pipes can be replicated. According to Kleinerman, this feature will be available soon in Preview.

 

Data Management and Governance Improvements:

 

6. The combination of tags and policies. You can now do  —  Private Preview now and will go into public preview very soon.

 

Expanding External Table Support and Native Iceberg Tables:

 

7. We will soon have support for external tables in Apache Iceberg. Keep in mind, however, that external tables are read-only and have certain limitations. Take a look at what Snowflake did in #9 below.

8. Snowflake is broadening its abilities to manage on-premises data by partnering with storage vendors Dell Technologies and Pure Storage. The integration is anticipated to be available in a private preview in the coming weeks.

9. We are excited to announce that Snowflake now fully supports Iceberg tables, which means these tables can now support replication, time travel, and other standard table features. This enhancement will greatly improve the ease of use within a Data Lake conceptual deployment. For any further inquiries or assistance, our expert in this area is Polita Paulus.

 

Improved Streaming Data Pipeline Support:

 

10. New Streaming Data Pipelines. The main innovation is the capability to create a concept of materialized tables. Now you can ingest streaming data as row sets. Expert in this area: Tyler Akidau

  • Funny—I presented on Snowflake's Kafka connector at Snowflake Summit 2019. Now it feels like ancient history.

 

Application Development Disruption with Streamlit and Native Apps:

 

11. Low code data application development via Streamlit: The combination of this and the Native Application Framework allows Snowflake to disrupt the entire Application Development environment. I would watch closely for how this evolves. It's still very early but this is super interesting.

12. Native Application Framework: I've been working with this tool for about three months and I find it to be a real game-changer. It empowers data professionals like us to create Data Apps, share them on a marketplace, and even monetize them. This technology is a significant step forward for Snowflake and its new branding.

 

Expanded SnowPark and Python Support:

 

13. Python Support in the Snowflake Data Cloud. More importantly, this is a major move to make it much easier for all “data constituents” to be able to work seamlessly within Snowflake for all workloads including Machine Learning. Snowflake has been making efforts to simplify the process of running data scientist workloads within its platform. This is an ongoing endeavor that aims to provide a more seamless experience.

14. Snowflake Python Worksheets. This statement is related to the previous announcement. It enables data scientists, who are used to Jupyter notebooks, to more easily work in a fully integrated environment within Snowflake.

 

New Workloads. Cybersecurity and OLTP! boom!

 

15. CYBERSECURITY. This was announced a while back, but it is being emphasized again to ensure completeness.

16. UNISTOREOLTP type support based on Snowflake’s Hybrid Table features. This was one of the biggest announcements by far. Snowflake now is entering a much larger part of data and application workloads by extending its capabilities beyond olap [big data. online analytical processing] into OLTP space which still is dominated by Oracle, SQL Server, mysql, postgresql, etc. This is a significant step that positions Snowflake as a comprehensive, integrated data cloud solution for all data and workloads.

 

Additional Improvements:

 

17. Snowflake Overall Data Cloud Performance Improvements. This is great, but with all the other "more transformative" announcements, I'll group this together. The performance improvements include enhancements to AWS capabilities, as well as increased power per credit through internal optimizations.

18. Large Memory Instances. They did this to handle more data science workloads, demonstrating Snowflake's ongoing commitment to meeting customers' changing needs.

19. Data Marketplace Improvements. The Marketplace is one of my favorite things about Snowflake. They mostly announced incremental changes.

 

Quick “Top 3” Takeaways for me from Snowflake Summit 2022:

 

  1. Snowflake is positioning itself now way beyond a cloud database or data warehouse. It now is defining itself as a full-stack business solution environment capable of creating business applications.
  2. Snowflake is emphasizing it is not just data but that it can handle “all workloads” – Machine Learning, Traditional Data Workloads, Data Warehouse, Data Lake, and Data Applications and it now has a Native App and Streamlit Development toolset.
  3. Snowflake is expanding wherever it needs to be in order to be a full data anywhere anytime data cloud. The push into better streams of data pipelines from Kafka, etc., and the new on-prem connectors allow Snowflake to take over more and more customer data cloud needs.

 

Snowflake at a very high level wants to:

 

  1. Disrupt Data Analytics
  2. Disrupt Data Collaboration
  3. Disrupt Data Application Development

 

Want more recap beyond just the features?

 

Here is a more in-depth take on the Keynote 7 Pillars that were mentioned:

Snowflake-related Growth Stats Summary:

  • Employee Growth:

2019:  938 Employees

2022 at Summit:  3992 Employees

  • Customer Growth:

2019:  948 Customers

2022 at Summit:  5944 Customers

  • Total Revenue Growth:

2019:  96M

2022 at Summit:  1.2B

 

Snowflake’s 7 Pillars of Innovations:

 

Let’s go through the 7 pillars of snowflake innovations:

  1. All Workloads – Snowflake is heavily focusing on creating an integrated platform that can handle all types of data and workloads, including ML/AI workloads through SnowPark. Their original architecture's separation of computing and storage is still a key factor in the platform's power. This all-inclusive approach to workloads is a defining characteristic of Snowflake's current direction.
  2. Global – Snowflake, which is based on SnowGrid, is a fully global data cloud platform. Currently, Snowflake is deployed in over 30 cloud regions across the three main cloud providers. Snowflake aims to provide a unified global experience with full replication and failover to multiple regions, thanks to its unique architecture of SnowGrid.
  3. Self-managed – At Snowflake, we are committed to ensuring that our platform remains user-friendly and straightforward to use. This is our priority and we continue to focus on it.
  4. Programmable – Snowflake can now be programmed using not only SQL, Javascript, Java, and Scala, but also Python and its preferred libraries. This is where Streamlit comes in.
  5. Marketplace – Snowflake emphasizes its continued focus on building more and more functionality on the Snowflake Marketplace (rebranded now since it will contain both native apps as well as data shares). Snowflake continues to make the integrated marketplace as easy as possible to share data and data applications.
  6. Governed – Snowflake stated that they have a continuous heavy focus on data security and governance.
  7. All Data – Snowflake emphasizes that it can handle not only structured and semi-structured data, but also unstructured data of any scale.

 

Conclusion:

 

We hope you found this article useful!

Today’s article recapped Snowflake Summit 2022, highlighting feature announcements and innovations. Snowflake is a full-stack business solution environment with seven pillars of innovation: all data, all workloads, global, self-managed, programmable, marketplace, and governed. We covered various topics such as cost governance, data management, external table support, and cybersecurity.

If you want more news regarding Snowflake and how to optimize your Snowflake accounts, be sure to check out our blog.

To Snowsight or Not to Snowsight – TAKE #2

To Snowsight or Not To Snowsight - TAKE #2 - May 2022: 

 

Back in October 2021, I published my original take on reasons why to move to Snowsight. It is in GA but based on my LinkedIn Polls and other DSH chatter, Snowsight is still not being embraced by many experienced Snowflake users it seems. I think if some of these changes below are made then users will begin to embrace the new interface!

 

ORIGINAL POLL ON LINKEDIN:   MAY 2022 POLL ON LINKEDIN: 

Snowsight vs Classic Console Poll October 2021
Snowsight vs Classic Console Poll October 2021

Snowsight vs Classic Console Usage Poll May 2022
Snowsight vs Classic Console Usage Poll May 2022
October 2021
102 Votes - Classic Console wins at 54% to 46%
against Snowsight
May 2022
53 Votes - Classic Console wins at 55% to 45%
against Snowsight - 8 months later

 

In my original Snowsight Article, I made a case that the new functionality in Snowsight makes it much better to use than the Classic Console Especially in these departments: Autosuggest/AutoComplete, Versioning, Sharing Worksheets, and Dashboards. These features take Snowflake's ease of use related to query and code collaboration to a much higher level of efficiency and collaboration.

All of that being said, I still see a huge resistance to using Snowsight in the Data Superhero (DSH) community and among experienced Snowflake Internal Sales Engineers. I guess part of the reason is that humans have a tough time with change. The other issue though is some MAJOR "inefficiencies" and "transparency" issues with Snowsight have been occurring in the last few months.

So, this grid below goes in-depth on the differences between Snowsight and Classic Console. In some cases, I consider that the Classic Console still outperforms Snowsight, but hopefully, the guys at Snowflake are working hard to improve these issues.

 

Snowflake versus Classic Console [May 2022]

Web Interface Functionality Classic Console Snowsight Comments
Show SQL based on GUI Form WINNER One of the great features of the Snowflake Classic Console was you could start creating just about any object in the GUI and then get the code syntax by using the Show SQL functionality.  Why this was left out of Snowsight seems like a tragic miss.
Create Procedure Example
Listing Procedures
WINNER Classic Console doesn’t have this functionality.
Create Function Example
Listing Functions
WINNER Classic Console doesn’t have this functionality.
Create Task Example
Listing Tasks
WINNER Classic Console doesn’t have this functionality.
Create Pipe Example
Listing Pipes
WINNER Classic Console doesn’t have this functionality.
Create Stage Example WINNER Classic Console doesn’t have this functionality.
Create Table Example WINNER Classic Console doesn’t have this functionality.
Create View Example WINNER Classic Console doesn’t have this functionality.
Usage Reporting WINNER While Snowsight has many more screens and better reports, the fact that it has default misleading reports and bugs makes the Classic Console the winner here.
Dependability WINNER Sadly, Snowsight still has weird bugs that happen all the time and you often come to a screen with an endless spinner.
New Worksheet Navigation WINNER One nice new feature of Snowsight is you can open worksheets in new windows.  This was not possible in the Classic Console.  Also, you could not share links with others.
Worksheet Sharing WINNER This is one of my favorite features of Snowsight.  This is not possible within the Classic Console.
Dashboards and Dashboard Sharing WINNER Dashboards are not in the Classic Console
Auto-Complete WINNER This is another one of my favorite features of Snowsight.
Executing highlighted code correctly WINNER Sadly, Snowsight often takes code above or below the highlighted code and tries to execute it which results in errors.  It becomes annoying.
Failure Reporting WINNER Sadly, Snowsight within the query history does a horrible job of surfacing the error while the Classic Console easily displays an error with 1 click.  So so so much better and makes troubleshooting on Snowsight ...  pretty much suck ...

 

Conclusion:

 

There you have it. This is just a highlight of features that I care about. Let me know what you think and what features you still love on the Classic Console work better there than on Snowsight.