Stambia component for Snowflake

Snowflake is an analytic data warehouse provided as Software-as-a-Service. In comparison to a traditional data warehouse offering, Snowflake is quite faster, easier to use and flexible.

The Stambia Component for Snowflake is built around the same philosophy to provide an easier, faster and flexible integration with Snowflake. Stambia's unified approach makes it easy for the users to get started with the Integration tasks and helps reduce the time to market significantly.

Discover how Stambia's unified approach makes it easy for Snowflake

Snowflake Component : different Use Cases

Simplicity and Agility

Customers choose Snowflake as their Data Warehouse solution, due to its ease of use, ability to scale and faster querying.

Below are some of the key features:

  • No hardware to select, install or configure
  • No software to install or configure
  • Maintenance & Management handled by Snowflake

These factors provide simplicity and agility in Implementing a data warehouse. Therefore, it is very important to have an Integration Layer, that as well, is simple and agile.

Snowflake agility simplicity
 

Migrate your On-Premise Data to Snowflake

Snowflake Data Migration With Stambia

Getting started with Snowflake means first moving your existing on-premise data to the new schema in Snowflake warehouse. Data Migration projects require a lot of planning as it involves moving your data phase by phase with conversion, assessment, re-conciliation all in place.

"Through 2019, more than 50% of data migration projects will exceed budget and timeline and/or harm the business, due to flawed strategy and execution"
source : Gartner

Data migration also requires automating and simplifying a lot of steps to bring agility in your implementations.

 

Integrate all your data

As part of an organizations data strategy, the ability to connect and read data from different technologies and file formats is very important. The Integration layer should be able to bring data from heterogenous technologies, and in doing so, should remain as consistent as possible to avoid ending up in a web of different kinds of isolated integration processes.

During the evolution of an organizations IT, the data strategy goes through a lot of changes, hence the architecture keeps evolving with time. An Integration layer, that is able to adapt to the needs and changes, is key to taking quick decision and reducing time to market.

Snowflake integrate all data
 

Performance and Scalability

Stambia performance and scalability

Snowflake processes queries using “virtual warehouses”. Each virtual warehouse is an MPP compute cluster composed of multiple compute nodes allocated by Snowflake from a cloud provider. Therefore, Snowflake can scale automatically to handle varying concurrency demands.

An ELT tool is most suitable when it comes to data integration. An ELT tool will complement Snowflake as it will use Snowflake's native capabilities and leverage it's performance to have optimized workloads.

Features of Stambia Component for Snowflake

Stambia component for Snowflake provides simple and easy way of connecting to a Snowflake warehouse and performing the data loads with an ELT approach.

Given below are some of it's features:

Flexible with any Cloud platform

Snowflake can be set-up on cloud platforms like Amazon AWS and Microsoft Azure. Data from various source systems can be staged in either of the following:

  • Amazon S3 Bucket as External Stage (or Named External Stage)
  • Microsoft Azure BLOB Storage (or Named External Stage)
  • Snowflake Internal Stage
Snowflake global schema storage
Snowflake Stambia Designer metadata

Stambia component for Snowflake provides users with a one time metadata configuration to choose either of the Stages on the Metadata itself. This reduces design time as inside the mapping the default stage (configured in the metadata) will be used.

 

Consistent in the way you design

To cater to various integration needs, involved in different types of evolving architectures, Stambia provides users with a Universal Data Mapper so as to keep designs consistent across different types of integration between Snowflake and various source technologies.

Stambia Mappings are easy to design due to the presence of "templates" specifically designed for Snowflake. Users actually focus on the business rules, to be executed while moving the data to Snowflake, in the mappings, and on execution the templates execute the native Snowflake scripts and sql orders.

This increases productivity as well as skips all the redundant steps of configuring the intermediate stages & integration approaches, in each mapping. On the other hand, it also provides the flexibility to change the Intermediate Stage in both Metadata and Mapping Level.

Snowflake Stambia Designer Mapping template
 

Quick migration of your on-premise data

Snowflake Automatic Data Migration With Stambia

Data Migration, as discussed above is the starting point in a Snowflake implementation. Stambia Replicator template designed for Snowflake helps migrating the one-time or history data from your on-premise data warehouse to Snowflake without the need to design mapping for each entity.

Apart from increasing productivity, the Replicator templates can be adapted to specific migration needs as these templates are open for customization. Thus, as part of your data migration strategy you can choose what to migrate, perform data conversions, reconciliations etc.

In order to know more about Data migration, take a look at a video about our "Replicator tool":

 

Ownership of Data Integration

One of the most important factors when choosing a Data Integration solution is the cost of ownership. Stambia provides simplicity in understanding the cost of ownership. With it's ELT approach, Stambia does not require specific hardware and does not restrict users to deploy in different environments. As well, there is no specificity in number of Sources or data volumes.

Being a unified solution, Stambia also reduces cost as the users doesn't need to look for a different solution in terms of changes in Use Cases and Architectures. The components of Stambia are simple to understand and easy to use, thereby, ensuring quick learning curve.

Stambia Mastered trajectory

Technical specifications and prerequisites

SpecificationsDescription

Protocol

JDBC, HTTP

Structured and semi-structured

XML, JSON, Avro (coming soon)

Storage

Storages supported to store temporary files for performances optimizations:
  • Amazon S3 Bucket as External Stage (or Named External Stage)
  • Microsoft Azure BLOB Storage (or Named External Stage)
  • Snowflake Internal Stage
Connectivity

You can extract data from :

  • Any relational database system such as Oracle, PostgreSQL, MSSQL, ...
  • Any NoSQL database system such as MongoDB, Elastic, ...
  • Any Cloud system such as Amazon Web Service (AWS), Google Cloud Platform (GCP), Microsoft Azure, ...
  • Any ERP applications such as SAP, Microsoft Dynamics, ...
  • Any SAAS applications such as Salesforce, ...

For more information, consult the technical documentation

Standard features

  • Reverse : The database structure can be reversed in a dedicated Metadata
  • DDL /DML Operations : DML/DDL operations can be performed on the database, such as: Insert, Update, Select, Delete, Create or Drop
  • Integration methods : Append, Incremental Update
  • Staging : The databases can be used as a staging area to make data transformation, mutualization, ... The following modes are supported:
    • staging as subquery
    • staging as view
    • staging as table
  • Reject : Reject rules can be defined to filter or detect data that does not fulfill the rules during integrations.
    • Three type of rules can be created: Fatal, Warning, Reject
    • Depending on the specified type, the rejected data will not be handle the same way
    • Rejects of previous executions can also be recycled
  • Replication : Replication of databases is supported. The replication source can be any rdbms, Flat files or XML files.
Advanced features
  • Slowly Changing Dimension Integrations : Integrations can be performed using Slowly Changing Dimension (SCD)
  • Load methods :
    • Generic load
    • Snowflake COPY loader
Stambia Version From Stambia Designer s18.3.1
Stambia Runtime version From Stambia Runtime S17.3.0
Notes
  • Development environment: any system supporting Stambia Designer (Windows, MacOS, Linux)
  • Runtime environment: any system supporting a Java environment (JRE 1.8 minimum)

Want to know more ?

Consult our resources

Anonymisation
Ask advices to our experts in Data integration.
Contact us
Anonymisation
Discover our training and certifications
To know more about it
Anonymisation
Your customized demonstration
Get your demo