Professional Documents
Culture Documents
Migration Guide
About Bigdata Dimension Labs
Bigdata Dimension Labs was built on the ideal that every business should be
positioned to make more profitable, data-driven decisions. To achieve this, they
have partnered with the best modern data sharing technologies in the world and
are recognized as centers of excellence for:
Bigdata Dimension Labs have empowered organizations such as 3M, Agile and
Bill Gates Foundation with 360° performance overview, real-time analytics for
better decision-making, standardized reporting and faster data access and
processing.
Snowflake reduces complexity with built-in Create a single source of truth to easily store,
performance so there’s no infrastructure integrate and extract critical insight from
to tweak, no knobs to turn and no tuning petabytes of structured and semi-structured
required. data (JSON, XML, AVRO).
Snowflake extends the data warehouse to the Snowflake supports the tools millions of
Data Sharehouse™, with direct, governed and business users already know how to use today.
secure data sharing within seconds, so enter-
prises can easily forge one-to-one,
one-to-many and many-tomany data sharing
relationships.
• What workloads are outdated and require a complete • Or, your current system uses welldesigned data and process-
es using standard ANSI SQL.
overhaul?
• What new workloads would you like to add that would • Or, you have timescale pressures to move off legacy equip-
ment
deploy easier in Snowflake?
Factors that lead to a lift-and-shift approach may • Or, you’ve chosen to focus on new business requirements
focus on what you already have: rather than reworking legacy processes.
Snowflake has an optimizer built from the ground up and Disaster recovery
architected for MPP and the cloud. Snowflake understands
Teradata has several disaster recovery scenarios. Many of
parallel execution plans and automatically optimizes them,
them require the purchase of another system, and purchasing
relieving our customers of this task. This means there are
software such as Unity to implement these scenarios. With
no indexes in Snowflake so you don’t have to migrate your
Snowflake, none of this is necessary. Snowflake leverages
USIs or NUSIs.
many of the built-in features of the cloud, such as the auto-
Workload management matic replication of data built into AWS. Snowflake is auto-
matically synced across multiple AWS availability zones, with
Workload management is unnecessary in a Snowflake more zones and other cloud computing platforms to come
environment due to its multicluster architecture, which available on a consistent basis. There is no work on your part
allows you to create separate virtual warehouses for your to establish this.
disparate workloads so as to avoid resource contention
completely. Separate dev/test environment
With Teradata, to perform development and testing, you need
Statistics collection additional servers, which means an additional cash outlay just for
Snowflake automatically captures statistics, relieving DBAs the hardware plus the time to configure the hardware. Not so
from having to set up jobs to collect statistics for perfor- with Snowflake as you can simply create another database in your
mance tuning. It’s automatic in Snowflake, so you no longer account and set it up for any purpose you need, such as dev or
have to remember to add new tables to the process when test. In addition, with Snowflake’s zero-copy clone feature, you
can instantly populate those databases with complete copies of
your data warehouse grows.
production data for no additional cost. With Teradata, you have to
endure the painstaking process of exporting your production
data from one system to import it to your dev or test server.
Once you have built your objects in Snowflake, you'll . If you choose to move your data manually, you will need to
want to move the historical data already loaded in your extract the data for each table to one or more delimited flat
Teradata system over to Snowflake. To do this, you can files in text format using Teradata Parallel Transporter (TPT).
use a third-party migration tool (see Appendix A), an ETL
(extract, transform, load) tool or a manual process to . Then upload these files using the PUT command into an
move the historical data. Choosing among these Amazon S3 staging bucket, either internal or external. We
options, you should consider how much data you have recommend these files be between 100MB and 1GB to take
to move. For example, to move 10s or 100s of advantage of Snowflake’s parallel bulk loading.
terabytes, or even a few petabytes of data, a practical
approach may be to extract the data to files and move it . After you have extracted the data and moved it to S3, you can
via a service such as AWS Snowball. If you have to move begin loading the data into your table in Snowflake using the
100s of petabytes or even exabytes of data, AWS Snow- COPY command. You can check out more details about our
mobile is likely the more appropriate option. COPY command in our online documentation.
Migrating BI tools
Many of your queries and reports are likely to use an exist-
ing business intelligence (BI) tool. Therefore, you’ll need to
account for migrating those connections from Teradata to
Snowflake. You’ll also have to test those queries and reports
to be sure you’re getting the expected results.
We have also built ingestion tool to automate and accelerate the migration process.
Whether your organization is fully staffed for a platform migration or you need additional expertise, Bddlabs have the skills
and tools to accelerate your journey to cloud-built data analytics, so you can reap the full benefits of Snowflake quickly. To
find out more, please contact the Bddlabs sales team at
If you still want to learn more, I encourage you to visit BDDLabs.com to gain more information
about modern cloud data sharing. You can also access documentation, view webinars, browse our
offers, view scoops of upcoming events, and get support. We also invite you to schedule a free
demo of our data sharehousing technology so your business can get started right away!
Email: solutions@bddlabs.com
Phone: 888-856-2238
At Bigdata Dimension Labs we make those frameworks and solutions. Our cloud-powered framework and solution enables
enterprises to become data driven, increase performance and provide the path for automagic modernization. Our solutions also
monetize the data you share externally (DaaS) and make the best use of the data others share with you - all with our flexible and
affordable cloud solutions.
The ultimate Snowflake cloud data warehouse companion Application modernization platform which is out of the
which uncovers the insights of the snowflake platform on box cloud native ready for Azure, AWS or Google – Built
Azure and AWS. with ageless application modernization framework for
the cloud that you will never have to code it again.
Application modernization platform which is out of the box Modern Talend ingestion & code standard platform to
cloud native ready for Azure, AWS or Google – Built with help you migrate from Teradata, Netezza, have a common
ageless application modernization framework for the cloud API ingestion engine.
that you will never have to code it again.
Modern Talend job scan platform that validate the code Modern Talend job scan platform that validate the code
standards and accelerate deployment standards and accelerate deployment Modern self
service providers and consumer’s data sharing platform
that eliminates the need to write API, FTP, Email, ETL and
manual processes to share the data.
• List of database objects that need to be migrated • Method for identifying process dependencies for data sets
• List of processes and tools that populate and pull data • Documentation of process dependencies for data sets
from the existing solution
IDENTIFY THE MIGRATION TEAM
• List of security roles, users and permissions • List of migration team members and roles
• List of Snowflake accounts that exist or need to be creat- • Contact information for all team members
ed
DEFINE THE MIGRATION DEADLINES
• Frequency of security provisioning processes AND BUDGET
• Document the existing solution into an as-is architecture • List of business expectations for the migration deadline
diagram
• Documented budget allocated for the migration project
DETERMINE A MIGRATION APPROACH
• List of processes to migrate as is • Completed estimation template for Snowflake virtual
warehouses
• List of processes that need re-engineering
DETERMINE THE MIGRATION OUTCOMES
• List of processes that need to be fixed
• List of high-level desired outcomes and assumptions at the
• Draft of migration deliverables completion of the migration
• To-be architecture diagram • Documented plan for communicating the migration project
wins to stakeholders
CAPTURE THE DEVELOPMENT AND
DEPLOYMENT PROCESSES
BYTEINT BYTEINT
SMALLINT SMALLINT
INTEGER INTEGER
BIGINT BIGINT
DECIMAL DECIMAL
FLOAT FLOAT
NUMERIC NUMERIC
CHAR Up to 64K CHAR Up to 16MB
VARCHAR Up to 64K VARCHAR Up to 16MB
LONG VARCHAR Up to 64K VARCHAR Up to 16MB
CHAR VARYING(n) CHAR VARYING(n)
REAL REAL
DATE DATE
TIME TIME
TIMESTAMP TIMESTAMP
BLOB BINARY Up to 8MB
CLOB VARCHAR Up to 16MB
BYTE BINARY
VARBYTE VARBINARY
GRAPHIC VARBINARY
JSON VARIANT
ARRAY ARRAY