Back FAQ

Frequently Asked Questions

The following are frequently asked questions regarding migrating workloads from Netezza to Hadoop/Spark using the Impetus Workload Migration Solution.

 

The Impetus Workload Transformation Solution is an accelerated service where the Impetus team actively collaborates with the customer. We offer workload transformation as a service for the following reasons:

  • The automated conversion percentage is dependent on the customer’s unique environment and requirements

  • Our underlying configurable, self-learning, and extensible grammar engine needs time to learn these new patterns.

Depending on your technology stack, the new converted environment may allow running the same BI tools. However, the Impetus Workload Transformation accelerators may automatically alter the underlying queries to ensure minimal end-user impact.

The Impetus Workload Transformation Solution also supports the transformation of Oracle Web Report, SSRS, and other reporting tools to modern Agile BI tools like Tableau along with the Netezza transformation.

Impetus sets up the data access based on the consumer’s behavior patterns such as the following:

  • Data access tools such as Spark, Hive JDBC, and ODBC

  • BI and analytical tools such as Tableau and SAS

  • Ad hoc querying tools such as Hue, Zeppelin, JDBC, and ODBC

  • Sub-second BI response tools such as Kyvos Insights

  • Interactive/search/discovery applications such as HBase, Phoenix, Solr, and more.

The converted code is SQL, which can be managed by your current SQL resources.

  • You can also use SQL editors and interfaces offered by Hadoop or cloud, such as Apache Zeppelin or Hue, and other tools, such as Toad, via JDBC access. 

  • For stored procedures, the Java/Scala/Python/Shell lightweight code is accompanied by Maven projects for easy maintenance.

The overall pricing depends on the number of unique queries running in Netezza. Impetus can establish the cost after the assessment phase.

Impetus recommends launching the transformation with a risk-free, comprehensive 4-8-week assessment, which includes a pilot (to prove the automagical conversion aspect, build confidence, and demonstrate ROI) and an end-to-end use case. The POC takes 6-9 weeks where we identify how much time it would take to transform all the workloads.

We’ve observed that the Impetus Workload Transformation Solution can save up to 70% of time compared to manual transformation. This is because we automate the entire process, end-to-end, from assessment and transformation to validation and execution.

Depending on the tool and usage pattern (push down query or other), the ETL tool will either be integrated with the new environment or, if needed, replaced with a Hive/Spark-based ETL.

No. There is no dependency on the Impetus tool after the transformation is complete.

The Impetus Workload Transformation Solution ensures that the end-to-end SLA is matched or will provide the Hadoop cluster size to match the same. If the Hadoop cluster size option is needed, it also includes cost comparisons.

The Impetus Workload Transformation Solution converts stored procedures in two phases:

  • First, it converts the SQL portion to HiveQL and Spark SQL.

  • Then it converts the procedures, such as loops and cursors, to a lightweight Java or Scala wrapper.

The Impetus Workload Transformation accelerator supports Ab Initio Transformation to Spark/Hadoop code. The output code can be managed through the application UI, Talend, or other tools that have Spark/Hive integrations, or AWS EMR/AWS Glue/AWS Pipeline in cloud.

All the UDFs and stored procedures are converted or replaced by a rich set of Spark/Hive UDF or reusable components. These UDF can be used for Netezza workloads or any greenfield development.

The Impetus Workload Transformation Solution supports AWS, Azure, and generic IaaS cloud providers that have Hive 1.3+ and/or Spark 1.6+ in the form of IaaS or PaaS, such as EMR and HDInsights.

The Impetus Workload Transformation Solution supports all Hadoop distros, including HDP, CDH, MapR, IBM Hadoop, Oracle Hadoop Appliance, AWS EMR, and Azure HDInsight that support Hive 1.3+ and/or Spark 1.6+.

The Impetus Workload Transformation Solution converts SQL and any Netezza UDFs to Hive/Spark-compatible SQL queries. Impetus leverages a configurable, self-learning, and extensible grammar engine for the conversion.

The Impetus Workload Transformation Solution is an accelerated end-to-end service that enables the transformation of Netezza to a low-cost and scalable Hadoop/Spark platform. In addition, it ensures that ingestion and consumption tools and processes, SLA, and DevOps work seamlessly after the transformation is complete.

The cloud serves as a low-cost scalable environment that can lower capital investments such as hardware, operational costs, on-demand usage-based costs, and outsourced infrastructure costs. The cloud also allows for the possibility of additional capabilities, such as advanced analytics, machine learning, streaming data, unstructured data, and more.

Many Netezza users are exploring various growth options to achieve greater scalability and tighter integration with their enterprise data lake and cloud computing initiatives.

IBM is offering its customers the IBM PureData System or IBM Db2 Analytics Accelerator for analytics. For cloud-based deployments, IBM has IBM DashDB. But these are expensive options and carry the risk of vendor lock-in.

Netezza users want to move their increasingly important enterprise analytics processing to a new platform — choosing to shift away from dependence on a single vendor architecture in order to take advantage of the flexibility that the open source (Hadoop/Spark) revolution is delivering.

Hadoop is an open source, low-cost, scalable data warehouse. It also has more capabilities to match the evolving data landscape such as advanced analytics, machine learning, streaming data, unstructured data, and more.

The Next Chapter for Netezza Begins Here