Major new version delivers multicloud data interoperability, automation, scalability and security.
Cirata, the company that automates Hadoop data transfer and integration to modern cloud analytics and AI platforms, today announced the launch of Data Migrator 3.0, with production-ready support for Apache Iceberg, expanded capabilities for Databricks Delta Lake, significant enterprise security enhancements and comprehensive extensibility. This major new version of Cirata’s technology empowers organizations to adopt open data architectures while improving scalability, automation, and security for their multicloud data management and analytics needs.
Also Read:Â Cloud Management Platform, Emma Secures $17Million in Series a Funding
“Organizations understand the need to avoid the lock-in, technology constraints and costs that come with isolated data storage and formats. They need solutions that make it possible to use their largest data assets wherever they are created, in a variety of formats, and with whatever future technologies emerge. They want data interoperability,” said Paul Scott-Murphy, Chief Technology Officer at Cirata. “Our production-ready, direct support for open table formats like Apache Iceberg and Delta Lake eliminates the constraints of closed data architectures, even if you have petabytes of data held in formats or locations that previously required lengthy, complex and risky efforts to modernize. Data Migrator 3.0 is a significant advancement for organizations that want to future-proof their data management, analytics and AI strategies.”
Also Read:Â CIO Influence Interview With Karthik Ranganathan, co-founder and co-CEO of Yugabyte
Cirata Data Migrator 3.0 introduces a range of new features that streamline data transfer between on-premises and cloud environments, reduce operational complexity, and increase data interoperability. Optimized for handling even the largest datasets, Data Migrator 3.0 removes computational overhead, offering improved performance and flexibility for enterprise-grade data transfers at scale. The improvements are particularly significant for enterprises adopting Apache Iceberg and Databricks Delta Lake within their data ecosystems. Key new features include:
- Data Interoperability with Support for Apache Iceberg: Organizations can now leverage Data Migrator to automate the migration of large-scale datasets in other formats to environments that use Apache Iceberg, supporting open data interoperability at scale.
- Expanded Support for Databricks Delta Lake: The new Data Migrator version further optimizes Databricks Delta Lake integration, improving performance and resource efficiency when transforming source Hive tables into Delta Lake formats. This ensures seamless data availability and scalability across cloud analytics environments.
- Enhanced Data Automation: The solution’s new internal scheduler eliminates the need for manual intervention when configuring recurring migrations, allowing users to define custom schedules for data transfers. This automation simplifies complex enterprise data workflows.
- Advanced Data Consistency: Cirata introduces enhanced verification mechanisms that simplify data consistency measures during migrations, providing robust, accurate data matching across storage systems without compromising performance.
- Enterprise-Grade Security: With built-in support for Hashicorp Vault, Data Migrator 3.0 ensures secure storage and management of credentials. It also expands role-based access control (RBAC) across all components, enhancing security and control for enterprise users.
Cirata Data Migrator is a fully automated solution that automates Hadoop data transfer and integration and moves on-premises HDFS data, Hive metadata, local filesystem, or cloud data sources to any cloud or on-premises environment, even while those datasets are under active change. Cirata Data Migrator requires zero changes to applications or business operations and moves data of any scale without production system downtime, business disruption, and with zero risk of data loss. Migration targets supported include the Hadoop Distributed File System, Alibaba Cloud Object Storage Service, Amazon S3, Azure Data Lake Storage Gen 2, Google Cloud Storage, IBM Cloud Object Storage and Oracle Object Storage.
[To share your insights with us as part of editorial or sponsored content, please write to psen@itechseries.com]