Cloud Wars
  • Home
  • Top 10
  • CW Minute
  • CW Podcast
  • Categories
    • AI and Copilots
    • Innovation & Leadership
    • Cybersecurity
    • Data
  • Member Resources
    • Cloud Wars AI Agent
    • Digital Summits
    • Guidebooks
    • Reports
  • About Us
    • Our Story
    • Tech Analysts
    • Marketing Services
  • Summit NA
  • Dynamics Communities
  • Ask Copilot
Twitter Instagram
  • Summit NA
  • Dynamics Communities
  • AI Copilot Summit NA
  • Ask Cloud Wars
Twitter LinkedIn
Cloud Wars
  • Home
  • Top 10
  • CW Minute
  • CW Podcast
  • Categories
    • AI and CopilotsWelcome to the Acceleration Economy AI Index, a weekly segment where we cover the most important recent news in AI innovation, funding, and solutions in under 10 minutes. Our goal is to get you up to speed – the same speed AI innovation is taking place nowadays – and prepare you for that upcoming customer call, board meeting, or conversation with your colleague.
    • Innovation & Leadership
    • CybersecurityThe practice of defending computers, servers, mobile devices, electronic systems, networks, and data from malicious attacks.
    • Data
  • Member Resources
    • Cloud Wars AI Agent
    • Digital Summits
    • Guidebooks
    • Reports
  • About Us
    • Our Story
    • Tech Analysts
    • Marketing Services
    • Login / Register
Cloud Wars
    • Login / Register
Home » 5 Ways to Accelerate Data Sharing Across Clouds
Data

5 Ways to Accelerate Data Sharing Across Clouds

Data Revolution
John FoleyBy John FoleyOctober 21, 2021Updated:December 13, 20214 Mins Read
Facebook Twitter LinkedIn Email
Data Sharing Across Clouds
Share
Facebook Twitter LinkedIn Email

As more databases move to the cloud, and as those databases grow and proliferate, technologies to move data are rapidly advancing to keep up with new requirements. Also, as technology advances, there have been further developments in data sharing across clouds.

The ability to move data has always been fundamental to data management. However, in the cloud, data movement is paramount. This is because hybrid clouds and multi-clouds, by definition, must be architected for data sharing. For instance, data integration, distribution, consumption, migration, and updates all involve data movement in and out of databases.

Developments in Cloud Data Sharing

A range of new and improved technologies enable high-speed, low-friction data distribution. Here are recent developments in five key areas:

1. APIs

On September 28th, Google Cloud announced the general availability of Apigee Integration. This is an API management and integration platform that makes it easier to connect data and applications via APIs. The solution has pre-built connectors for Google’s Cloud SQL and BigQuery databases, with more connectors planned. This makes it easier for developers to tap into data sources when building applications.

2. Change Data Capture

Recently, Google Cloud introduced Datastream, a cloud service that provides change data capture and replication. Change data capture is an automated process that ensures data synchronization and consistency between a primary database and others. Datastream works not only with Google Cloud Storage, but also with Oracle, MySQL, BigQuery, Cloud SQL, and Cloud Spanner. Many other vendors have their own change data capture capabilities.

3. ETL & ELT

Tools to extract, transform, and load (ETL) have long been the standard way of moving data from databases, apps, and other sources into data warehouses. In the cloud, data transformation sometimes happens as the last step, which is reflected in a slightly different abbreviation, ELT. Either way, these capabilities are essential for data integration as well as establishing pipelines of quality-controlled “clean data” for analysis.

4. Data Fabrics

InterSystems, Oracle, Talend, and other vendors offer data fabrics that weave together data wherever it may be. For instance, this includes databases, apps, and IoT. These many-to-many middleware fabrics support data streaming and also replication for real-time analytics.

5. Distributed Databases

A growing number of new-generation cloud databases are designed to be distributed with portions of the database running on different servers and potentially across geographic regions. A distributed database functions as one logical database, albeit with a distributed architecture. Microsoft’s Azure Cosmos DB, CockroachDB, DataStax Astra DB, and YugabyteDB are examples of distributed databases.

Follow the Money

Data movement makes data management more complex. It requires added attention to security and governance. Yet, data-in-motion technologies are essential to data lakes, data warehouses, data clouds, and other data-driven business operations.

Venture investment is pouring into these products and services. For example, Matillion, a data-integration and ETL specialist, just pulled in $150 million in series E funding, for a total of $310 million in venture capital at a market valuation of $1.5 billion.

Also, Fivetran recently announced plans to acquire HVR for $700 million, bringing together Fivetran’s automated data integration and HVR’s data replication. According to Fivetran, the goal is to make data access “as simple and reliable as electricity.”

Fivetran also revealed $565 million in additional venture funding, led by Andreessen Horowitz. So, that brings its total funding to $730 million at a $5.6 billion valuation.

Beyond Big Data

As databases expand to petabytes and beyond, standard approaches to data distribution and replication may not be sufficient. New solutions for very large databases (VLDBs) will ultimately be required.

In one interesting example, a project is underway by a handful of U.S. national labs to develop processes that support petabyte data transfers. The initiative, the Petascale DTN Project, involves data transfer nodes (DTNs) connected by high-speed networking. An article in The Next Platform describes the project, with diagrams and a link to a paper by the team behind it.

When gigabyte-speed connections aren’t enough, it’s time to bring in the 18-wheelers. My favorite example of big-time data transfer is AWS’s Snowmobile. This is a 45-foot-long shipping container that can move 100 petabytes at a time by driving down the road to an AWS data center. AWS says it could take 20 years to transfer that much data over a 16-Gbps connection.

Yet, data transfer by truck can only be an interim solution. So, state-of-the-art technologies must continue to advance to take on bigger workloads at faster speeds.

Data transfer gets even more complicated when entire databases are moved to the cloud. For more on that topic, see my article, “Cloud Vendors Confront ‘Highest Risk’ Projects: Database Migration.”

No one said data movement would be easy. But it’s absolutely essential, as the best way to unlock the tremendous value of data is through distribution and sharing.

Data Revolution featured
Share. Facebook Twitter LinkedIn Email
John Foley
  • LinkedIn

John is founder of the Cloud Database Report and host/Sr. Analyst for the Data Revolution channel on Acceleration Economy. For more than 20 years, John has covered database management systems and data warehouses, and the ongoing challenges businesses face with data quality, policy, performance, and scale. He also writes and podcasts regularly about the latest trends and innovations in cloud database platforms, including data integration, analytics, machine learning, data transformation, autonomous management, and hybrid clouds. John digs into real-world use cases and best practices that lead to data-driven insights and actions. Recently he helped drive strategy as a communications leader at Oracle, IBM, and MongoDB. That first-hand industry experience informs his perspective and analysis.

Related Posts

Snowflake Ventures Invests in Diskover to Tackle Unstructured Data at Scale

June 25, 2025

Microsoft, Gong Detail How AI and Integration Partnership Drives Higher Sales Performance

June 24, 2025

AI Industrialization of America Rolls On as AWS Plans Data Centers in Coal Country

June 24, 2025

Snowflake to Acquire Crunchy Data to Power Agentic AI with PostgreSQL Integration

June 24, 2025
Add A Comment

Comments are closed.

Recent Posts
  • Snowflake Ventures Invests in Diskover to Tackle Unstructured Data at Scale
  • Microsoft, Gong Detail How AI and Integration Partnership Drives Higher Sales Performance
  • AI Industrialization of America Rolls On as AWS Plans Data Centers in Coal Country
  • Snowflake to Acquire Crunchy Data to Power Agentic AI with PostgreSQL Integration
  • AWS Data Centers Opening in Coal Country: ‘AI-Industrialization’

  • Ask Cloud Wars AI Agent
  • Tech Guidebooks
  • Industry Reports
  • Newsletters

Join Today

Most Popular Guidebooks

Accelerating GenAI Impact: From POC to Production Success

November 1, 2024

ExFlow from SignUp Software: Streamlining Dynamics 365 Finance & Operations and Business Central with AP Automation

September 10, 2024

Delivering on the Promise of Multicloud | How to Realize Multicloud’s Full Potential While Addressing Challenges

July 19, 2024

Zero Trust Network Access | A CISO Guidebook

February 1, 2024

Advertisement
Cloud Wars
Twitter LinkedIn
  • Home
  • About Us
  • Privacy Policy
  • Get In Touch
  • Marketing Services
  • Do not sell my information
© 2025 Cloud Wars.

Type above and press Enter to search. Press Esc to cancel.

  • Login
Forgot Password?
Lost your password? Please enter your username or email address. You will receive a link to create a new password via email.