Enterprise network infrastructure connecting on-premises data center to cloud services through dedicated fiber connections for low-latency hybrid integration
Published on May 17, 2024

Contrary to common belief, achieving low-latency hybrid cloud performance is not about choosing a single “best” connection, but about mastering the end-to-end data path across both legacy and cloud environments.

  • A dedicated connection like AWS Direct Connect is not a magic bullet; its performance is dictated by precise BGP configuration and overcoming on-premises bottlenecks.
  • The greatest risks to performance and security often lie in subtle firewall misconfigurations and inefficient routing that sends local traffic on a costly round-trip to the cloud.

Recommendation: Shift focus from the “pipe” to the “packet journey.” Adopt a Zero Trust security model and use BGP attributes intentionally to enforce routing policies that treat your hybrid network as a single, unified fabric.

For any network architect tasked with bridging a legacy on-premises data center with the public cloud, the promise of infinite scalability often crashes into a harsh reality: crippling latency. You’ve designed a state-of-the-art architecture in AWS or Azure, yet applications feel sluggish, data transfers crawl, and the entire system feels less like a seamless extension and more like two distant islands connected by a fragile bridge. The business demands performance, but the network seems to have its own plans.

The standard advice is predictable: implement a dedicated connection like AWS Direct Connect or simply “optimize your VPNs.” While these are components of the solution, they are far from the whole story. This approach overlooks the fundamental physics of data transfer, the hidden complexities of routing protocols, and the insidious security risks that emerge when two disparate network philosophies are bolted together. It treats the symptoms—high latency—without diagnosing the underlying disease: a disjointed network fabric.

But what if the key to low latency wasn’t just about a bigger, faster pipe? What if it was about mastering the complete, end-to-end packet journey? This guide adopts that very perspective. We will move beyond a simple product comparison to dissect the architectural discipline required for true hybrid integration. We will explore why data has “gravity,” how to control its path with surgical precision, and how to build a secure, unified network that performs as a single, cohesive entity, finally delivering on the true promise of the hybrid cloud.

This article provides a comprehensive roadmap for network architects. The following sections break down the critical challenges and strategic solutions for building a high-performance, low-latency hybrid network.

Why moving petabytes to the cloud is slower than you think?

The first miscalculation in many hybrid cloud strategies is underestimating the sheer “gravity” of data. While cloud storage is virtually limitless, the physical constraints of network bandwidth are very real. The idea of quickly shifting petabytes of data from an on-premises data center to a cloud provider over a standard internet connection is a logistical fallacy. The network physics are unforgiving; even with a relatively fast connection, the time required can be staggering. For context, transferring large datasets can take an astonishing 120 days for just 100TB over a typical 100Mbps connection.

This challenge is not merely about bandwidth; it’s about the entire data transfer pipeline. Large-scale migrations are complex projects that require significant optimization at both the source and destination. As demonstrated by TIM Brasil’s successful petabyte-scale migration to Google Cloud, success hinges on meticulous planning. Their team had to optimize the number of transfer agents and fine-tune networking settings just to saturate a 20 Gbps Partner Interconnect link. This underscores a critical lesson: the advertised speed of your connection is a theoretical maximum, not a guaranteed throughput.

Factors like protocol overhead, disk I/O on the source servers, and the number of individual files all impact the actual transfer rate. A migration involving millions of small files will perform drastically differently than one involving a few massive files, even with the same total data volume. Therefore, before even considering the long-term architecture, architects must perform a realistic assessment of the initial data seeding and ongoing synchronization requirements. This initial step often reveals that a dedicated, high-throughput connection is not a luxury, but a baseline requirement for any serious hybrid initiative.

How to set up AWS Direct Connect for consistent throughput?

Once the reality of data gravity sets in, AWS Direct Connect often emerges as the logical solution. It provides a private, dedicated physical connection between your on-premises infrastructure and AWS. However, procuring a Direct Connect circuit is only the beginning. Achieving consistent, low-latency throughput is not automatic; it is the result of precise network engineering, primarily centered on the Border Gateway Protocol (BGP). This is where the work of a network architect truly begins.

The connection is essentially a Layer 2 link. To make it useful, you must establish a Layer 3 BGP session between your routers and the AWS Direct Connect endpoint. This session is what allows you to exchange routing information, making your on-premises network and your Amazon VPCs aware of each other. The configuration involves setting up virtual interfaces (VIFs), configuring BGP with MD5 authentication, and ensuring 802.1Q VLAN encapsulation is correctly implemented. A single misstep here can lead to an unstable connection or a complete failure to establish peering.

As the diagram illustrates, the architecture relies on this BGP session to create a predictable and high-performance path. To properly configure this, you must follow a clear sequence of steps:

  1. Create a Connection: Order a connection at an approved AWS Direct Connect location to establish the physical network link from your premises to an AWS Region.
  2. Configure On-Premises Routers: Ensure your edge routers support BGP and BGP MD5 authentication. You must also configure the physical port with the correct 802.1Q VLAN encapsulation to match the VIF settings.
  3. Set Up Virtual Interfaces (VIFs): In the AWS console, configure a VIF for your connection. This will be a private VIF for accessing a VPC or a transit VIF for connecting to a Transit Gateway, which is essential for more complex, multi-VPC architectures.
  4. Establish BGP Peering: Using the VLAN ID, peer IP addresses, and BGP authentication key provided by AWS, establish the BGP session between your router and the AWS endpoint.
  5. Verify Connectivity: Thoroughly test the path using tools like `traceroute` and `ping`. The network trace should clearly show traffic traversing the Direct Connect identifier, confirming that you are no longer routing over the public internet.

Only after this meticulous configuration and verification process will Direct Connect deliver on its promise of consistent, low-latency performance. It is an active engineering task, not a passive subscription service.

VPN vs dedicated leased lines: which is more cost-effective?

While Direct Connect offers the gold standard in performance, it comes with higher costs and longer deployment times. This leads every network architect to a critical decision point: for which workloads is a Site-to-Site VPN sufficient, and when is a dedicated line non-negotiable? The answer lies in a nuanced analysis of cost, performance, and use case, as a purely cost-based decision can be misleading.

A Site-to-Site VPN uses the public internet to create an encrypted tunnel between your on-premises network and your cloud environment. Its primary advantages are speed of deployment and lower upfront costs. You can establish a VPN connection in minutes or hours, making it ideal for pilot projects, temporary needs, or initial, low-volume data migrations. However, its performance is inherently variable. Since it relies on the public internet, it is susceptible to unpredictable latency and packet loss. Furthermore, standard AWS VPNs have an aggregate bandwidth limit of 1.25 Gbps per tunnel, which can be a significant constraint.

In contrast, AWS Direct Connect provides a private, predictable, and high-throughput connection. It bypasses the public internet entirely, resulting in consistent low latency. This makes it the superior choice for latency-sensitive applications like real-time services, VoIP, or high-frequency data streaming. The cost savings can also be substantial for data-heavy workloads. For example, using a dedicated connection can lead to $481 per month in savings for every 10TB of outbound data compared to internet transfer rates.

The following table, based on a recent comparative analysis, summarizes the key trade-offs:

VPN vs. AWS Direct Connect Comparison
Criterion Site-to-Site VPN AWS Direct Connect
Maximum Bandwidth 1.25 Gbps (aggregate limit) Up to 100 Gbps
Network Latency Variable (public Internet-dependent) Consistent, predictable low-latency
Deployment Time Minutes to hours Days to weeks
High Availability Built-in (dual tunnels by default) Requires multiple connections or VPN backup
Encryption Encrypted by default (IPsec) Not encrypted by default (MACsec available at select locations)
Best Use Case Quick setup, lower upfront cost, pilots/migrations, temporary connectivity Latency-sensitive workloads, high-throughput requirements, consistent data transfers, compliance needs

Ultimately, the most cost-effective strategy is often a hybrid of both. Use VPNs for non-critical workloads and as a backup for Direct Connect, while reserving the dedicated line for production traffic and applications where performance is paramount. The decision is not “either/or” but rather about right-sizing the connection to the business requirement.

The firewall misconfiguration that exposes internal networks

Achieving a high-speed, low-latency connection between your on-premises data center and the cloud is a significant accomplishment. However, this new, wider digital highway can also become a fast lane for security threats if not properly managed. The single most common and dangerous vulnerability in a hybrid environment is not a sophisticated zero-day exploit, but a simple firewall misconfiguration. In the complex world of hybrid networking, it’s frighteningly easy to create a rule that inadvertently exposes your entire internal network.

Traditional network security was built on a “castle-and-moat” model, with a strong perimeter firewall protecting a trusted internal network. The hybrid cloud shatters this model. The security perimeter is no longer a clear line but a porous, distributed surface that spans both on-prem and cloud environments. Cloud misconfigurations are a leading factor in security incidents, and recent studies show they are responsible for 19% of all breaches, with an average cost of $4.41 million per incident. The risk is not theoretical; it is a clear and present financial danger.

A common misconfiguration involves creating overly permissive firewall rules. For instance, an engineer troubleshooting a connectivity issue might temporarily open a wide range of ports with an “allow any” rule from a cloud-based source to an on-premises server. If this temporary rule is forgotten, it creates a permanent backdoor. Another frequent error is mismanaging Network Address Translation (NAT) rules, which can unintentionally expose internal, private IP addresses to the public internet or the entire cloud VPC.

The image above metaphorically represents this risk: a seemingly robust security mesh with a single, subtle gap. This one gap is all a threat actor needs. In a hybrid world, security cannot be an afterthought. Every routing change and every firewall rule modification must be scrutinized through the lens of security. This requires robust change management processes, regular security audits, and automated tools that can detect and alert on risky configurations before they can be exploited.

Route optimization: ensuring local traffic stays local

A fast and secure connection is in place, but a new, more subtle problem emerges: performance degradation for on-premises users. The cause is often found in suboptimal routing. Without careful management, your cloud provider can become the “default route” for all traffic, a phenomenon known as “tromboning” or “hairpinning.” This is where traffic between two local on-premises sites is unintentionally sent all the way to the cloud and back, introducing unnecessary latency and cost.

The key to preventing this lies in mastering BGP route optimization. BGP is the protocol that allows you to control how traffic flows between your on-premises network and the cloud. By using BGP attributes with intent, you can enforce policies that ensure local traffic stays local, and only cloud-destined traffic traverses the dedicated link. This is not a “set and forget” configuration; it’s a discipline of “Routing Intentionality,” where every path is deliberately chosen.

For example, using AS Path Prepending, you can artificially make the path through one connection seem longer (and therefore less desirable) to influence inbound traffic from the cloud. On your own network, you can use Local Preference attributes to dictate which outbound path your local routers should prefer. A higher local preference value makes a route more attractive. By assigning a higher preference to local routes, you ensure that inter-site traffic never leaves your corporate WAN. Another crucial technique is precise route summarization. Instead of advertising your entire internal network address space to the cloud, you should only advertise the specific subnets that need to be reachable, minimizing the risk of the cloud becoming an accidental transit network.

A thorough audit of your BGP configuration is essential to ensuring your network behaves as intended. This checklist provides a framework for that audit.

Action Plan: Auditing Your BGP Routing Policies

  1. Map Traffic Endpoints: List all network ingress and egress points between your on-premises sites and your cloud VPCs to define all possible traffic paths.
  2. Inventory Current Routes: Document all current BGP route announcements being advertised to and received from your cloud provider, including current Local Preference and AS-Path attributes.
  3. Assess Path Coherence: Using traceroute and other monitoring tools, actively test traffic paths between on-premises sites. Confront the actual data path with your intended design to identify any “tromboning.”
  4. Evaluate Path Predictability: Verify that traffic paths are symmetric (the path from A to B is the same as from B to A). Asymmetric routing can break stateful firewalls and cause elusive connectivity issues.
  5. Develop an Integration Plan: Based on your findings, create a phased plan to implement or adjust route summarization and BGP attributes to correct any deviations and enforce your desired routing policies.

By treating routing as a continuous optimization process rather than a one-time setup, you can ensure low latency for both cloud-bound and local traffic, creating a truly efficient hybrid network.

How to establish secure VPN tunnels between AWS and Azure?

The modern enterprise is rarely a single-cloud environment. It’s common for an organization to use AWS for some services and Azure for others. This multi-cloud reality introduces another layer of complexity: establishing secure, reliable, and low-latency connectivity *between* public clouds. While you could route this traffic back through your on-premises data center, this creates the exact “hairpinning” effect we seek to avoid. The solution is to build a direct, secure tunnel between your AWS and Azure environments.

Establishing an IPsec VPN tunnel between an AWS Virtual Private Gateway and an Azure Virtual Network Gateway is a well-documented but highly detailed process. The primary goal is to create a redundant and resilient connection that can automatically fail over if one tunnel goes down. This is not a simple, single-tunnel setup. True high availability requires a minimum of four redundant tunnels—a primary and secondary tunnel for each of the two gateway instances on both sides.

To manage this complexity and enable automatic failover, using BGP is essential. By establishing BGP sessions over the VPN tunnels, you allow the cloud providers to dynamically exchange routing information. If a tunnel fails, BGP will automatically withdraw the routes associated with it, and traffic will be redirected through a working tunnel. This provides a level of resilience that static routing cannot match. Furthermore, it is strongly recommended to use Route-Based VPNs instead of Policy-Based VPNs, as they provide far more flexibility when adding or changing networks in the future without having to rebuild the entire tunnel configuration.

Successfully architecting a multi-cloud VPN requires adherence to several best practices:

  • Use Route-Based VPNs: They are more flexible and scalable for inter-cloud connectivity, allowing you to add new networks without reconfiguring security policies.
  • Implement Full Redundancy: Set up four tunnels by configuring a pair for each of the two Virtual Network Gateways on both the AWS and Azure sides.
  • Leverage BGP for Failover: Implement BGP to manage automatic failover between tunnels, ensuring high availability across cloud providers.
  • Plan Your IP Schema: Establish a centralized, non-overlapping IP address schema from day one to prevent complex and error-prone NAT configurations later.
  • Consider Virtual Appliances: For throughput requirements beyond the standard 1.25 Gbps IPsec limit, consider using third-party Network Virtual Appliances (NVAs) from vendors like Cisco, Palo Alto Networks, or Fortinet, available in both cloud marketplaces.

Building this inter-cloud bridge transforms your separate cloud deployments into a more cohesive ecosystem, enabling applications in AWS to communicate securely and efficiently with services in Azure.

Why your 10GbE switch is the choke point of your network?

After optimizing your cloud connections, firewall rules, and routing policies, you might find latency issues persist. In these cases, it’s time to look inward. The bottleneck is often not in the cloud or the WAN link, but within your own data center. Specifically, your trusted 10GbE access or core switch, once the pinnacle of performance, can become the primary choke point in a modern hybrid architecture.

The problem is one of aggregate demand. A single large-scale data migration or a burst of activity from multiple applications can easily saturate a 10GbE link. For example, to truly saturate a high-speed dedicated cloud connection during a data transfer, a migration process might require 20 Gbps of total bandwidth—10 Gbps for reading data from your on-premises file systems and another 10 Gbps for uploading that data to the cloud simultaneously. A single 10GbE switch port simply cannot handle this combined load, causing packet queuing, increased latency, and a throttled migration.

This issue is compounded when multiple services contend for the same limited internal bandwidth. Your switch isn’t just handling the cloud connection; it’s also managing all the east-west traffic between your internal servers, storage arrays, and user workstations. The public internet connection, for all its variability, may offer more raw bandwidth than the congested internal link leading to it.

Loss and latency will be orders of magnitude higher across internet links than across internal networks.

– Network World, Cloud connectivity performance analysis

While this statement from Network World is generally true, it assumes the internal network is not the bottleneck. In a high-demand hybrid scenario, the congestion on your 10GbE switch can introduce latency that rivals or even exceeds that of the WAN link. The solution requires a holistic view of your network capacity. This may involve upgrading your core network to 25, 40, or even 100GbE, or implementing Link Aggregation (LAG) to bond multiple 10GbE ports together. It also necessitates careful traffic shaping and Quality of Service (QoS) policies to ensure that critical, latency-sensitive applications are prioritized over bulk data transfers.

Key Takeaways

  • Low latency is a function of end-to-end network architecture, not just the connection type you choose.
  • BGP route optimization is non-negotiable for controlling traffic paths, managing costs, and preventing performance degradation in a hybrid environment.
  • The traditional security perimeter has dissolved; a Zero Trust model is the only viable approach for protecting legacy and cloud assets together.

How to implement a Zero Trust strategy in a legacy network environment?

We’ve established that the hybrid cloud dissolves the traditional network perimeter, making tools like firewalls insufficient on their own. The modern solution to this challenge is a Zero Trust security architecture. The guiding principle is simple but profound: “never trust, always verify.” This means no user, device, or application is trusted by default, whether it’s inside or outside the network. Every access request must be explicitly verified.

Implementing Zero Trust in a greenfield, cloud-native environment is one thing; retrofitting it into a legacy on-premises network with aging hardware and flat network segments is a far greater challenge. However, it is not only possible but essential for securing a hybrid enterprise. The key is to shift the focus from network-centric controls (like firewalls at the perimeter) to identity- and data-centric controls.

The foundation of any Zero Trust strategy is modernizing Identity and Access Management (IAM). This means implementing strong Multi-Factor Authentication (MFA) for every user and service, eliminating shared credentials, and moving towards certificate-based authentication. This single step can have a massive impact; research shows that this approach can lead to an 88% elimination of the credential attack surface. Next, you must tackle the flat legacy network. Since re-architecting the physical network is often infeasible, the solution lies in software-based micro-segmentation. By deploying agents on endpoints and servers, you can create and enforce granular security policies that control traffic flow between applications, regardless of the underlying network topology. This effectively creates a secure micro-perimeter around each critical workload.

The implementation of a Zero Trust framework in a hybrid environment is a strategic journey, not a single project. It follows a logical progression:

  1. Modernize IAM: Make strong MFA the cornerstone of your strategy before touching the network. This is the biggest and fastest win.
  2. Deploy Micro-segmentation: Use software-based agents to enforce Zero Trust policies on your existing flat legacy network, bypassing its physical limitations.
  3. Create Containment Zones: For un-modernizable legacy systems like mainframes, create heavily monitored “containment zones” with strict ingress/egress proxying and inspection.
  4. Implement Data-Centric Security: Use data discovery, classification, and encryption tools to protect the data itself, ensuring it remains secure even if the network or endpoints are compromised.
  5. Enforce Verification: Apply the “never trust, always verify” principle consistently, requiring verification for every application, device, and user before granting access to any resource.

By adopting this strategy, you build a security model that is resilient, adaptable, and capable of protecting your organization’s assets across both your legacy data center and the dynamic world of the cloud.

To put these principles into practice, the next logical step is to conduct a full audit of your current network topology, routing policies, and security posture. Begin by mapping your data flows and identifying your most critical, latency-sensitive applications to build a network that truly serves the needs of your business.

Written by Erik Jensen, Principal Data Scientist & AI Systems Architect focused on data integrity and algorithms.