Edge, edge, everywhere an edge

Where is the edge? That question has popped up in a bunch of recent meetings as we talk about cloud computing and its “long-haired, hippy” sibling, edge computing. Even for us techie types, the edge can be difficult to pin down. Roughly 10 years ago, Cisco introduced the term “fog computing” – and given all the choices that exist to place compute these days, finding the edge is more like playing hide and seek in the mist than ever.

The edge will always vary for different use cases, so to describe the edge, let’s borrow a leaf from John Gage’s ’80s playbook. While at Sun Microsystems, he coined the phrase “the network is the computer.” John was describing access to compute that existed out on the network beyond workstations. In a slight adaptation, I would like to propose “the network is the edge.”

Maybe more specifically, every network transition your data zips through has the potential to be the edge. So, where should you put your edge? While data privacy and management of the infrastructure are always factors that could sway the outcome, the cost and relative importance of latency for your application are the key factors driving the decision.

Take cars, for instance. They have cameras or LiDar generating a high volume of data that a neural network consumes in order to make split-second driving decisions. The inference model needs to perform the forward propagation within milliseconds—very low latency. Many of these “embedded edge” or “product edge” devices either generate large volumes of data that can be difficult to transfer to another location quickly enough, or it is simply cheap enough to make these decisions on the local device. If you need at-source, low-latency decisioning or visualization, then the embedded edge will be the first edge to consider.

Devices and applications that still need low latency but can accommodate the time introduced by sending data upstream another hop or two may leverage an edge in the telco provider’s network. For example, maybe a smart city device leverages compute at a 5G MEC in order to determine an action. The 5G MEC provides low-latency connections and allows for some of the compute logic and decisioning to reside beyond the embedded edge in order to be scaled across a large number of devices.

This “mobile edge” is generally where the fog edge begins. A combination of edge choices that might be a private enterprise cloud or datacenter, telco cloud, internet data center, or exchange—really any variation of edges that precede the hyperscaler cloud edge that is most often thought of today.

Functions that evaluate and manipulate the flow of data might be deployed in a CDN edge that is a network hop or two from the local ISP. Inference models, games, and real-time financial applications might allow for compute to be deployed at a well-connected peering exchange or meet-me datacenter that has a unique blend of IP transit and peering. Training a large language model might require GPU deployment in a core data center or a hyperscaler. Depending on the latency requirements, we can evaluate the overall network path from devices, data, and users and find a network edge that provides the best balance between latency and cost for the application.

A diagram of a network Description automatically generated


In the evolving landscape of edges, businesses are increasingly turning to a range of solutions to bolster their strategic initiatives and achieve their objectives. By leveraging different edge environments, companies can ensure that they are optimizing their operations for efficiency, scalability, and security. This multi-edge approach allows for optimizing that balance between latency and cost, as well as the agility to adapt to market demands and technological advancements. As a result, choosing an edge provider that can accommodate a wide array of use cases has become complicated.

At Zenlayer, we understand the importance of latency and the intricacies of evaluating cost, privacy, security, and deployment flexibility. Our hyperconnected cloud is deployed in the fastest-growing and hardest-to-reach markets, and we strive to provide the lowest latency edge for your applications.

We have over 280 PoPs that are hyperconnected to local networks and also have access to private backbones for low-latency connectivity to other cities or regions. In zenConsole, you can evaluate latency between our PoPs and end-user networks to assess the best location for your bare metal or virtual machine deployment. You can also use our console to evaluate the latency between PoPs and seamlessly deploy a layer 2 point-to-point or fully meshed layer 3 connection between PoPs in other cities or regions.

In each of our PoPs, we architect a BGP blend with a wide range of local peers and transit providers in order to optimize traffic for our customers. We partner with AWS, Google, Microsoft, Oracle, Alibaba, and other public clouds to provide direct connections from our PoPs to these leading hyperscalers and flexible deployment models to connect your private data center to our network.

Here is an example of the Dhaka PoP and the latency to local eyeball networks in Bangladesh.


As technology evolves and the edge expands, choosing a partner that can accommodate a wide range of compute and low latency options is critical. Zenlayer’s hyperconnected cloud exemplifies our dedication to providing agile, cost-effective, and strategically located services to drive your business forward in the ever-evolving edge ecosystem.




Share article :


Register for our webinar: Solving Asia: How to Fast-Track your Digital Infrastructure Expansion