IPoC: A New Core Networking Protocol for 5G Networks

Greg White
Distinguished Technologist, Network Technologies

Oct 29, 2018

5G is the latest iteration of cellular network technology developed to meet the growing traffic demand for both smartphones and homes. With beamforming and frequency bands reaching millimeter waves, 5G promises many benefits:

  • Higher speeds
  • Lower latency
  • The ability to connect many more devices

However, current de jure standards and protocols, designed for earlier technologies, have the potential to dilute these promises. To address the limitations of current networks, CableLabs developed the IP over CCN (IPoC) protocol, a compelling new solution to meet the new, more robust requirements of 5G.

Why a New Solution?

The primary goal of the fourth generation (4G or LTE) technology was access to the Internet, so the technology utilized IP networking, the packet routing technology historically and currently used in the Internet.

IP networking has been around since the mid-1970s and has served us remarkably well, but it isn’t without flaws. The purpose of the Internet Protocol is to allow a computer at one fixed location in the network to exchange information with another computer at a fixed location in the network.  For mobile devices (that clearly aren’t at a fixed location) this has never been a great fit, and LTE technology had to develop complicated IP over IP tunneling mechanisms (the LTE Evolved Packet Core (EPC)) to enable mobility.

Furthermore, in the majority of cases, a mobile application wants to fetch specific data (say the text and images of a blog post) but doesn’t really care which computer it talks to in order to get it. As a result, to improve network efficiency and performance, network operators (both mobile and fixed) have implemented complex Content Distribution Networks in order to try to redirect the mobile application to the nearest server or cache that has the requested data.

In LTE-EPC, all of a user’s IP traffic is tunneled through a centralized choke point (or anchor) in the mobile operator’s core network, which eliminates the ability to serve data from a nearby cache. Also, as a mobile device moves in the network, the EPC needs to create new tunnels and tear down old ones in order to ensure that the user’s data reaches them.

These limitations are widely acknowledged by standards-setting groups. They are currently soliciting input to introduce new protocols that will pave the way for 5G to meet the demands of next-generation technologies, specifically:

  • Improve the efficiency and performance of the network mobility plane, compared to today’s LTE standards,
  • Support non-IP network protocols, of which Content Centric Networking is a leading candidate.

Benefits of Content Centric Networking

Content Centric Networking: A networking paradigm that emphasizes content by making it directly addressable and routable. Learn more here.

CCN offers several key advantages over IP networking:

  • It employs “stateful forwarding” which elegantly and efficiently supports information retrieval by mobile client devices without the need for tunneling or a location registration protocol
  • It addresses content directly rather than addressing end hosts, which means that it enables in-network caching, processing and intelligent packet forwarding, allowing it to excel in content retrieval optimization, allowing data to be easily retrieved from an on-path cache
  • It supports a client device using multiple network attachments (e.g., radio links) simultaneously, providing greater reliability and performance.
  • Its design meets the needs of large-data and IoT applications

For many new applications, CCN provides a much better fit for purpose than the Internet Protocol.

IP over CCN (IPoC): A New Way to Handle IP

In spite of the significant improvements Content Centric Networking offers over current IP networking, the reality is that all of today’s applications, both client and server, are built to use IP networking. We developed IPoC as the solution to this issue. IP over CCN (IPoC) protocol is a general-purpose tunneling protocol that enables delivery of IP traffic over a Content Centric Network (CCN) or a Named Data Network (NDN).

IPoC enables deployment of CCN as the core networking protocol for 5G, both for new, native CCN applications and as a mobility plane for existing IP applications, replacing the LTE-EPC. As a result, IPoC saves the IP investment and allows a full transition to the new CCN protocol.

With this approach:

  • Native CCN applications reap the benefits of tunnel-free anchorless networking, along with the latency and efficiency gains that come from in-network caching.
  • Existing IP-based applications can be supported with a mobility management solution that is simpler than the existing LTE-EPC. Gone are the special-purpose tunnel management functions that create and destroy tunnels as mobile devices move in the network.
  • The need for network slicing to accommodate both IP and CCN and the complications and overhead entailed in running two core networks in parallel are eliminated.

IPoC Performance in Mobile Networks

With the assistance of two PhD students from Colorado State University, we developed simulation models and conducted performance and efficiency testing of the protocol in comparison to LTE-EPC. In our simulation study, we implemented the IPoC protocol using the Named Data Networking (NDN) simulator ndnSim (which implements a CCN-like semantic) and used mobile communication as the driving example, comparing IPoC-over-NDN protocol performance against GTP-over-IP. We found that the protocol overhead and performance impact of IPoC is minimal, which makes it suitable for immediate deployment.  The report on this study includes links to the source code as well.

Want to Take a Closer Look?

IPoC can be best understood as a transition technology. Providing a shim layer and allowing CCN to act as a mobility plane for legacy IP applications, it accommodates the current protocol standards while opening the door for deployment of native CCN applications and the benefits they offer.

The 5G standardization project is seeking new mobility solutions for 5G, and we believe CCN and IPoC would be a great solution to address the needs. We have submitted a definition of the IPoC protocol as an Internet-Draft to the Internet Research Task Force (IRTF) Information Centric Networking Research Group. In addition, we have developed a proof-of-concept implementation of the IPoC protocol on Linux.

Interested in learning more? Subscribe to our blog and recieve updates on 5G by clicking below. 



Re-Inventing the Internet

Greg White
Distinguished Technologist, Network Technologies

Feb 8, 2016

The Internet Protocol – IP. It has become practically synonymous with networking, and over the last 40 years the adoption of IP has radically changed the world economy and the way that people interact globally, with CableLabs and the cable industry being a major force in that adoption. The past 15+ years in particular have been marked by the transitioning of the significant majority of industries, services, and human interactions into versions that are mediated by the world wide web, itself built on the foundation of the Internet Protocol. As this transition has taken place, the way that we use the network has radically evolved, from a simple point-to-point (or end-to-end) model with static endpoints, into one of massively distributed computing and storage accessed almost entirely by a huge number of mobile and nomadic client devices (with content and virtualized services being deployed on demand wherever it is most advantageous).

Alongside this usage evolution, network engineers have evolved in the way that they build IP networks, constructing ever more complex systems to try to manage the flow of communication, and have even taken on a change of the IP protocol (from IPv4 to IPv6) in an attempt to solve its most pressing shortcoming, the size of the address space. However, it is becoming more apparent that the address space issue is not the only problem with IP, and that we may be nearing the day when a more fundamental re-invention is needed.

A New Approach to Networking

Information Centric Networking is an emerging networking approach that aims to address many of the shortcomings inherent in the existing Internet Protocol. The field of Information Centric Networking was sparked by work at Stanford University in 1999-2002, which proposed a radical re-thinking of the network stack and a fundamental change in the way information flows in the global data network. Since that time, a number of researchers have developed proposals for the operational model, semantics, and syntax of an Information Centric Network, and today, one technology appears to be leading the pack and to be poised for potential success. This approach to Information Centric Networking, being developed in parallel by two projects: Content-Centric Networking (CCN) and Named Data Networking (NDN), appears to be gaining mindshare in the research community and in industry, and promises to significantly improve network scalability and performance, and reduce cost over a network built on the Internet Protocol. CCN/NDN provides native, and elegant, support for client mobility, multipath connectivity, multicast delivery and in-network caching, many of which are critical for current and future networks, and all of which require inefficient and/or complex managed overlays when implemented in IP. Further, CCN/NDN provides a much richer addressing framework than that which exists in IP (which could eliminate significant sources of routing complexity), and it provides a fundamentally more secure communication model, even for in-the-clear communications.

By moving away from a "host-centric" view of networking functions, where the core protocol (IP) is entirely devoted to delivering data from one specific host to another, to a "content-centric" view, where content objects are identified and routed solely by the use of globally unique names, the CCN/NDN approach provides a more elegantly scalable, faster, and more efficient network infrastructure for the majority of traffic on the Internet today. To get a sense of how big a mind shift this is, consider this: in CCN/NDN devices don’t have addresses at all. A device can retrieve content by requesting it by name, without needing to have a way of identifying a server where that content is stored, or even identifying itself.

At CableLabs, we are experimenting with this new protocol and are deep into investigating the applications that might drive its adoption. Because CCN/NDN is built with mobility, privacy and efficient content distribution in mind from the beginning, we see synergies with the direction that cable networks are going. We see this as enabling a convergence of fixed networking, mobile networking and content distribution – a “Fixed-Mobile-Content Convergence”.

As IP is baked into the networking field (equipment, software stacks, applications, services, engineering knowledge, business models, even national policies), it may seem daunting to consider the use of a non-IP protocol. However, while IP has been a phenomenally successful networking protocol for the last 40 years, as technology and time inevitably marches on it is reasonable to believe that we won’t be utilizing it forever.  And furthermore, while replacing IP with another protocol will certainly bring implementation challenges, it doesn’t follow that doing so necessarily means changing the way applications and users interact with the network. In other words, the web will still look and act like the web, but it will become more streamlined and efficient.

To be clear, CCN/NDN is not a mature protocol ready for immediate deployment. A number of R&D issues are still being worked in academic and industry research groups (including CableLabs), and implementations (both endpoint stacks and network gear) are still largely in the reference implementation or proof-of-concept stage. However, development work is active and confidence is high that the protocol will be ready for deployment in the next 3-5 years.

As part of this investigation, CableLabs has released a publication describing how an existing Content Distribution Network (CDN) evolves to one that leverages the benefits of CCN.

Greg White is a Distinguished Technologist at CableLabs. 


How DOCSIS 3.1 Reduces Latency with Active Queue Management

Greg White
Distinguished Technologist, Network Technologies

Jun 6, 2014

At CableLabs we sometimes think big picture, like unifying the worldwide PON standards, and sometimes we focus on little things that have big impacts.

In the world of cable modems a lot has been said about the ever-increasing demand for bandwidth, the rock-solid historical trend where broadband speeds go up by a factor of about 1.5 each year.  But alongside the seemingly unassailable truth described by that trend is the nagging question: "Who needs a gigabit per second home broadband connection, and what would they use it for?"  Are we approaching a saturation point, where more bits-per-second isn't what customers focus on?  Just like no one pays much attention to CPU clock rates when they buy a new laptop anymore, and megapixel count isn't the selling point it once was for digital cameras.

It's clear that aggregate bandwidth demand is increasing, with more and more video content (at higher resolutions and greater color depths) being delivered, on-demand, over home broadband connections.  DOCSIS 3.1 will deliver this bandwidth in spades.  But, in terms of what directly impacts the user experience, what is increasingly getting attention is latency.

Packet Latency and Application Performance

Latency is a measure of the time it takes for individual data packets to traverse a network, from a tablet to a web server, for example.  In some cases it is referred to as "round trip time," both because that's easier to measure, and because for a lot of applications, it’s what is most important.

For network engineers, latency has long been considered important for specific applications.  For example, it has been known since the earliest days of intercontinental telephone service that long round-trip times cause severe degradation to the conversational feeling of the connection.  It’s similar for interactive games on the Internet, particularly ones that require quick reaction times, like first-person-shooters. Latency (gamers call it "lag" or "ping time") has a direct effect on a player's chance of winning.


Based on data from M. Bredel and M. Fidler, "A Measurement Study regarding Quality of Service and its Impact on Multiplayer Online Games", 2010 9th Annual Workshop on Network and Systems Support for Games (NetGames), 16-17 Nov. 2010.


But for general Internet usage, web browsing, e-commerce, social media, etc. the view has been that these things aren't sensitive to latency, at least at the scales that modern IP networks can provide.  Does a user really care that much if their Web browsing traffic is delayed by a few hundred milliseconds?  That's just the blink of an eye – literally.  Surely this wouldn't be noticed, let alone be considered a problem.  It turns out that it is a problem, since a typical webpage can take 10 to 20 round-trip-times to load.  So, an eye-blink delay of 400ms in the network turns into an excruciating 8-second wait.  That is a killer for an e-commerce site, and for a user's perception of the speed of their broadband connection.

And, it turns out that more bandwidth doesn't help.  Upgrading to a "faster" connection (if by faster we mean more Mbps) will have almost zero impact on that sluggish experience.  For a typical webpage, increasing the connection speed beyond about 6 Mbps will have almost no effect on page load time.

Screen Shot 2014-06-06 at 8.53.40 AM

Graphic from SPDYEssentials, Roberto Peon & William Chan, Google Tech Talk, 12/8/11

If Latency is Indeed Important, What Can We Do About It? 

There is a component of latency that comes from the distance (in terms of network link miles) between a client and a server.  Information currently travels at near the speed of light on network links, and, while some network researchers have bemoaned that "Everybody talks about the speed of light, but nobody ever does anything about it" (Joe Touch, USC-ISI), it seems that once we harness quantum entanglement we can call that one solved.  Until that day comes, we have to rely on moving the server closer to the client by way of Web caches, content-delivery networks, and globally distributed data centers.  This can reduce the round-trip-time by tens or hundreds of milliseconds.

Another component of latency comes from the time that packets spend sitting in intermediate devices along the path from the client to the server and back.   Every intermediate device (such as a router, switch, cable modem, CMTS, mobile base station, DSLAM, etc.) serves an important function in getting the packet to its destination, and each one of them performs that function by examining the destination address on the packet, and then making a decision as to which direction to forward it.  This process takes some time, at each network hop, while the packet waits for its turn to be processed.    Reducing the number of network hops, and using efficient networking equipment can reduce the round-trip-time by tens of milliseconds.

But, there is a hidden, transient source of significant latency that has been largely ignored – packet buffering.  Network equipment suppliers have realized that they can provide better throughput and packet-loss performance in their equipment if they include the capability of absorbing bursts of incoming packets, and then playing them out on an outgoing link.  This capability can significantly improve the performance of large file transfers and other bulk transport applications, a figure of merit by which networking equipment is judged.    In fact, the view has generally been that more buffering capability is better, since it further reduces the chance of having to discard an incoming packet due to having no place to put it.

The problem with this logic is that file transfers use the Transmission Control Protocol (TCP) to control the flow of packets from the server to the client.  TCP is designed to try to maximize its usage of the available network capacity.  It does so by increasing its data rate up to the point that the network starts dropping packets.  If network equipment implements larger and larger buffers to avoid dropping packets, and TCP by its very design won't stop increasing its data rate until it sees packet loss, the result is that the network has large buffers that are being kept full whenever a TCP flow is moving files, and packets have to sit and wait in a queue to be processed.

Waiting in Line

Buffering in the network devices at the head of the bottleneck link between the server and the client (in many cases this is the broadband access network link) have the most impact on latency.  And, in fact, we see huge buffers supported by some of this equipment, with buffering delays of hundreds or thousands of milliseconds being reported in many cases, dwarfing the latency caused by other sources.  But it is a transient problem.  It only shows up when TCP bulk traffic is present.  If you test the round-trip time of the network connection in the absence of such traffic, you would never see it.  For example, if someone in the home is playing an online game or trying to make a Skype call, whenever another user in the home sends an email with an attachment, the game or Skype call will experience a glitch.

Is There a Solution to This Problem? 

Can we have good throughput, low packet loss and low latency?  At CableLabs, we have been researching a technology called Active Queue Management where the cable modem and CMTS will keep a close eye on how full their buffers are getting, and as soon as they detect that TCP is keeping the buffer full, they will drop just enough packets to send TCP the signal that it needs to slow down, so that more appropriate buffer levels can be maintained.

This technology shows such promise for radically improving the broadband user experience, that we've mandated it be included in DOCSIS 3.1 equipment (both cable modem and CMTS).  We've also amended the DOCSIS 3.0 specification to strongly recommend that vendors add it to existing equipment (via firmware upgrade) if possible.

On the cable modem, we took it a step further and require that devices implement an Active Queue Management algorithm that we've specially designed to optimize performance over DOCSIS links.  This algorithm is based on an approach developed by Cisco Systems, called "Proportional Integral Enhanced Active Queue Management" (or, more commonly known by its less tongue-tying acronym: PIE), and incorporates original ideas from CableLabs and several of our technology partners to improve its use in cable broadband networks.

By implementing Active Queue Management, cable networks will be able to reduce these transient buffering latencies by hundreds or thousands of milliseconds, which translates into massive reductions in page load times, and significant reductions in delays and glitches in interactive applications like video conferencing and online gaming, all while maintaining great throughput and low packet loss.

Active Queue Management: A little thing that can make a huge impact.

Greg White is a Principal Architect at CableLabs who researches ways to make the Internet faster for the end user.  He hates waiting 8 seconds for pages to load. 

For more information on Active Queue Management in DOCSIS, download the white paper.