ipj26.1
ipj26.1
ipj26.1
In This Issue Twenty-five years ago, we published the first issue of The Internet Protocol
Journal (IPJ). Since then, 87 issues for a total of 3,316 pages have been
produced. Today, IPJ has about 20,000 subscribers all around the world.
From the Editor....................... 1 In the early days of IPJ, most of our readers preferred the paper edition, but
over time preferences have shifted steadily to a situation where only some
1,200 print subscribers remain. The rest are downloading the PDF version.
ALTO...................................... 2 This shift in reading habits is likely related to the changes in technology that
have taken place in the last 25 years. Lower costs and higher-resolution
displays and printers, as well as improvements in Internet access technolo-
Wi-Fi Privacy.......................... 12
gies, have made the online “experience” a lot better than it was in 1998.
Twenty-Five Years Later......... 23 In this issue, we will first look at two areas of work taking place
in the Internet Engineering Task Force (IETF). The Application-Layer
Traffic Optimization (ALTO) protocol aims to make network state such
Supporters and Sponsors....... 49 as topology, link availability, routing policies, and path cost metrics in-
formation available to applications in a standardized manner. The next
article concerns the thorny topic of tracking of users and their devices
Thank You!........................... 50 on the Internet. This area is complex, with many potential solutions,
including the use of randomized Media Access Control (MAC) addresses
as described by members of the MAC Address Device Identification for
Network and Application Services (MADINAS) Working Group in the
IETF.
Our final article is a look back at the last 25 years of Internet technology
development. As we did with our 10th anniversary issue in 2008, we asked
Geoff Huston to provide an overview of the many changes that have
taken place in this period. At the end of his article, you will find a list of
previously published articles from IPJ on numerous aspects of Internet
technologies. All back issues are, of course, available from our website.
Let me take this opportunity to thank all the people who make IPJ pos-
sible. We are grateful to all our sponsors and donors, without whose
generous support this publication would not exist. Our authors deserve a
round of applause for carefully explaining both established and emerging
technologies. They are assisted by an equally insightful set of reviewers
and advisors who provide feedback and suggestions on every aspect of
our publications process. The process itself relies heavily on two individu-
You can download IPJ als: Bonnie Hupton, our copy editor, and Diane Andrada, our designer.
back issues and find Thanks go also to our printers and mailing and shipping providers. Last,
subscription information at: but not least, our readers provide encouragement, suggestions, and feed-
www.protocoljournal.org back. This journal would not be what it is without them.
I
n today’s Internet, network-related information (for example,
topology, link availability, routing policies, and path cost metrics)
are usually hidden from the application layer. As a result, end-
points make network-unaware decisions that may lead to suboptimal
service placement and selection decisions, sometimes resulting in poor
user experience and unnecessary inter-Internet Service Provider (ISP)
traffic. Previous approaches to this problem space have considered
snooping on the lower layers to determine the state and capabilities
of the network, but such techniques require applications to be aware
of lower-layer components (for example, routing protocols) and, fur-
thermore, if left unspecified, can potentially overload key network
resources.
Figure 1, on the next page, depicts the main ALTO abstract objects
involved in a network. In this figure, application endpoints are rep-
resented as ANEs clustered in two groups with provider-defined
identifiers “PID1” and “PID2.” An endpoint can select to communi-
cate with another endpoint based on the network properties that the
ALTO server exposes. For instance, in a Content Delivery Network
(CDN), ANEs correspond to client and server hosts, and a specific
content (for example, a movie) is in general replicated in more than
one server instance. A client host can decide to retrieve the content by
selecting the server instance that provides the higher communication
bandwidth according to the exposed ALTO information. Each of the
abstract objects that are illustrated in Figure 1 is further elaborated in
the following sections.
ALTO Maps
An ALTO server organizes the network information using the concept
of maps. Maps can be constructed from physical information, logical
information, or a combination thereof. ALTO supports four types of
maps, as shown in Figure 1:
• The Network Map lists all the endpoint groups that the ALTO server
tracks. This map includes PIDs that uniquely identify each group.
• The Entity Property Map describes the properties of each ANE in
the network, including the geolocation or the connectivity type (for
example, fiber or wireless) of an ANE.
• The Cost Map provides the cost information (for example, hop
count, latency, or bandwidth) between each pair of PIDs enclosed in
the network map, where a PID identifies a group of endpoints.
• The Endpoint Cost Map provides finer-grained cost information
between specific endpoints.
ALTO Client
ALTOALTO
Server
Server
ALTO Client
ANE: Endpoint
ANE: Router ANE: Router ANE: Router
ANE: Endpoint
ANE: Endpoint
ANE: Endpoint
ALTO Extensions
In addition to its base abstractions and maps, the ALTO protocol
supports the following extensions to enable a richer network-aware
application experience:
• The Information Resource Directory (IRD) lists the services an
ALTO server provides and the locations from where you can access
these services.
• The Cost Calendar provides a set of cost values as a function of
time, allowing applications to know not only where to connect to,
but also when.
• Incremental Updates using Server-Sent Events (SSEs) allow an
ALTO server to expose cost values as delta updates, reducing the
amount of server-client data exchanged.
• The CDNI Advertisement exposes a CDNI Footprint and Capability
Advertisement Interface (FCI)[26].
• The Path Vector extension exposes the set of ANEs along the path
between two endpoints and the performance properties of these
ANEs.
• The Extended Performance Cost Metrics enrich ALTO with ad-
vanced metrics such as network one-way delay, one-way delay
variation, one-way packet-loss rate, hop count, and bandwidth.
• The Entity Properties generalize the concept of ALTO endpoint
properties by presenting them as entity property maps.
CDNI FCI
RFC 9241
Server
ALTO Incremental
Discovery ALTO OAM ALTO Transport
Deployment Update
RFC 7286
RFC 7971 RFC 8895
RFC 8686
History of ALTO
The ALTO Working Group was established in 2008 with an initial
charter to develop a request/response protocol that would allow hosts
to extract enough state information from a network to make optimized
server selection decisions. The working group’s first charter focused on
the optimization of Peer-to-Peer (P2P) applications, with the first four
RFCs introducing the problem statement[11] and requirements[4], the
base protocol[1], and support for server discovery[5].
The current ALTO Working Group charter was approved in 2021 with
the goal to focus on three operational areas: (1) support for modern
transport protocols such as HTTP/2 and HTTP/3[22]; (2) development
of OAM mechanisms[23], and (3) collection of deployment experi-
ences[24]. These three areas constitute the current highest priorities of
the ALTO Working Group.
Four additional RFCs that had originated from the second charter have
also been published since then: (1) support of property maps for gener-
alized entities[10], (2) a new Footprint and Capabilities Advertisement
Interface (FCI) protocol for CDNI[12], (3) a new Internet Assigned
Numbers Authority (IANA) registry for tracking cost modes supported
by ALTO[3], and (4) extensions to the cost map and ALTO property
map services to allow the application to identify optimized paths[14].
Figure 3: Mapping of RFC 7285 Entities onto the OpenALTO Software Architecture
AI
RFC 7285
+-------------------------------------------------------------------+ Meta
| Network Region | XR V2X IoT CDN Science (...)
| | Verse
| +-----------+ |
| | Routing | | OpenALTO Client
| +--------------+ | Protocols | |
| | Provisioning | +-----------+ |
| | Policy | | |
| +--------------+\ | |
| \ | | Northbound Interface
OpenALTO Server Stack
| \ | |
| +-----------+ \+---------+ +--------+ |
| |Dynamic | | ALTO | ALTO Protocol | ALTO | | ALTO Services Layer
| |Network |.......| Server | ==================== | Client | |
| |Information| +---------+ +--------+ |
| +-----------+ / / | Network Cost Property
| / ALTO SD Query/Response / |
Map Map Map (...)
| / / |
| +----------+ +----------------+ |
| | External | | ALTO Service | |
| | Interface| | Discovery (SD) | |
| +----------+ +----------------+ |
| | | Southbound Interface
+-------------------------------------------------------------------+
|
+------------------+
| Third Parties |
| | SDN Controller
| Content Providers|
+------------------+ Mobile Edge Mini
RAN WAN DCN
Core Core Net (...)
Future Perspectives
The ALTO protocol initially started with the goal of supporting the
optimization of P2P applications in 2008, then evolved to incorpo-
rate extensions for the support of CDNs in 2014, and today it is
well-positioned to support the requirements of new advanced edge
computing applications such as augmented reality, vehicle networks,
and the metaverse, among others. Because this new class of applica-
tions requires stringent Quality of Experience (QoE) performance,
the ALTO protocol becomes a key component to enable collaborative
application/network schemes.
The Internet Protocol Journal is published under the “CC BY-NC-ND” Creative Commons
Licence. Quotation with attribution encouraged.
This publication is distributed on an “as-is” basis, without warranty of any kind either
express or implied, including but not limited to the implied warranties of merchantability,
fitness for a particular purpose, or non-infringement. This publication could contain technical
inaccuracies or typographical errors. Later issues may modify or update information provided
in this issue. Neither the publisher nor any contributor shall have any liability to any person
for any loss or damage caused directly or indirectly by the information contained herein.
W
i-Fi technology has revolutionized communication and
become the preferred technology and sometimes the only
networking technology used by devices such as smart-
phones, tablets, and Internet-of Thing (IoT) devices.
This privacy concern affects all layers of the protocol stack, from the
lower layers involved in the actual access to the network (for example,
you can use the L2 and L3 addresses to obtain a user’s location) to
higher-layer protocol identifiers and user applications[1]. In particular,
IEEE 802 MAC addresses have historically been an easy target for
tracking users[2]. Attackers who are equipped with surveillance equip-
ment can “monitor” Wi-Fi packets and track the activity of Wi-Fi
devices. After the association between a device and its user is made,
identifying the device and its activity is sufficient to deduce informa-
tion about what the user is doing, without the user’s consent.
b0 (U/L bit):
0: unicast
1: multicast
b1:
0: globally unique (OUI enforced)
1: locally administrated
b7 b6 b5 b4 b3 b2 b1 b0
However, such address changes may affect the user experience and
the efficiency of legitimate network operations. For a long time, net-
work designers and implementers relied on the assumption that a given
machine in a network implementing IEEE 802 technologies would be
represented by a unique network MAC address that would not change
over time, despite the existence of tools to flush out the MAC address
to bypass some network policies. When this assumption is broken, ele-
ments of network communication may also break.
For example, sessions established between the end device and net-
work services may be lost and packets in translation may suddenly
be without a clear source or destination. If multiple clients implement
fast-paced RCM rotations, network services may be over-solicited by
a small number of stations that appear as many clients.
At the same time, some network services rely on the client station pro-
viding an identifier, which can be the MAC address or another value.
If the client implements MAC rotation but continues sending the same
static identifier, then the association between a stable identifier and
the station continues despite the RCM scheme. There may be environ-
ments where such continued association is desirable, but others where
the user privacy has more value than any continuity of network service
state.
When a device changes its MAC address, other devices on the same
LAN may fail to recognize that the same machine is attempting to
communicate with them. Additionally, multiple layers implemented at
upper layers have been designed with the assumption that each node
on the LAN, using these services, would have a MAC address that
would stay the same over time (a persistent MAC address).
The IEEE 802.1 Working Group has also published a specification that
defines a local MAC address space structure, known as the Structured
Local Address Plan (SLAP). This structure designates a range of
local MAC addresses for protocols using a CID assigned by the IEEE
Registration Authority. Another range of local MAC addresses is desig-
nated for assignment by administrators. The specification recommends
a range of local MAC addresses for use by IEEE 802 protocols[11].
Work within the IEEE 802.1 Security Task Group on privacy recom-
mendations for all IEEE 802 network technologies has also looked into
general recommendations on identifiers, reaching the conclusion that
temporary and transient identifiers are preferable in network technol-
ogy designs if there are no compelling reasons of service quality for a
newly introduced identifier to be permanent. This work has been spec-
ified in the recently published IEEE P802E: “Recommended Practice
for Privacy Considerations for IEEE 802 Technologies”[12]. The IEEE
P802E specification will form part of the basis for the review of user
privacy solutions applicable to IEEE Std 802.11 (aka Wi-Fi) devices as
part of the RCM[7] efforts.
Some solutions might mitigate this privacy threat, such as the use of
temporary addresses[15] and opaque IIDs[16,17]. Additionally, [18] pro-
poses an extension to DHCPv6 that allows a scalable approach to
link-layer address assignments where preassigned link-layer address
assignments (such as by a manufacturer) are not possible or unnec-
essary. [19] proposes extensions to DHCPv6 protocols to enable a
DHCPv6 client or relay to indicate a preferred SLAP quadrant to the
server, so that the server may allocate MAC addresses in the quadrant
requested by the client or relay.
Not only can you use MAC and IP addresses for tracking purposes,
but some DHCP options allow you to also carry unique identifiers.
These identifiers can enable device tracking even if the device adminis-
trator takes care of randomizing other potential identifications such as
link-layer addresses or IPv6 addresses. [20] introduces anonymity pro-
files, designed for clients that wish to remain anonymous to the visited
network. The profiles provide guidelines on the composition of DHCP
or DHCPv6 messages, designed to minimize disclosure of identifying
information.
Existing Solutions
One possible solution is to use 802.1X with Wi-Fi Protected Access
2/3 (WPA2/WPA3). At the time of association to a Wi-Fi access point,
802.1X authentication coupled with WPA2 or WPA3 encryption
schemes allows for the mutual identification of the client device or of
the user of the device and an authentication authority. The authen-
tication exchange is protected from eavesdropping. In this scenario,
you can obfuscate the identity of the user or the device from exter-
nal observers. However, the authentication authority is in most cases
under the control of the same entity as the network access provider,
thus making the identity of the user or device visible to the network
owner. This scheme is therefore well-adapted to enterprise environ-
ments, where a level of trust is established between the user and the
enterprise network operator.
It is also worth mentioning that most evolved client device OSes already
offer RCM schemes, enabled by default (or easy to enable) on client
devices. With these schemes, the device can change its MAC address,
when not associated, after using a given MAC address for a semi-
random duration window. These schemes also allow for the device to
manifest a different MAC address in different Service Set Identifiers
(SSIDs). Different OSes follow slightly different approaches, which
are also evolving with the new releases. Such a randomization scheme
enables the device to limit the duration of exposure of a single MAC
address to observers. In the IEEE 802.11-2020 specification, MAC
address rotation is not allowed during a given association session, and
thus rotation of MAC address can occur only through disconnection
and reconnection.
References
[1] Carlos J. Bernardos, Juan Carlos Zúñiga, and Piers O’Hanlon,
“Wi-Fi Internet Connectivity and Privacy: Hiding your tracks
on the wireless Internet,” IEEE Conference on Standards for
Communications and Networking (CSCN), October 2015.
[2] James Vincent, “London’s bins are tracking your smartphone,”
The Independent, August 2013,
https://www.independent.co.uk/life-style/gadgets-
and-tech/news/updated-london-s-bins-are-tracking-your-
smartphone-8754924.html
[3] Piers O’Hanlon, Joss Wright, and Ian Brown, “Privacy at the link
layer,” Contribution at W3C/IAB workshop on Strengthening
the Internet Against Pervasive Monitoring (STRINT), February
2014. https://www.w3.org/2014/strint/papers/35.pdf
[4] Marco Gruteser and Dirk Grunwald, “Enhancing Location
Privacy in Wireless LAN Through Disposable Interface Identifiers:
A Quantitative Analysis,” Mobile Networks and Applications,
Volume 10, No. 3, pp. 315-325, 2005.
[5] Jeremy Martin, Travis Mayberry, Colin Donahue, Lucas Foppe,
Lamont Brown, Chadwick Riggins, Eric C. Rye, and Dane
Brown, “A Study of MAC Address Randomization in Mobile
Devices and When It Fails,” arXiv:1703.02874v2 [cs.CR], 2017.
[6] Institute of Electrical and Electronics Engineers (IEEE), “IEEE
802.11aq-2018 - IEEE Standard for Information Technology—
Telecommunications and Information Exchange Between
Systems Local and Metropolitan Area Networks—Specific
Requirements Part 11: Wireless LAN Medium Access Control
(MAC) and Physical Layer (PHY) Specifications Amendment 5:
Preassociation Discovery,” IEEE 802.11, 2018.
[7] Institute of Electrical and Electronics Engineers (IEEE),“IEEE
802.11 Randomized and Changing MAC Addresses Study Group
CSD on User Experience Mechanisms,” doc.:IEEE 802.11-
20/1346r1, 2020.
[8] Institute of Electrical and Electronics Engineers (IEEE), “IEEE
802.11 Randomized and Changing MAC Addresses Topic
Interest Group Report,” doc.:IEEE 802.11-19/1442r9, 2019.
[9] Institute of Electrical and Electronics Engineers (IEEE), “IEEE
802.11 Randomized and Changing MAC Addresses Study Group
PAR on User Experience Mechanisms,” doc.:IEEE 802.11-
20/742r5, 2020.
_______________________
T
he Internet is not quite as young and spritely as you might have
thought. Apple’s iPhone, released in 2007, is now 16 years old,
and YouTube is an ageing teenager at 18 after its initial release
in 2005, and these two examples are relatively recent additions to the
Internet. The first web browser, Mosaic, was released some 30 years
ago in 1993. Going back further, the Internet emerged from its early
Advanced Research Projects Agency (ARPA) roots in the form of the
National Science Foundation Network (NSFNET) in 1986. At the
start of 1983, the ARPA Network (ARPANET) had a flag day and
switched over to use the Transmission Control Protocol (TCP). Going
back further, in 1974 Vint Cerf and Bob Kahn published the first
academic paper describing the protocol and the underlying archi-
tectural framework of a packet-switched network that became the
Internet. This achievement was built upon earlier foundations, where
numerous efforts in the late 1960s showed the viability of a packet-
switched approach to computer networking. These packet-switched
networking efforts included a program led by Donald Davies at the
UK National Physics Laboratory, an effort in the US in the form of an
ARPA project led by Larry Roberts, and Louis Pouzin’s work in France
with the CYCLADES network. This work, in turn, has some of its
antecedents in work by Paul Baran at the RAND Corporation on
distributed communications and packet-switched networks, published
between 1960 and 1964. The Internet has managed to accumulate a
relatively lengthy pedigree.
And it has been a wild ride. The Internet has undergone numerous
cycles of economic boom and bust, each of which is right up there with
history’s finest episodes of exuberant irrational mania. It has managed
to trigger a comprehensive restructuring of the entire global commu-
nications enterprise and generated a set of changes that have already
altered the way in which we now work and play. That’s quite a set of
achievements in just 25 years!
By the early 2000s, the Internet had finally made it into the big time.
The job was apparently done, and the Internet had prevailed. But
then came a new revolution, this time in mobility services, where after
numerous clumsy initial efforts by others, the iPhone entered the mar-
ket with a seamless blend of sleek design and astounding capability.
The mobile carriage sector struggled to keep up with the new levels of
rapacious demand for Internet-based mobile data. The Internet then
took on the television networks, replacing the incumbent broadcast
and cable systems with streaming video. But the story is not over by
any means. Communications continues to drive our world, and the
Internet continues to evolve and change.
The evolutionary path of any technology can often take strange and
unanticipated turns and twists. At some points simplicity and mini-
malism can be replaced by complexity and ornamentation, while at
other times a dramatic cut-through exposes the core concepts of the
technology and removes layers of superfluous additions. The techni-
cal evolution of the Internet appears to be no exception, and this story
contains these same forms of unanticipated turns and twists.
Transmission
It seems like a totally alien concept these days, but the Internet Service
Provider (ISP) business of 1998 was still based around the technol-
ogy of dial-up modems. The state-of-the-art of modem speed had been
continually refined, from 9600 bps to 14.4 kbps, to 28 kbps, to finally
56 kbps, squeezing every last bit out of the phase amplitude space
contained in an analogue voice circuit. Analogue modems were capri-
cious, constantly being superseded by the next technical refinement,
unreliable, difficult for customers to use, and on top of that, they were
slow! Almost everything else on the Internet had to be tailored to
download reasonably quickly over a modem connection. Web pages
were carefully composed with compressed images to ensure a rapid
download, and plain text was the dominant medium as a consequence.
It could only get better.
The next generation of mobile services, 2G, was used in the 1990s.
It was still predominately a voice service, and while it could theoreti-
cally support data access at speeds of 100 kbps, this data-transfer rate
was largely aspirational, and the mobile network was predominantly
used by the Short Message Service (SMS) as an adjunct to voice. The
intersection of the Internet and mobile services occurred with the
introduction of 3G mobile services. The 3G architecture could push
IP connectivity directly out to the handset, supporting data-transfer
speeds of 1–2 Mbps. This network capability, coupled with a new
generation of handsets, first with the BlackBerry in 2002 and then
the iPhone in 2007, transformed the mobile service into a mass-
market consumer service. The high margins available from the service
captured the attention of the traditional voice industry, and the
result was a large-scale opening up of radio spectrum to create an
Internet access environment that quickly rivalled the wire-line access
market in size, but totally outpaced it in terms of revenue margins. This
massive escalation of demand created further pressures on the capac-
ity of the mobile system, and in 2009 the mobile sector introduced
4G services, opening up more spectrum bands, and also adding
Multiple-Input Multiple Output (MIMO) to the mobile device to
achieve greater deliverable capacity to each connected device. Over
time these services were to deliver peak download speeds of 50 to
100 Mbps. The industry was also selling hundreds of millions of
mobile devices per year. 4G dispensed with circuit-switched services,
and it exclusively used packet switching. In 2018 5G was introduced.
5G can be supported over more spectrum bands, including a high-band
millimetre spectrum at 24–47Ghz. These higher carrier frequencies
permit multi-gigabit data services, but they come at a cost of higher
density of base-station towers to compensate for the lower propaga-
tion distances.
Wi-Fi and Bluetooth A second radio technology that has also transformed the Internet
emerged in 1998, and it could be argued that it has become so funda-
mental that it has weaved itself so naturally into our environment that
it all but disappeared. The combination of low-power radio systems
and unlicensed radio spectrum allocation, or Wi-Fi, and subsequently
Bluetooth, has been transformational. The combination of efficient
battery technology, computer chips that operate with low power
consumption, and the unwiring of the last few meters in the home
and office completely changed our collective of technology, and it is
only because of our desire to use products that are portable, unobtru-
sively wearable, and powerful enough to be useful that the component
technologies such as batteries and processors have been pushed in this
direction over this period. While large bands of radio spectrum space
have been allocated to cellular mobile service operators, the intensity
of use and the utility of use of radio spectrum peaks in the unlicensed
spectrum space used by Wi-Fi and Bluetooth. It could be argued that
the economic value of these unlicensed spectrum bands exceeds the
exclusively licensed cellular radio systems by orders of magnitude.
It could also be argued that the untethering of the last meter of the
Internet transformed the Internet, and digital technologies in general,
from a specialist pursuit into the consumer product space.
Satellite Mobile data services, Wi-Fi, and Bluetooth really revolutionised the
Internet, taking it from a “destination you visit” to an “always-on
utility in your pocket.” The Internet was now a set of applications
on a device that went with you everywhere. Always available, always
connected, no matter where you might be or what you might be
doing. But that was not exactly the full truth. Head out into remote
country far enough, or head onto the world’s oceans, and your con-
nection options quickly disappeared, leaving only somewhat expensive
satellite-based services.
These satellite services have been in operation since the early 1960s,
but the high launch costs, limited capacity, and competing interests of
terrestrial providers have meant that these services were often operated
at the margins of viability. The best example is Motorola’s Iridium
project of the late 1990s, where even before the entire service constel-
lation of satellites was launched, the $5B Iridium project was declared
bankrupt. Starlink, a recent entrant in the satellite service area, is
using a constellation of some 4,000 low-earth-orbiting spacecraft
and appears so far to have been able to break through this financial
barrier. Using reusable launch vehicles, smaller (and lighter) satellites,
transponder arrays on board, and a new generation of digital-signal-
processing capabilities, Starlink is in a position to offer retail access
services of 100 Mbps or more to individual customers. The low
altitude of the spacecraft means that the Starlink service competes
directly with terrestrial access services in terms of performance. The
introduction of inter-spacecraft laser links means that the system
can provide a service in any location, and the limiting factor, as with
the Iridium effort decades ago, is obtaining the necessary clearances
and licenses to have customers located in the respective national
geographies. Starlink is certainly revolutionary in terms of capacity,
coverage, and cost. The questions are whether it is sufficiently revolu-
tionary and whether it can scale up to provide a high-capacity service
to hundreds of millions of users. At this point in time these questions
are not easy to answer, but the limitations inherent in Low Earth Orbit
(LEO)-based services point to a potential advantage in terrestrial-based
access networks. Nevertheless, Starlink is redefining the Internet access
market space in many countries, and setting price/performance bench-
marks that their terrestrial competitors now have to match.
While it was not going to stop here, squeezing even more capacity
from the network was now proving to be a challenge; 622-Mbps IP
circuits were being deployed, although many of them were constructed
using 155-Mbps Asynchronous Transfer Mode (ATM) circuits using
router-based load balancing to share the IP load over four of these
circuits in parallel. Gigabit circuits were just around the corner, and
the initial exercises of running IP over 2.5-Gbps Synchronous Digital
Hierarchy (SDH) circuits were being undertaken in 1998.
In some ways 1998 was a pivotal year for IP transmission. Until this
time, IP was still just one more data application that was positioned
as just another customer of the telco’s switched-circuit infrastructure.
This telco infrastructure was designed and constructed primarily to
support telephony. From the analogue voice circuits to the 64K digital
circuit through to the higher-speed trunk bearers, IP had been run-
ning on top of the voice network infrastructure. Communications
infrastructure connected population centres where there was call vol-
ume. The Internet had different demands. Internet traffic patterns did
not mirror voice traffic, and IP performance is sensitive to every addi-
tional millisecond of delay. Constraining the Internet to the role of an
overlay placed on top of a voice network was showing signs of stress,
and by 1998 things were changing. The Internet had started to make
ever larger demands on transmission capacity, and the driver for
further growth in the network infrastructure was now not voice, but
data. It made little sense to provision an ever-larger voice-based switch-
ing infrastructure just to repackage it as IP infrastructure, and by 1998
the industry was starting to consider just what an all-IP high-speed
network would look like, building an IP network all the way from the
photon in a fibre-optic cable all the way through to the design of the
Internet application.
Fibre Optics At the same time, the fibre-optic systems were changing with the intro-
duction of Wave Division Multiplexing (WDM). Older fibre equipment
with electro-optical repeaters and Plesiochronous Digital Hierarchy
(PDH) multiplexors allowed a single fibre pair to carry around
560 Mbps of data. WDM allowed a fibre pair to carry multiple chan-
nels of data using different wavelengths, with each channel supporting
a data rate of up to 10 Gbps. Channel capacity in a fibre strand was
between 40 and 160 channels using Dense WDM (DWDM). Combined
with the use of all-optical amplifiers, the most remarkable part of this
entire evolution in fibre systems is that a cable system capable of an
aggregate capacity of a terabit can be constructed today for much the
same cost as a 560-Mbps cable system of the mid-1990s. That’s a cost-
efficiency improvement of a factor of one million in a decade. The drive
to deploy these high-capacity DWDM fibre systems was never based
on expansion of telephony. The explosive growth of the industry was
all about supporting the demand for IP. So, it came as no surprise that
at the same time as the demand for IP transmission was increasing
there was a shift in the transmission model where instead of plugging
routers into telco switching gear and using virtual point-to-point cir-
cuits for IP, we started to plug routers into wavelengths of the DWDM
equipment and operate all-IP networks in the core of the Internet.
Network Management In network operations, we are seeing some stirrings of change, but it
appears to be a rather conservative area, and adoption of new network
management tools and practices takes time.
At that point, all the plans for an orderly transition were discarded,
and many network administrators scrambled to obtain IPv4 addresses,
further depleting the IPv4 pools. The central pool of IPv4 addresses,
operated by the Internet Assigned Numbers Authority (IANA), was
exhausted in February 2011. The Asia Pacific Network Information
Centre (APNIC) depleted its IPv4 pool in April of that year, the Réseaux
IP Européens Network Coordination Centre (RIPE NCC) 18 months
later, the Latin America and Caribbean Network Information Centre
(LACNIC) in 2014, and the American Registry for Internet Numbers
(ARIN) in 2015. We had expected that this situation would motivate
network operators to hasten their plans for IPv6 deployment, yet,
perversely, that did not happen. Less than 1% of the Internet user
base was using IPv6 in 2011. Five years later, as each of the Regional
Internet Registries (RIRs) ran down their remaining pools of IPv4
addresses, this Internet-wide IPv6 user count had increased to just 5%.
In 2023, the process is still underway, and some 35% of the Internet
user base has IPv6 capability. I’m not sure anyone is willing to predict
how long this anomalous situation of running the IPv4 Internet “on an
empty tank” will persist.
NATs How has the Internet managed to continue to operate, and even grow,
without a supply of new IPv4 addresses? In a word, the answer is
“NATs.” While the Network Address Translator (NAT) concept
received little fanfare when it was first published, it has enjoyed
massive deployment over the past 25 years, and today NATs are ubiq-
uitous. The application architecture of the Internet has changed, and
we are now operating a client/server framework. Servers have perma-
nent IP addresses, while clients “borrow” a public IPv4 address to
complete a transaction and return it back to a common pool when
they are done. Time-sharing IP addresses, and also using the 16-bit
source port field in TCP and the User Datagram Protocol (UDP), has
managed to extend the IPv4 address space by some 20 bits, making the
IPv4+NAT address space up to a million times larger than the origi-
nal 32-bit IPv4 address space. In practice, the story is a little more
complicated than that, and some very large service providers have
reached logistical limits in using NATs to compensate for the exhaus-
tion of IPv4 addresses. This situation has motivated these providers to
transition to a dual-stack mode of operation, and they are relying on a
dual-stack host behaviour that prefers to use IPv6 when possible, thus
relieving the pressure on the IPv4 NAT functions
The leisurely pace of the IPv6 transition is partly due to this altered
role of addresses, as we no longer require every connected device to
have a persistently assigned globally unique IP address.
IPv6 and NATs are not the only areas of activity in the Internet layer in
the past 25 years. We have tried to change many parts of the Internet
layer, but interestingly, few, if any, of the proposed changes have
managed to gain any significant traction out there in the network.
The functions performed at the Internet layer of the protocol stack are
no different from those of 25 years ago. IP Mobility, Multicast, and
IP Security (IPSec) are largely Internet layer technologies that have
failed to gain significant levels of traction in the marketplace of the
public Internet.
QoS Quality of Service (QoS) was a “hot topic” in 1998, and it involved
the search for a reasonable way for some packets to take some form of
expedited path across the network, while other packets took an undif-
ferentiated path. We experimented with various forms of signalling,
packet classifiers, queue-management algorithms, and interpretations
of the Type of Service bits in the IPv4 packet header, and we explored
the QoS architectures of Integrated and Differentiated Services in
great detail. However, QoS never managed to get established in main-
stream Internet service environments. In this case, the Internet took a
simpler direction, and in response to not enough network capacity we
just augmented the network to meet demand.
MPLS The switch from circuit switching to packet switching has never man-
aged to achieve universal acceptance. We have experimented with
putting circuits back into the IP datagram architecture in various
ways, most notably with the Multi-Protocol Label Switching (MPLS)
technology. This technology used the label-swapping approach that
was previously used in X.25, Frame Relay and ATM virtual circuit-
switching systems, and it created a collection of virtual paths from
each network ingress to each network egress across the IP network.
The original idea was that in the interior of the network you no
longer needed to load up a complete routing table into each switch-
ing element, and instead of performing destination-address lookup
you could perform a much smaller, and hopefully faster, label lookup.
This performance differentiator did not eventuate and switching pack-
ets using the 32-bit destination address in a fully populated forwarding
table continued to present much the same level of cost efficiency at
the hardware level as virtual circuit label switching.
The largest of these unresolved issues lies in the trust we place in the
inter-domain routing system of the Internet. There is no overall orches-
tration of the routing system. Each network advertises reachability
information to its adjacent networks and selects what it regards as
the “best” reachability information from the set received from these
same network peers. This mutual trust that each network places in
all other networks can, and has, been abused in various ways. The
effort to allow each routing entity to distinguish between what is a
“correct” item of routing information and what is a “false” route
has a rich history of initiatives that have faltered for one reason or
another. The most recent effort in this space is built upon the founda-
tions of the number system, and it uses the association of a public/
private key pair with the current holders of addresses and autonomous
system numbers, allowing these holders to issue signed authorities about
the use of these number resources in the context of routing, and by
coupling these authorities with the information being propagated in
the routing system, the intention being that unauthorized use cases will
be detected.
RPKI This effort, the Resource Public Key Infrastructure (RPKI), has
achieved some level of acceptance in the networking space, and in
2023 around one-third of all route objects have associated RPKI
credentials. The work is still “in progress” because the more challeng-
ing aspect of this work is to associate verifiable credentials with the
propagation route through a network that does not impose onerous
burdens on the routing system and is not overly fragile in its operation.
The extended period where the routing system has operated in a state
that essentially cannot be trusted has prompted the application layer
to generate its own mechanisms of trust. These days it is largely left
to Transport Layer Security (TLS) to determine whether a client
has reached its intended server. Given that we have been unable to
construct a secured routing system for many decades, the question
arises whether there is still the same level of need for such a system that
we had some 25 years ago, given that the application space sees this
problem as largely solved through the close-to-ubiquitous use of TLS.
This tension between the Internet layer and the upper layers of the
protocol stack is also evident in the way in which we have addressed
the perennial issue of location and identity. One of the original sim-
plifications in the IP architecture was to bundle the semantics of
identity, location, and forwarding into an IP address. While that has
proved phenomenally effective in terms of simplicity of applications
and simplicity of IP networks, it has posed some serious challenges
when considering mobility, routing, protocol transition, and network
scaling. Each of these aspects of the Internet would benefit consider-
ably if the Internet architecture allowed identity to be distinct from
location. Numerous efforts have been directed at this problem over the
past decade, particularly in IPv6, but so far, we really haven’t arrived
at an approach that feels truly comfortable in the context of IP. The
problem we appear to have been stuck on for the past decade is that
if we create a framework of applications that use identity as a ren-
dezvous mechanism and use an IP layer that requires location, then
how is the mapping between identity and location distributed in an
efficient and suitably robust manner? The transport layer of the
protocol stack has also looked at the same space and developed some
interesting approaches, as we will see in the next section.
Transport
Back in 1998 the transport layer of the IP architecture consisted of
UDP and TCP, and the network use pattern was around 95% TCP and
5% UDP. It has taken all of the intervening 25 years, but this picture
has finally changed.
Other initiatives in the transport space that are also worthy of note
include Multipath TCP and QUIC.
Multipath TCP allows separate TCP states to control the flows pass-
ing across each network path to optimise throughput. It also can
permit flow migration, allowing a logical TCP flow to switch from one
network path to another while preserving integrity. The interesting
aspect of this behaviour is that the control of the multipath behaviours
is, in the first instance, under the control of the application rather than
the host platform. This response was an early one to recognize the
increasing capacity and diversity in edge networks, and how we could
respond to this situation at the transport session level.
Numerous lessons can be drawn from the QUIC experience. Any use-
ful public communications medium needs to safeguard the privacy and
integrity of the communications that it carries.
It shifted the innovation role from the large and lumbering telco opera-
tors into the nimbler world of software. QUIC takes it one step further,
and pushes the innovation role from platforms to applications, just at
the time when platforms are declining in relative importance within
the ecosystem. From such a perspective, the emergence of an appli-
cation-centric transport model that provides faster services, a larger
repertoire of transport models, and encompassing comprehensive
encryption were inevitable developments.
By 1998 the AltaVista search engine had made its debut, and these
content-collation portals were already becoming passé. This change,
from compiling directories and lists to active search, completely
changed the Internet. These days we simply assume that we can type
any query we want to into a search engine and the search machinery
will deliver a set of pointers to relevant documents. And every time it
occurs our expectations about the quality and utility of search engines
are reinforced. Content is also changing as a result, as users no lon-
ger remain on a site and navigate around the site. Instead, users are
driving the search engines, and pulling the relevant pages without
reference to any other material. But it has not stopped there. Search
engines are morphing into “instant answer machines,” where instead
of providing a set of pointers to sources where there is a high level
of correlation between the source and the question, the search engine
attempts to extract material from the source and show what it believes
is the answer to the implicit question in the search term. Even this
process is just a way point in a longer journey, and today we are seeing
Artificial Intelligence (AI) chat bots appearing, where the underlying
data set that has been indexed by the search machinery is now being
used as a corpus of data to drive an AI chat engine. The interaction is
based on a natural language model.
Social Media A related area of profound change has been the rise of social media. The
television, radio, film, and print industries had evolved to use content
mediators, compilers, and editors to curate their content, and the wide-
spread deployment of highly capable user devices allowed end users to
directly perform content production without the need to engage with
mediators or producers. This situation has transformed many societies,
and the social media platforms, including YouTube, Flickr, Face-
book, Instagram, and TikTok, have been rocketed into societal promi-
nence, prompting major debates about the role of these platforms and
levels of societal influence that such platforms can generate.
However, using this model comes at a price, and in this case the price
lies in the motivations of the platforms that perform ad delivery. The
critical objective now is to engage the user for longer periods, so that
they can present more ads and glean more information about the user’s
profile. Merely informing a user is largely a transactional interaction,
whereas entertaining a user can be far more lucrative in terms of
generating advertising revenue because of the longer attention span.
This model has been highly successful for some content players,
particularly the current giants of streaming content, and it’s there-
fore unsurprising that the largest entities in the content world, such as
Alphabet, Microsoft, Amazon, and Apple, are more valuable in terms
of market capitalization than their counterparts in the carriage world.
We are now seeing the next round of the friction between content
and carriage, where the access network operators are arguing that the
content players should contribute to the costs of access carriage.
The Domain Name System (DNS) also merits a mention in this sec-
tion. From one perspective, little has changed in this space, and the
DNS name-resolution protocol hasn’t changed to any appreciable
extent. In some sense that’s true, but at the same time there have been
some significant changes.
The second major theme of change in the DNS concerns the larger
issue of pervasive monitoring in the DNS, highlighted by the Snowden
revelations of 2013. Most Internet transactions start with a call to
the DNS, and the meta-data contained in DNS queries and responses
provides a rich real-time profile of user activity, both in general and
potentially on a user-by-user basis. This situation has prompted
a concerted effort to improve the privacy aspects of the DNS as a
protocol. One approach has been to take the existing use of DNS
across a TCP session and add TLS to the TCP session, so the contents
of the interaction between the client and the DNS server are imper-
vious to third-party inspection or manipulation. This approach can
be taken a step further with DNS over Hypertext Transfer Protocol
Secure (HTTPS)/2, where the DNS payload has a lightweight HTTP
wrapper in addition to TLS. This approach allows DNS traffic to
be melded in with all other HTTP traffic as a further step of obscuring
DNS transactions. More recently we have seen DNS over QUIC, using
QUIC faster session start times and fast open capabilities to improve
the performance of the arrangement, and DNS over HTTPS/3, which
combines QUIC with HTTP object semantics.
This work on DNS privacy has extended into the scenarios of the
recursive resolver talking with authoritative name servers, although
it’s unclear as to the extent of the security benefits (because the end
user is not identified directly in such queries), nor is session reuse as
feasible in this scenario.
Cloud In many ways applications and services have been the high frontier of
innovation in the Internet in this period. An entire revolution in open
interconnection of content elements has taken place, and content is
now a very malleable concept. It is no longer the case of “my com-
puter, my applications, my workspace” or “your server, your content”
but an emerging model where not only the workspace for each user
is held in the network, but where the applications and services them-
selves are part of the network, and all are accessed through a generic
mechanism based around permutations of the HTTPS access model.
This world is one of the so-called Cloud Services. The world of cloud
services takes advantage of abundance in computation, storage, and
communications resources, and rather than a network facilitating users
to connect to service delivery points, the cloud model inverts the model
and attempts to bring replicant copies of content and services closer to
the user. If distance equates to cost and performance in the network-
ing world, then the cloud model dramatically shortens the distance
between consumer and content, with obvious implications in terms
of cost and performance reductions. The clouded Internet can achieve
extremely challenging performance and cost objectives by changing
the provisioning model of the content and service from “just-in-time”
on-demand service to “just-in-case” pre-provisioning of local caches
so that the local cache is ready if a local client accesses the service.
Cyber Hostility
We still are under relentless attack at all levels. We are beset by data
leaks, surveillance, profiling, disruption, and extortion.
Why can’t we fix this problem? We’ve been trying for decades, and we
just can’t seem to get ahead of the attacks. Advice to network opera-
tors to prevent the leakage of packets with forged source addresses
was published more than two decades ago, in 2000. Yet massive UDP-
based attacks with forged source addresses still persist today. Aged
computer systems with known vulnerabilities continue to be connected
to the Internet and are readily transformed into attack bots.
The larger picture of hostile attacks is not getting any better. Indeed,
it’s getting much worse. If any enterprise has a business need to main-
tain a service that is always available for use, then any form of in-house
provisioning is just not enough to withstand attack. These days only
a handful of platforms can offer resilient services, and even then, it’s
unclear whether they could withstand the most extreme of attacks.
IoT What makes this scenario even more depressing is the portent of the
so-called Internet of Things (IoT). In those circles where Internet prog-
nostications abound and policy makers flock to hear grand visions of
the future, we often hear about the boundless future represented by
this Internet of Things. This phrase encompasses some decades of the
computing industry’s transition from computers as esoteric pieces of
engineering affordable only by nations to mainframes, desktops, lap-
tops, handheld devices, and now wrist computers.
Where next? In the vision of the IoT, we are going to expand the
Internet beyond people and press on using billions of these chatter-
ing devices in every aspect of our world. What do we know about the
“things” that are already connected to the Internet? Some of them are
not very good. In fact, some of them are just plain stupid. And this stu-
pidity is toxic, in that their sometime-inadequate models of operation
and security affect others in potentially malicious ways.
But what we tend to forget is that all of these devices are built on layers
of other people’s software that is assembled into a product at the cheap-
est possible price point. It may be disconcerting to realise that the web
camera you just installed has a security model that can be summarised
with the phrase: “no security at all,” and it’s actually offering a view
of your house to the entire Internet. It may be slightly more discon-
certing to realise that your electronic wallet is on a device that is using
a massive compilation of open-source software of largely unknown
origin, with a security model that is not completely understood, but
appears to be susceptible to be coerced into being a “yes, take-all-you-
want” device. It would be nice to think that we have stopped making
mistakes in code, and from now on our software in our things will be
perfect. But that’s hopelessly idealistic. It’s just not going to happen.
Software will not be perfect. It will continue to have vulnerabilities.
Our ability to effectively defend the network and its connected hosts
continues to be, on the whole, ineffectual. Anyone who still has trust
in the integrity of the systems that make up the digital world is just
hopelessly naive. This space is toxic and hostile, and we still have no
idea how we can shift it to a different state that can resist such erosive
and insidious attacks. But somehow, we are evidently not deterred by
all this information. Somehow each of us has found a way to make the
Internet work for us.
This relationship was never going to last, and it resolved itself in ways
that in retrospect were quite predictable. From the telco perspective,
it quickly became apparent that the only reasons the telco was being
pushed to install additional network capacity at ever-increasing rates
were demands from the ISP sector. From the ISP perspective, the only
way to grow at a rate that matched customer demand was to become
one’s own carrier and take over infrastructure investment. And, in var-
ious ways, both outcomes occurred. Telcos bought up ISPs, and ISPs
became infrastructure carriers.
All this activity generated considerable investor interest, and the rapid
value escalation of the ISP industry and then the entire Internet sector
generated the levels of wild-eyed optimism that are associated only
with an exceptional boom. By 2000 almost anything associated with
the Internet, whether it was a simple portal, a new browser develop-
ment, a search engine, or an ISP, attracted investor attention, and the
valuations of Internet start-ups achieved dizzying heights. Of course,
one of the basic lessons of economic history is that every boom has an
ensuing bust, and in 2001 the Internet collapse happened. The bust
was as inevitable and as brutal as the preceding boom was euphoric.
But, like the railway boom and bust of the 1840s, after the wreckage
was cleared away what remained was a viable, and indeed a valuable,
industry.
By 2003 the era of the independent retail ISP was effectively over. But
it reshaped itself dramatically with the introduction of mobile services.
It was the old telco sector that had secured spectrum allocations in
the bidding wars in the early 2000s, and while they had thought that
mobile voice would be the reason why these investments would make
sense, it was mobile Internet services that proved to be the lasting
service model. In this period, the Internet was the amalgam of the larg-
est of the original ISP, the transformed cable television operators, and
the mobile providers. Each national regime was populated with some
three to five major service providers, and the business started to stabi-
lise around this model.
Into this world came the content world of the Internet, using cloud-
based models of service delivery to circumvent communications
bottlenecks in the long-haul transit sector. They embarked on service
models that included advertiser-funded models of content generation
and delivery and direct subscription models, and the result has been so
effective that the value of this sector is far greater than the traditional
ISP and carriage sector. The content world is now the major funder of
subsea cable systems, and the carriage world has been very reluctantly
pushed into an undistinguished commodity role as a result.
But this process has also been reflected within these edge devices. We
started with a model of a highly capable and complicated operating
system platform, and relatively simple applications that used plat-
form services. Some 25 years ago the release of Windows 98 was a
Big Thing, and rightly so. As these edge devices become more capable
and have higher processing capability, more local storage applications
have elected to take on more of the responsibility in terms of the user’s
experience. In doing so they no longer rely on the release schedules of
the platform provider, and they are no longer as concerned about the
level of control being exercised by this platform provider and gaining
an essential level of self-control. Modern browsers (Chrome and a few
far smaller fellow travellers) are far more complex than most operat-
ing systems, and they continue to subsume functions and roles that the
platform previously carried out. With DNS over HTTPS, the task of
DNS name resolution can be transformed to an application function,
rather than a common platform function. With QUIC, the transport
protocol itself has been subsumed into the application space.
Not only have we seen the commoditisation of the network over the
past 25 years, we have also seen similar commoditisation pressures on
the end-device platforms and on the operating systems used on these
devices. Even the browser space has been commoditised. The brunt
of competitive differentiation in this industry has been pushed up the
protocol stack into the content and service economy, and there is the
distinct feeling that even in that space competitive differentiation is
perhaps a misnomer, and what we have is a synthetic form of competi-
tion between a select small group of digital service-delivery behemoths
that in any other time and context would probably be called a cartel.
What Now?
It’s been a revolutionary quarter-century for us all, and the Internet has
directly or indirectly touched the lives of almost every person on this
planet. Current estimates put the number of regular Internet users at
one half of the world’s population.
Over this period, some of our expectations were achieved and then
surpassed with apparent ease, while others remained elusive. And
some things occurred that were entirely unanticipated. At the same
time, very little of the Internet we have today was confidently predicted
in 1998, while many of the problems we saw in 1998 remain problems
today.
GEOFF HUSTON AM, B.Sc., M.Sc., is the Chief Scientist at APNIC, the Regional
Internet Registry serving the Asia Pacific region. He has been closely involved with the
development of the Internet for many years, particularly within Australia, where he
was responsible for building the Internet within the Australian academic and research
sector in the early 1990s. He is author of numerous Internet-related books, and was
a member of the Internet Architecture Board from 1999 until 2005. He served on the
Board of Trustees of the Internet Society from 1992 until 2001. At various times Geoff
has worked as an Internet researcher, an ISP systems architect, and a network opera-
tor. E-mail: gih@apnic.net
Emerald Sponsors
Corporate Subscriptions