Internet DRAFT - draft-zha-dln-framework
draft-zha-dln-framework
Network Working Group Y. Zha
Internet Draft Huawei Technologies
Intended status: Informational
Expires: April 2017
October 31, 2016
Deterministic Latency Network Framework
draft-zha-dln-framework-00
Abstract
More and more real time applications, such as VR gaming, high
frequency trading, Internet of vehicle, require accurate latency
guarantee or bound, during service provisioning. Providing End-to-
End latency guarantee across Internet is increasingly attractive
while challenging. The main problem with latency guarantee on
Internet is that packet scheduling is still best-effort, and
congestion that further introduces latency uncertainty cannot be
avoided. This document presents a way of describing latency
information that mostly rely on the congestion management such as
queue scheduling, and a framework to use such information to
guarantee End-to-End latency.
Status of this Memo
This Internet-Draft is submitted in full conformance with the
provisions of BCP 78 and BCP 79.
Internet-Drafts are working documents of the Internet Engineering
Task Force (IETF), its areas, and its working groups. Note that
other groups may also distribute working documents as Internet-
Drafts.
Internet-Drafts are draft documents valid for a maximum of six
months and may be updated, replaced, or obsoleted by other
documents at any time. It is inappropriate to use Internet-Drafts
as reference material or to cite them other than as "work in
progress."
The list of current Internet-Drafts can be accessed at
http://www.ietf.org/ietf/1id-abstracts.txt
The list of Internet-Draft Shadow Directories can be accessed at
http://www.ietf.org/shadow.html
Zha Expires April 30 2017 [Page 1]
Internet-Draft DLN Framework October 2016
This Internet-Draft will expire on April 30, 2017.
Copyright Notice
Copyright (c) 2014 IETF Trust and the persons identified as the
document authors. All rights reserved.
This document is subject to BCP 78 and the IETF Trust's Legal
Provisions Relating to IETF Documents
(http://trustee.ietf.org/license-info) in effect on the date of
publication of this document. Please review these documents
carefully, as they describe your rights and restrictions with
respect to this document. Code Components extracted from this
document must include Simplified BSD License text as described in
Section 4.e of the Trust Legal Provisions and are provided without
warranty as described in the Simplified BSD License.
Table of Contents
1. Introduction .................................................2
2. Conventions used in this document ............................3
3. End-to-End Latency Guarantee Problem .........................3
3.1. Current QoS framework ...................................4
3.2. Need of Modeling of Latency .............................5
3.3. Need of Measurement of Flow Latency .....................6
4. Latency Information Modeling and Interface Framework .........7
4.1. Latency Aware PHB .......................................7
4.2. Latency Aware Interface .................................9
5. Security Considerations ......................................9
6. IANA Considerations ..........................................9
7. Acknowledgments ..............................................9
8. References ...................................................9
8.1. Normative References ....................................9
8.2. Informative References .................................10
1. Introduction
Latency sensitive applications, such as VR gaming, high frequency
trading, Internet of vehicle, have become more and more popular.
These applications often require a guaranteed or bounded End-to-
End packet transmission latency. For example, the End-to-End
latency bounds of and VR gaming are 1 millisecond and 5
millisecond corresponding. Hence, mechanism achieving End-to-End
latency guarantee or bound for latency sensitive applications is
highly desirable [I-D.liu-dln-use-cases].
Zha Expires April 30, 2017 [Page 2]
Internet-Draft DLN Framework October 2016
The well-known End-to-End QoS (Quality of Service) model is
DiffServ (Differentiated Service) [RFC2474] that defines PHB (Per
Hop Behavior) to provide different, relative service levels (or
priority) for different packets. However, one key factor for
latency during packet transmission is congestion, and congestion
management such as packet queue scheduling further introduces
latency uncertainty. DiffServ which only cares about providing
differentiated service based on packet priority is not sufficient
for the application that requires guaranteed End-to-End latency.
Therefore, mechanism is desirable that can model the packet queue
scheduling latency, then give the latency guarantee or bound of
certain packet flow in each network node. It is also useful to
define interface to expose such latency information to the upper
level of the network to facilitate End-to-End latency guaranteed
or bounded service.
2. Conventions used in this document
The key words "MUST", "MUST NOT", "REQUIRED", "SHALL", "SHALL
NOT", "SHOULD", "SHOULD NOT", "RECOMMENDED", "MAY", and "OPTIONAL"
in this document are to be interpreted as described in [RFC2119].
In this document, these words will appear with that interpretation
only when in ALL CAPS. Lower case uses of these words are not to
be interpreted as carrying [RFC2119] significance.
3. End-to-End Latency Guarantee Problem
End-to-End latency is one of the most important metrics of network
service provisioning. Latency sensitive service has become the
main driving force of deterministic low latency network, which
requires End-to-End latency guarantee or bound. Unlike relative
low latency transport, End-to-End guaranteed or bounded latency
requires the knowledge of the worst latency performance across the
network, and the capability of latency guarantee in each network
node. Current packet network is mainly best-effort, so End-to-End
latency guarantee or bound is still difficult to achieve.
The key issue is that latency guarantee requires latency
description, or the capability of the network device, in order to
further provide latency guarantee for certain traffic flow. In
this section, the challenge of End-to-End latency guarantee, as
well as the need of latency description for the network device is
presented.
Zha Expires April 30, 2017 [Page 3]
Internet-Draft DLN Framework October 2016
3.1. Current QoS framework
End-to-End QoS provisioning is the typical approach today to
provide better performance including latency, for the latency
sensitive services. DiffServ [RFC2474] is a well-known QoS
mechanism to provide differentiated service quality to different
classes of traffic flow. DiffServ model is shown in Figure 1.
+-------+
| |------------+
+-->| Meter | |
| | |-+ |
| +-------+ | |
| V V
+------------+ +--------+ +---------+ +-----------+
| | | | | Shaper/ | | |
Packets =>| Classifier |=>| Marker |=>| Dropper |=>| Scheduler |=>
| | | | | | | |
+------------+ +--------+ +---------+ |-----------+
Figure 1 - DiffServ Model
For classifiers, DiffServ uses code-points to describe the service
level and drop level of packets. The service level and drop level
are essentially relative performance description (or priority) for
different classes of traffic flows. One key factor for latency is
congestion, and congestion management such as packet queue
scheduling further introduces latency uncertainty. A deterministic
latency performance (e.g. guarantee or bound) description is still
missing while necessary for latency sensitive traffic flows.
For traffic conditioning policies, PHBs of DiffServ are applied,
including BE (Best Effort), EF (Expedited Forwarding), CS (Class
Selector) and AF (Assured Forwarding). These PHBs are designed for
only a few flow aggregates, which imply that they cannot provide
differentiated services for a potentially large amount of latency
sensitive traffic flows with different latency requirements.
For service provisioning (or packet scheduling) policies, they are
decoupled from traffic conditioning policies and haven't received
much attention. For example, in the case where an EF packet
arrives at a network device that contains other EF packets in the
queue, the latency of scheduling the EF packet is impacted by the
size of the queue. Moreover, the order of different latency aware
Zha Expires April 30, 2017 [Page 4]
Internet-Draft DLN Framework October 2016
flows arriving the EF queue hasn't been considered. Hence the
specific packet scheduling policy in specific network device is
import to the latency performance of the packet.
In summary, the current DiffServ model is not sufficient for the
application that requires guaranteed End-to-End latency. The
problems can be listed as below.
a) Current QoS mechanism like DiffServ is not well defined to
support deterministic latency performance. For example,
relative service level or packet priority can not address the
congestion factor for latency, as well as the congestion
management such as packet scheduling that introduces latency
uncertainty. Current PHBs can only support a few flow
aggregates which are not sufficient for different latency
requirements.
b) There is no user-/device-specific latency performance
specification, or no control plane mechanism to assign user-
/device-specific latency requirement to the network devices
along the path. As a result, network device has no idea how
fast a packet must be forwarded, and cannot adopt a suitable
mechanism (e.g. queue scheduling) to guarantee latency.
c) There is no mechanism supporting the measurement of flow
latency inside of a network device, especially given certain
PHB type and code-points of the flow. Such measurement will
make End-to-End latency more visible, and thus is crucial for
End-to-End latency oriented OAM.
d) Service provisioning (or packet scheduling) policies are not
specified. Packet scheduling policy and queue status are also
key factors of latency and its uncertainty. Therefore packet
scheduling policy must be considered to provide deterministic
latency service for time sensitive flows.
In a nutshell, how to explain the QoS value or how to make sure
the QoS value can be used to guarantee latency performance is not
well defined yet. Some extension to the current QoS model (e.g.
new PHB) could be useful to solve these problems.
3.2. Need of Modeling of Latency
As mentioned in problem section, QoS value or packet priority
cannot guarantee deterministic low latency. In another word, the
same QoS value or priority doesn't guarantee same latency
performance. In network device, various forwarding mechanisms and
Zha Expires April 30, 2017 [Page 5]
Internet-Draft DLN Framework October 2016
interfaces introduce different latency that may be linked to the
same priority code. There is still lack of latency performance
information that can be used to provide latency guarantee service.
The principle of Diffserv is focus on providing, describing
differentiated service of traffic flow but not deterministic
latency. Instead, queuing and scheduling, as a main part of
latency and latency uncertainty, is out of DiffServ's major
concern.
PHB provides standard way of modeling of device forwarding
behavior as well as how to handle each traffic flow. However, PHB
description does not include queuing or scheduling information. In
reality, latency is dominated by congestion control scheme, which
is mainly queuing and scheduling in the network device to take
care of multiple traffic flows arriving at the same port
simultaneously.
Therefore, extension to the current QoS model (e.g. new PHB) is
desirable as a standard way to describe the latency performance of
network device. With such standard latency model, network device
is enabled to better manage the forwarding mechanism (e.g. queue
scheduling) for the packet, in order to guarantee End-to-End
latency for the service.
3.3. Need of Measurement of Flow Latency
Flow latency measurement is also very crucial to make sure the
latency bound is not violated and useful for End-to-End latency
aware OAM mechanism. There is a need to support the measurement of
flow latency inside of a network device, especially given certain
PHB type and code-points of the flow.
Existing technologies such as OWAMP [RFC4656] and TWAMP [RFC5357]
is focused on providing one way and two way IP performance metrics.
Latency is one of metrics that can be used for End-to-End
deterministic latency provisioning. Use OWAMP/TWAMP protocols or
extension on that to support measurement of flow latency
performance is feasible.
Overlay based End-to-End latency measurement is another approach
commonly adopted by service providers like CDN vendor. Such
approach can be further enhanced by latency measurement inside
network device, for better service provisioning, e.g. traffic
steering and path selection.
Zha Expires April 30, 2017 [Page 6]
Internet-Draft DLN Framework October 2016
4. Latency Information Modeling and Interface Framework
To provide better End-to-End latency guarantee, an extension of
QoS framework is proposed to provide more accurate latency
information of network device. Mechanisms for flow latency
measurement and latency information exchange are also proposed.
This work may introduce new interfaces or extension on existing
interfaces.
4.1. Latency Aware PHB
Latency aware PHB provides latency performance information on
network device with more accurate description on latency bound of
queuing and scheduling mechanisms. In addition to classical PHB
which focus on classifier, marker and shaper/dropper, latency
aware PHB includes queuing and scheduling as well. The latency
factor is mainly introduced by queuing and scheduling algorithm
that handles congestion of multiple traffic flows. Congestion
latency can be up to hundreds of milliseconds regarding buffer
size. Implementations of various queuing, scheduling algorithms,
and QoS policies introduce latency uncertainty. Moreover,
different links cause different latency, as 1GE and 10GE will
certainly cause different latency.
The queue scheduling information model is proposed to describe
latency based service capability and capacity of network nodes. As
the assigned deterministic latency bound must be guaranteed, the
network needs to know what kinds of latency based services can be
provided by a network node for a flow with specific traffic
profile.
For latency based service capability, by referring to DiffServ
design, a differentiated service model called latency slicing
model is defined as follows. A network node provides multiple
classes of latency-bounded services, and if a flow is allocated to
a service class, then all its shaped packets can be sent out from
the network node with waiting time no more than the predefined
latency class bound. Such a latency-bounded service class is
called a latency slice. For example, a network node may have
multiple latency slices with latency bounds of 50us, 100us, 250us,
1ms, etc. Notice that actual max flow latency may vary due to
acceptance of new flows or finish of existing flows, but it will
never exceed the corresponding latency slice bound.
Zha Expires April 30, 2017 [Page 7]
Internet-Draft DLN Framework October 2016
The maximum packet length cannot exceed a predefined value called
MaxSDU, otherwise other flows latency requirement may be possibly
violated.
A latency aware flow with specific traffic profile can be accepted
by a latency slice only if the latency of any existing latency
aware flow does not violate the latency bound of its latency slice.
For example, when the traffic profile is defined by the leaky
token model (b, r), two parameters MaxBurst B and MaxRate R
representing the maximum allowable burst size and average rate are
introduced to formulate the latency slice service capacity. In
this case, a new flow can be accepted as long as b<=B and r<=R.
MaxBurst and MaxRate can be adjusted to accommodate more latency
sensitive flows in a particular latency slice, and lead to service
capacity reduction of other latency slice.
By hiding queue scheduling implementation details, the latency
slicing information model is shown in Figure 2.
+--------------+--------------+
| Name | Elements |
+--------------+--------------+
|NodeID | |
+--------------+--------------+
|SliceID | |
+--------------+--------------+
|LatencyBound | |
|--------------+--------------+
|MaxSDU | |
|--------------+--------------+
|MaxRate | |
|--------------+--------------+
|MaxBurst | |
|--------------+--------------+
Figure 2 - Latency Slice Information Model
More detail of this latency Slice information model is under
discussion and will be available in the next version.
Zha Expires April 30, 2017 [Page 8]
Internet-Draft DLN Framework October 2016
4.2. Latency Aware Interface
The queuing and scheduling information is typically localized in
the network device. In another word, the current hop has no
knowledge of how the flow was scheduled on last hop. So it is hard
to guarantee End-to-End latency if latency information only
affects locally. Latency performance such as queuing and
scheduling information in network device needs to be exposed for
End-to-End latency guarantee service provisioning.
Basically, more information needs to be exchanged than the current
DiffServ code-points. Existing approach such as [RFC7297] defines
CPP that contains information of connection attributes, latency,
loss, and so on.
With latency performance information defined on previous section,
an interface or mechanism to exchange such information is proposed.
A simple approach can be an interface from network device to
controller, which collects all the latency performance information
of each hop, and then make a decision how to serve the flow at
each hop. In this case, the controller tells each hop how to serve
the flow with queuing, scheduling information that can be
understood by each hop.
The detail of latency aware interface is under discussion and will
be available in the next version.
5. Security Considerations
TBD
6. IANA Considerations
This document has no actions for IANA.
7. Acknowledgments
This document has benefited from reviews, suggestions, comments
and proposed text provided by the following members, listed in
alphabetical order: Jinchun Xu and Hengjun Zhu.
8. References
8.1. Normative References
[RFC2119] S. Bradner, "Key words for use in RFCs to Indicate
Requirement Levels", BCP 14, RFC 2119, March 1997.
Zha Expires April 30, 2017 [Page 9]
Internet-Draft DLN Framework October 2016
[RFC3393] C. Demichelis, "IP Packet Delay Variation Metric for IP
Performance Metrics (IPPM) ", RFC 3393, November 2002.
[RFC2474] K. Nichols, "Definition of the Differentiated Services
Field (DS Field) in the IPv4 and IPv6 Headers", RFC 2474,
December 1998.
[RFC4656] S. Shalunov, "A One-way Active Measurement Protocol
(OWAMP)", RFC 4656, September 2006.
[RFC5357] K. Hedayat, "A Two-Way Active Measurement Protocol
(TWAMP)", RFC 5357, October, 2008.
[RFC7279] M. Boucadair, "IP Connectivity Provisioning Profile
(CPP)", RFC 7297, July, 2014.
8.2. Informative References
[I-D.finn-detnet-problem-statement]
Finn, N. and P. Thubert, "Deterministic Networking Problem
Statement", draft-ietf-detnet-problem-statement-01 (work in
progress), September 2016.
[I-D.finn-detnet-architecture]
Finn, N., Thubert, P., and M. Teener, "Deterministic Networking
Architecture", draft-ietf-detnet-architecture-00 (work in
progress), September 2016.
[I-D.liu-dln-use-cases]
Liu, X., "Deterministic Latency Network Use Cases", draft-liu-dln-
use-cases-00 (work in progress), October 2016.
Authors' Addresses
Yiyong Zha
Huawei Technologies
Email: zhayiyong@huawei.com
Zha Expires April 30, 2017 [Page 10]