Internet DRAFT - draft-zzhang-pim-multicast-scaling-considerations

draft-zzhang-pim-multicast-scaling-considerations







pim                                                             Z. Zhang
Internet-Draft                                          Juniper Networks
Intended status: Informational                                 R. Parekh
Expires: 27 April 2023                                             Cisco
                                                              H. Bidgoli
                                                                   Nokia
                                                                Z. Zhang
                                                                     ZTE
                                                         24 October 2022


                    Multicast Scaling Considerations
          draft-zzhang-pim-multicast-scaling-considerations-00

Abstract

   This informational document discusses various multicast scaling
   aspects, compares different multicast technologies with respect to
   scaling, and suggests a general approach of combined solutions to
   scale multicast.  This discussion is independent of IPv4/IPv6 or
   MPLS/SRv6 data planes.

Status of This Memo

   This Internet-Draft is submitted in full conformance with the
   provisions of BCP 78 and BCP 79.

   Internet-Drafts are working documents of the Internet Engineering
   Task Force (IETF).  Note that other groups may also distribute
   working documents as Internet-Drafts.  The list of current Internet-
   Drafts is at https://datatracker.ietf.org/drafts/current/.

   Internet-Drafts are draft documents valid for a maximum of six months
   and may be updated, replaced, or obsoleted by other documents at any
   time.  It is inappropriate to use Internet-Drafts as reference
   material or to cite them other than as "work in progress."

   This Internet-Draft will expire on 27 April 2023.

Copyright Notice

   Copyright (c) 2022 IETF Trust and the persons identified as the
   document authors.  All rights reserved.

   This document is subject to BCP 78 and the IETF Trust's Legal
   Provisions Relating to IETF Documents (https://trustee.ietf.org/
   license-info) in effect on the date of publication of this document.
   Please review these documents carefully, as they describe your rights



Zhang, et al.             Expires 27 April 2023                 [Page 1]

Internet-Draft              Multicast Scaling               October 2022


   and restrictions with respect to this document.  Code Components
   extracted from this document must include Revised BSD License text as
   described in Section 4.e of the Trust Legal Provisions and are
   provided without warranty as described in the Revised BSD License.

Table of Contents

   1.  Different Scaling Aspects . . . . . . . . . . . . . . . . . .   2
     1.1.  Scaling in the number of receivers  . . . . . . . . . . .   3
     1.2.  Scaling in the number of trees  . . . . . . . . . . . . .   3
   2.  Multicast Tunnels . . . . . . . . . . . . . . . . . . . . . .   3
   3.  New Multicast Technologies  . . . . . . . . . . . . . . . . .   4
     3.1.  BIER  . . . . . . . . . . . . . . . . . . . . . . . . . .   4
     3.2.  BIER-TE . . . . . . . . . . . . . . . . . . . . . . . . .   5
     3.3.  CGM2  . . . . . . . . . . . . . . . . . . . . . . . . . .   5
   4.  Multicast Tunnel Segmentation . . . . . . . . . . . . . . . .   6
   5.  Signaling for Tunneling and Segmentation  . . . . . . . . . .   7
     5.1.  MVPN as Flow Overlay Signaling for Tunneling  . . . . . .   7
     5.2.  PIM/mLDP as Flow Overlay Signaling over BIER  . . . . . .   8
     5.3.  Flow Overlay Signaling and Tunnel Segmentation  . . . . .   8
   6.  Overall Considerations for Multicast Scaling  . . . . . . . .   8
     6.1.  Observations  . . . . . . . . . . . . . . . . . . . . . .   9
     6.2.  Considerations  . . . . . . . . . . . . . . . . . . . . .   9
       6.2.1.  Reduce Number of Underlay Tunnels or Amount of Tunnel
               Binding Signaling . . . . . . . . . . . . . . . . . .   9
       6.2.2.  Scale Up Segmentation Points  . . . . . . . . . . . .  10
       6.2.3.  Scale out Segmentation Points . . . . . . . . . . . .  11
   7.  Summary . . . . . . . . . . . . . . . . . . . . . . . . . . .  12
   8.  Informative References  . . . . . . . . . . . . . . . . . . .  12
   Authors' Addresses  . . . . . . . . . . . . . . . . . . . . . . .  15

1.  Different Scaling Aspects

   IP Multicast [RFC1112] architecture involves an IP multicast group/
   destination address that, together with the source address,
   identifies a multicast tree rooted at the First Hop Router (FHR) that
   the source is attached to.  Multicast traffic is forwarded along the
   tree, which is typically set up by Protocol Independent Multicast
   (PIM) [RFC7761].

   In this document, each (source, group) pair is referred to a flow or
   a tree.  Typically, each flow is for a "piece of content", e.g., an
   audio or video content.

   While a bidirectional tree [RFC5015] can be used for multicast among
   a set of sources and receivers using only group/destination address,
   it is typically used only in a small domain and is not considered in
   this document.



Zhang, et al.             Expires 27 April 2023                 [Page 2]

Internet-Draft              Multicast Scaling               October 2022


1.1.  Scaling in the number of receivers

   With the above-mentioned multicast tree, each tree node only need to
   replicate to a minimum number of downstream nodes, which could be
   transit or leaf nodes.  Except in the case of Broadband Node Gateway
   (BNG) connecting to a huge number of home subscribers or in the case
   of a spine switch connecting to a large number of leaf switches in a
   Data Center, the number of replication is small, yet the number of
   total receivers can be unlimited as the tree grows lengthwise and
   width-wise.

1.2.  Scaling in the number of trees

   The number of (source, group) pairs could be huge.  Each (source,
   group) pair would need a tree, with each replication point being a
   tree node.  Each tree node needs to have state specifically for each
   tree both in forwarding plane and control plane (that is used to set
   up the forwarding plane state).

   The number of multicast trees that a network can support may have to
   be limited because of the tree state, yet the number of multicast
   trees transiting a network may be huge (simply because of the huge
   number of multicast flows).

   Chances are, many of the flows have a common set of ingress and
   egress points in the transit network.  In that case, instead of
   setting up individual multicast (source, group) trees across this
   network, tunnels can be to transport those individual trees.  For
   example, a single mLDP [RFC6388] tunnel can be used to transport
   thousands of IP multicast flows, greatly reducing the number of trees
   in the network.

2.  Multicast Tunnels

   A multicast tunnel also corresponds to a multicast tree and requires
   corresponding state on tree nodes, though it is not identified by a
   (source, group) pair.

   In case of mLDP [RFC6388] or RSVP-TE P2MP [RFC4875] tunnel, the tree
   is identified by an mLDP FEC or RSVP P2MP Session Object in the
   control plane, and a label in the forwarding plane.  The tree will
   likely have transit (non-root/leaf) tree nodes.

   An MPLS SR-P2MP [I-D.ietf-pim-sr-p2mp-policy] tunnel is identical to
   mLDP/RSVP-TE in the forwarding plane.  It's only different from the
   latter in the control plane - different control plane identifier and
   different setup procedures, but still requires per-tree state on all
   tree nodes.



Zhang, et al.             Expires 27 April 2023                 [Page 3]

Internet-Draft              Multicast Scaling               October 2022


   An SRv6-P2MP tunnel is also similar to mLDP/RSVP-TE/SR MPLS P2MP
   tunnel even in data plane - it's just that the data plane identifier
   is now part of IPv6 destination address.

   An Ingress Replication (IR) [RFC7988] tunnel is a special/degraded
   multicast tree in that it does not have transit tree nodes.  The tree
   root (ingress) replicates traffic and tunnel to tree leaves directly
   using unicast, without the need for tree state on transit routers.
   Obviously, it does not have efficient replication - the ingress may
   need to send multiple copies through common downstream nodes - but it
   may still be desired in certain situation.

   We refer the tunnels to as underlay tunnels and multicast traffic
   (e.g.  IP Multicast) that the undelay tunnels carry to as overlay
   flows.  Notice that, an underlay tunnel could be carried by another
   underlay tunnel (e.g. a part of an mLDP tunnel is transported by
   RSVP-TE P2MP tunnel).

3.  New Multicast Technologies

   The IP multicast and non-IR tunnels described above require per-tree/
   tunnel state on transit tree nodes.  While use of tunnels removes
   individual IP Multicast trees in the tunnels' region, the number of
   tunnels may still be large.  New multicast technologies have been
   developed to remove the per-tree/tunnel state while still allow
   efficient replication, as summarized below.

3.1.  BIER

   Bit Index Explicit Replication (BIER) [RFC8279] is a multicast
   architecture in which forwarding is based on a BitString in the BIER
   header preceding the payload.  Each Bit Position (BP) that is set in
   the BitString represents a BIER Forwarding Egress Router (BFER) that
   needs to receive the packet.  A BIER Forwarding Router (BFR) checks
   the set BPs to determine the set of next hop BFR to replicate the
   packet to, by repeatedly using a BP as lookup key in a BIER Index
   Forwarding Table (BIFT).  A clever way ensures that the number of
   lookup is bound by the number of replications, no matter how many and
   which BPs are set.

   The length of the BitString is limited by maximum transmission unit
   (MTU) of the BIER domain, the reserved space in a packet for
   payloads, and more importantly the implementation tradeoff in the
   forwarding plane.  A typical value is 256.  While larger BitString
   lengths could be used, it may be suboptimal if the BFERs for a packet
   are sparsely distributed.





Zhang, et al.             Expires 27 April 2023                 [Page 4]

Internet-Draft              Multicast Scaling               October 2022


   If the number of BFERs is larger than the BitString length, there are
   two ways to make it work:

   *  Send multiple copies - each copy is for a different set of BFERs.
      The same BP in different copies corresponds to different BFERs.
      This not only requires multiple copies but also multiple BIFTs -
      one BIFT for each set.

   *  Divide the BIER domain into per-region sub-domains to use tunnel
      segmentation (Section 4).  A segmentation point decapsulates the
      BIER packets received in an upstream region and re-encapsulate
      BIER packets for downstream regions.

   BIER can be considered as a tunneling technology - overlay flows
   (e.g.  IP multicast) are tunneled over BIER, even though there are
   not per-tunnel state in a BIER domain.

   The BIER architecture includes a flow overlay on top of BIER layer
   that does BIER signaling and forwarding, which is in turn over a
   routing underlay that forwards BIER traffic from BIER Forwarding
   Ingress Routers (BFIRs) to BFERs.  The purpose of flow overlay is for
   BFERs to signal BFIRs that they are interested in certain overlay
   flows so that the BFIRs can derive what BitString to use for an
   overlay flow.

3.2.  BIER-TE

   BIER does not provide per-flow traffic engineering capability.  BIER
   Traffic Engineering (BIER-TE) [I-D.ietf-bier-te-arch] does so without
   requiring per-flow state inside the BIER domain.

   With BIER-TE, a BP may indicate a replication branch - not just a
   BFER anymore.  Because of that, with the same BitString length, fewer
   BFERs will be covered by a packet - whether we send multiple copies
   or use tunnel segmentation.

3.3.  CGM2

   With BIER and BIER-TE, the BPs in the BitString have "global"
   ("subdomain-wide" to be strict) significance and all BFRs in a
   subdomain must have entries for all of them in the BIFTs.  However,
   the BIER-TE concept can be augmented to work with BPs of local
   significance.  This is referred to as Carrier Grade Minimalist
   Multicast (CGM2) [I-D.eckert-bier-cgm2-rbs].

   The CGM2 concept could be more easily understood through
   [I-D.chen-pim-srv6-p2mp-path], which uses SRv6 SIDs to encode
   replication branches.  Replacing the SIDs with BPs you get CGM2.



Zhang, et al.             Expires 27 April 2023                 [Page 5]

Internet-Draft              Multicast Scaling               October 2022


   Obviously CGM2 scales better than [I-D.chen-pim-srv6-p2mp-path] in
   that many fewer bits are needed in the packet header so this document
   focuses on CGM2.

   While CGM2 does use fewer BIFT entries, it does not save on the
   number of bits needed to encode all replication branches.  In fact,
   it uses more bits to turn the flat-structured BitString in BIER-TE to
   recursive/hierarchical constructs, and that also leads to more
   complicated forwarding behavior.

   Given this limitation, while CGM2 can be used as underlay tunnels, it
   is not really a good fit for Carrier Grade multicast when overlay IP
   multicast is not used (i.e. the end receivers are directly encoded as
   BPs).

4.  Multicast Tunnel Segmentation

   An end-to-end (E2E) multicast flow may need to go through a vast
   network consisting of many ASes and/or IGP areas (referred to
   regions).  When tunneling overlay multicast flows, different types/
   instances of tunnels may need to be used in different regions.  This
   is referred to as tunnel segmentation [RFC6514] [RFC7524], and it is
   because of the following reasons:

   *  Due to technological or administrative reasons, different tunnel
      technologies may have to be used in different regions.  For
      example, one region may use mLDP while another may use IR.

   *  For the same reasons, different tunnel instances of the same type
      may have to be used in different regions.

   *  Even if the same tunnel could be established across multiple
      regions, different tunnel instances in different regions may be
      desired for better optimization.  For example, if a set of flows
      have very different sets of egress points, it is not desired to
      carry them in a single across-region tunnel but they may be
      carried by a single intra-region tunnel in one region and
      different intra-region tunnels in other regions.

   *  In case of IR, instead of directly replicating to all egress
      points, it is more efficient to ingress replicate to segmentation
      points who in turn ingress replicate to the next set of
      segmentation and/or egress points.

   *  In case of BIER, when the number of BFERs is larger than the
      BitString length, segmentation may be used together with smaller
      subdomains.




Zhang, et al.             Expires 27 April 2023                 [Page 6]

Internet-Draft              Multicast Scaling               October 2022


   At the segmentation points, overlay state is needed to stitch the
   upstream segments and downstream segments together.

5.  Signaling for Tunneling and Segmentation

   To carry multicast traffic over a tunnel, the tunnel ingress needs to
   know what flows to be put onto which tunnel.  Depending on the tunnel
   type, it may also need to know the tunnel egresses.  The tunnel
   egresses may also need to know that they need to join the tunnel.
   This requires signaling - note that this is not for setting up the
   underlay tunnel, but for "flow overlay".

5.1.  MVPN as Flow Overlay Signaling for Tunneling

   Consider some E2E IP multicast trees of a customer with part of them
   go over a VPN provider network.  The PE routers tunnel the traffic
   across the provider network based on PIM-MVPN or BGP-MVPN signaling
   specified in [RFC6037] [RFC6513] and [RFC6514].  In particular, Data
   MDT or I/S-PMSI A-D routes are used to announce the binding of
   overlay flows to underlay tunnels.  The binding could be inclusive
   (one tunnel for all flows in a VPN and to all egress PEs - referred
   to as an inclusive tunnel) or selective (one tunnel for one or more
   flows and only to the egress PEs that need to receive the traffic -
   referred to as a selective tunnel).  While selective binding prevents
   multicast traffic from being sent to PEs that do not need to receive
   traffic, inclusive binding can significantly reduce the number of
   tunnels needed (only one tunnel is used for each VPN but traffic is
   sent to all the PEs of the VPN).

   Two other features of BGP-MVPN can further reduce the number of
   underlay tunnels.  One is to use the same tunnel for flows in
   different VPNs (referred to as tunnel aggregation in this document).
   This can be done for all flows or just for some selective flows in
   the VPNs, achieved by advertising a de-multiplex VPN label in the
   PMSI Tunnel Attribute (PTA) attached to the PMSI A-D routes and
   imposing the de-multiplex label before imposing the tunnel label to
   the traffic.  The other is inter-as aggregation where per-PE
   inclusive tunnels are confined to the local AS while per-AS inclusive
   tunnels are used outside the local AS.  This is achieved with Inter-
   AS I-PMSI A-D routes [RFC6514] and the concept is further explained
   and extended to per-Region Aggregation (Section 6.2 of
   [I-D.ietf-bess-evpn-bum-procedure-updates]).

   The same BGP-MVPN procedures can be applied to Global Table Multicast
   (GTM) for multicast in the default/master routing instance over an
   underlay network, as specified in [RFC7716].





Zhang, et al.             Expires 27 April 2023                 [Page 7]

Internet-Draft              Multicast Scaling               October 2022


   Note that this section is about MVPN as Flow Overlay Signaling for
   any kind of tunneling technology including BIER [RFC8556], and the
   next section is about two flow overlay signaling methods specifically
   for BIER.

5.2.  PIM/mLDP as Flow Overlay Signaling over BIER

   Consider an E2E IP multicast tree with part of it tunneled over BIER.
   The flow overlay signaling can be PIM, which is both the signaling
   protocol for the E2E tree and the overlay signaling protocol for
   tunneling over BIER [I-D.ietf-bier-pim-signaling].

   The above applies to IP multicast in both the default/global routing
   instance and in VRFs in case of MVPN, and it replaces PIM-MVPN
   [RFC6037] [RFC6513] in this context.

   Similarly, an E2E mLDP tree can have part of it tunneled over BIER.
   In this case, the flow overlay signaling is mLDP as specified in
   [I-D.ietf-bier-mldp-signaling-over-bier].

   Note that, "tunnel segmentation" is originally documented in BGP-MVPN
   (Section 4) and it refers to that a PE-PE multicast provider tunnel
   can be instantiated with different types/instances of tunnels in
   different ASes/regions.  Strictly speaking, an IP multicast tree
   going over different tunnels in different regions is different from
   the MVPN "tunnel segmentation".  However, in the rest of the document
   we use "tunnel segmentation" for both situations.

5.3.  Flow Overlay Signaling and Tunnel Segmentation

   In case of BGP-MVPN as overlay signaling (for any kind of tunnels -
   see Section 5.1) with tunnel segmentation, the segmentation points
   maintain overlay state in the form of I-PMSI A-D routes that are for
   all flows or S-PMSI A-D routes for specific flows, instead of overlay
   (e.g., IP) multicast tree state.

   In case of PIM/mLDP as BIER flow overlay signaling (Section 5.2),
   when the BIER domain is divided into multiple sub-domains for
   segmentation purpose, the overlay state that the segmentation points
   maintain is the overlay multicast tree state itself (i.e., IP
   multicast tree state or mLDP tree state).

6.  Overall Considerations for Multicast Scaling

   With all the background laid out above, we have the following
   observations and considerations.





Zhang, et al.             Expires 27 April 2023                 [Page 8]

Internet-Draft              Multicast Scaling               October 2022


6.1.  Observations

   For a massive number of flows to reach a massive number of receivers,
   the best solution is IP multicast (Section 1.1) carried over tunnels
   (Section 1.2).

   With massive number of receivers and a large span of an E2E network,
   an E2E multicast overlay flow may need to be carried by multiple
   tunnels/segments of different types or instances in different parts
   of the E2E network (Section 1.2, Section 4).

   Even if BIER/BIER-TE/CGM2 is supported on all devices E2E, it may be
   impractical to encode all BFERs or BIER-TE/CGM2 replication branches
   in a single BitString (flat or recursive), and sending multiple
   copies is just not efficient hence undesired or impractical.  Tunnel
   segmentation needs to be used so that different sub-domains can be
   used in different regions of the E2E network, and that requires the
   overlay state/signaling at the segmentation points.

   However, the segmentation points may become the scaling bottleneck
   due to the overlay state/signaling.  That may be mitigated by
   solutions described below.

6.2.  Considerations

   As observed above, to massively scale multicast in both dimensions
   (number of receivers and number of flows) and over a vast network, IP
   multicast over underlay tunnel segments is needed.  This section
   focuses on how to reduce the number of underlay tunnels and how to
   scale segmentation points.

6.2.1.  Reduce Number of Underlay Tunnels or Amount of Tunnel Binding
        Signaling

   As discussed in Section 5.1, underlay tunnels can be reduced by use
   of following:

   *  Inclusive tunnels

   *  Tunnel aggregation

   *  Per-region aggregation

   Notice that while they all reduce the number of underlay tunnels
   (which is useful, otherwise the underlay network need to maintain
   more tunnel state unless BIER/BIER-TE/CGM2 is used), tunnel
   aggregation does not reduce the overlay signaling of binding between
   tunnel and overlay flows.



Zhang, et al.             Expires 27 April 2023                 [Page 9]

Internet-Draft              Multicast Scaling               October 2022


   Therefore, if segmented selective tunnels are used, or if there are
   just too many VPNs to support (such that even the number of I-PMSI
   A-D routes is too large), segmentation point scaling is still needed
   as discussed below.

   On the other hand, the trend of multicast application seems to be
   that it is mainly for large scale delivery of real-time high rate
   data that most likely need to be sent everywhere, e.g., broadcasting
   of World Cup Soccer or Chinese Spring Festival Gala.  In this case
   inclusive tunnels in a few VPNs may very well suffice.

   Note that, while [RFC6514] and [RFC6625] specifies that the source/
   group length fields of S-PMSI A-D routes to be either 0 (wildcard) or
   32/128 (IPv4/IPv6 host address), it could be extended to have lengths
   between 0 and 32/128.  With that, a set of overlay flows not only can
   share the same underlay tunnel but can also share the same S-PMSI A-D
   route.

   Part of an underlay tunnel itself can be stacked on top of another
   multicast tunnel.  When multiple upper tunnels stack on top of the
   same lower tunnel, the lower tunnel's domain does not need to
   maintain state for the upper tunnels.  Examples include mLDP over
   RSVP-TE/mLDP/other P2MP tunnels ([RFC7060], mLDP over MVPN [RFC6514]
   where mLDP is the C-multicast protocol, and mLDP over BIER
   [I-D.ietf-bier-mldp-signaling-over-bier].

6.2.2.  Scale Up Segmentation Points

   The scaling burden on a segmentation point falls on both control
   plane (for overlay signaling) and forwarding plane.  While a typical
   router implementation supports much lower number of multicast flows
   than unicast ones, it does not mean the multicast scale cannot be
   increased.

6.2.2.1.  Forwarding Plane Scaling

   A modern edge router can handle hundreds of thousands if not millions
   of (unicast) routes in the forwarding plane.  For multicast
   forwarding, the scaling property is actually similar to the unicase
   case - the only difference is that the lookup key in case of IP
   multicast is (source, group/destination) pair instead of just a
   destination address.  The forwarding action is based on the
   forwarding instructions (e.g. a set of replication branches),
   referred to as forwarding nexthop associated with the route that is
   looked up.






Zhang, et al.             Expires 27 April 2023                [Page 10]

Internet-Draft              Multicast Scaling               October 2022


   On the other hand, while many multicast routes can have the same
   forwarding nexthop just like in unicast case, on segmentation points
   the multicast forwarding nexthop sharing is limited as explained in
   Section 2.1 of [I-D.zzhang-bess-mvpn-evpn-segmented-forwarding].

   Therefore, a modern router designed with multicast scaling in
   consideration (e.g. has enough memory for multicast state especially
   more unshared forwarding nexthop) should be able to handle hundreds
   of thousands if not millions of multicast flows.

   Of course, scaling must always be considered together with
   convergence.  For example, when something happens, how fast can the
   FIB state be updated.  This is further discussed below.

6.2.2.2.  Control Plane Scaling

   While it might be challenging for some PIM [RFC7761] implementations
   to handle hundreds of thousands of multicast flows due to the
   following reasons:

   *  Periodic refreshes due to soft state nature

   *  Potentially huge number of impacted flows when an interface goes
      up/down

   Overlay multicast scaling does not have to rely on PIM (so no soft
   state refresh problem) and is less prone to topology changes.

   Taking the example of BGP-MVPN [RFC6514] as the overlay protocol,
   overlay multicast interest is signaled with BGP-MVPN type-6/7
   (C-Multicast Auto-Discovery or A-D) routes, augmented with type-4
   (Leaf A-D) routes in case of selective IR/BIER tunnels.  None of the
   above-mentioned scaling concerns are applicable here.

   It is possible that an underlay topology change could impact some
   underlay tunnels.  A good implementation should be able to minimize
   the impact to the overlay forwarding state.  In the BIER case, no
   overlay forwarding state needs to be changed at all.

6.2.3.  Scale out Segmentation Points

   Consider that between two segmentation regions there are 2N
   segmentation points (Regional Border Routers, or RBRs) in parallel.
   They are divided into N pairs where each pair is responsible for 1/N
   of overlay flows (a pair is used for redundancy purposes).  By
   increasing the number N, we can scale out the segmentation points
   (whether scale up is done at the same time or not).




Zhang, et al.             Expires 27 April 2023                [Page 11]

Internet-Draft              Multicast Scaling               October 2022


7.  Summary

   As discussed above, existing and in-development multicast solutions,
   when implemented properly and deployed with appropriate combinations,
   can scale very well to massive number of multicast flows with massive
   number of receivers over a vast network:

   *  Use IP multicast to scale for massive number of receivers
      (Section 1.1)

   *  Use tunneling to scale for massive number of flows (Section 1.2)

   *  Use inclusive tunnels and/or aggregation to reduce the number of
      underlay tunnels needed (Section 5.1)

   *  Use BIER to achieve efficient replication without requiring per-
      tree state (Section 3.1)

   *  Use BIER-TE/CGM2 to achieve per-flow traffic steering without
      requiring per-tree state (Section 3.2, Section 3.3)

   *  Use tunnel segmentation to scale for a vast network that may
      deploy different types/instances of tunnels in different regions
      (Section 4)

   *  Scale up (Section 6.2.2) and scale out (Section 6.2.3) tunnel
      segmentation points

   It's worth pointing out that the above is independent of IPv4/IPv6 or
   MPLS/SRv6 data planes.

8.  Informative References

   [I-D.chen-pim-srv6-p2mp-path]
              Chen, H., McBride, M., Fan, Y., Li, Z., Geng, X., Toy, M.,
              Gyan Mishra, S., Wang, A., Liu, L., and X. Liu, "Stateless
              SRv6 Point-to-Multipoint Path", Work in Progress,
              Internet-Draft, draft-chen-pim-srv6-p2mp-path-06, 30 April
              2022, <https://www.ietf.org/archive/id/draft-chen-pim-
              srv6-p2mp-path-06.txt>.











Zhang, et al.             Expires 27 April 2023                [Page 12]

Internet-Draft              Multicast Scaling               October 2022


   [I-D.eckert-bier-cgm2-rbs]
              Eckert, T. and B. (. Xu, "Carrier Grade Minimalist
              Multicast (CGM2) using Bit Index Explicit Replication
              (BIER) with Recursive BitString Structure (RBS)
              Addresses", Work in Progress, Internet-Draft, draft-
              eckert-bier-cgm2-rbs-01, 9 February 2022,
              <https://www.ietf.org/archive/id/draft-eckert-bier-cgm2-
              rbs-01.txt>.

   [I-D.ietf-bess-evpn-bum-procedure-updates]
              Zhang, Z., Lin, W., Rabadan, J., Patel, K., and A.
              Sajassi, "Updates on EVPN BUM Procedures", Work in
              Progress, Internet-Draft, draft-ietf-bess-evpn-bum-
              procedure-updates-14, 18 November 2021,
              <https://www.ietf.org/archive/id/draft-ietf-bess-evpn-bum-
              procedure-updates-14.txt>.

   [I-D.ietf-bier-mldp-signaling-over-bier]
              Bidgoli, H., Kotalwar, J., Wijnands, I., Mishra, M.,
              Zhang, Z., and E. Leyton, "M-LDP Signaling Through BIER
              Core", Work in Progress, Internet-Draft, draft-ietf-bier-
              mldp-signaling-over-bier-01, 12 November 2021,
              <https://www.ietf.org/archive/id/draft-ietf-bier-mldp-
              signaling-over-bier-01.txt>.

   [I-D.ietf-bier-pim-signaling]
              Bidgoli, H., Xu, F., Kotalwar, J., Wijnands, I., Mishra,
              M., and Z. Zhang, "PIM Signaling Through BIER Core", Work
              in Progress, Internet-Draft, draft-ietf-bier-pim-
              signaling-12, 25 July 2021,
              <https://www.ietf.org/archive/id/draft-ietf-bier-pim-
              signaling-12.txt>.

   [I-D.ietf-bier-te-arch]
              Eckert, T., Menth, M., and G. Cauchie, "Tree Engineering
              for Bit Index Explicit Replication (BIER-TE)", Work in
              Progress, Internet-Draft, draft-ietf-bier-te-arch-13, 25
              April 2022, <https://www.ietf.org/archive/id/draft-ietf-
              bier-te-arch-13.txt>.

   [I-D.ietf-pim-sr-p2mp-policy]
              (editor), D. V., Filsfils, C., Parekh, R., Bidgoli, H.,
              and Z. Zhang, "Segment Routing Point-to-Multipoint
              Policy", Work in Progress, Internet-Draft, draft-ietf-pim-
              sr-p2mp-policy-05, 2 July 2022,
              <https://www.ietf.org/archive/id/draft-ietf-pim-sr-p2mp-
              policy-05.txt>.




Zhang, et al.             Expires 27 April 2023                [Page 13]

Internet-Draft              Multicast Scaling               October 2022


   [I-D.zzhang-bess-mvpn-evpn-segmented-forwarding]
              Zhang, Z. and J. Xie, "MVPN/EVPN Segmentated Forwarding
              Options", Work in Progress, Internet-Draft, draft-zzhang-
              bess-mvpn-evpn-segmented-forwarding-00, 20 December 2018,
              <https://www.ietf.org/archive/id/draft-zzhang-bess-mvpn-
              evpn-segmented-forwarding-00.txt>.

   [RFC1112]  Deering, S., "Host extensions for IP multicasting", STD 5,
              RFC 1112, DOI 10.17487/RFC1112, August 1989,
              <https://www.rfc-editor.org/info/rfc1112>.

   [RFC4875]  Aggarwal, R., Ed., Papadimitriou, D., Ed., and S.
              Yasukawa, Ed., "Extensions to Resource Reservation
              Protocol - Traffic Engineering (RSVP-TE) for Point-to-
              Multipoint TE Label Switched Paths (LSPs)", RFC 4875,
              DOI 10.17487/RFC4875, May 2007,
              <https://www.rfc-editor.org/info/rfc4875>.

   [RFC5015]  Handley, M., Kouvelas, I., Speakman, T., and L. Vicisano,
              "Bidirectional Protocol Independent Multicast (BIDIR-
              PIM)", RFC 5015, DOI 10.17487/RFC5015, October 2007,
              <https://www.rfc-editor.org/info/rfc5015>.

   [RFC6037]  Rosen, E., Ed., Cai, Y., Ed., and IJ. Wijnands, "Cisco
              Systems' Solution for Multicast in BGP/MPLS IP VPNs",
              RFC 6037, DOI 10.17487/RFC6037, October 2010,
              <https://www.rfc-editor.org/info/rfc6037>.

   [RFC6388]  Wijnands, IJ., Ed., Minei, I., Ed., Kompella, K., and B.
              Thomas, "Label Distribution Protocol Extensions for Point-
              to-Multipoint and Multipoint-to-Multipoint Label Switched
              Paths", RFC 6388, DOI 10.17487/RFC6388, November 2011,
              <https://www.rfc-editor.org/info/rfc6388>.

   [RFC6513]  Rosen, E., Ed. and R. Aggarwal, Ed., "Multicast in MPLS/
              BGP IP VPNs", RFC 6513, DOI 10.17487/RFC6513, February
              2012, <https://www.rfc-editor.org/info/rfc6513>.

   [RFC6514]  Aggarwal, R., Rosen, E., Morin, T., and Y. Rekhter, "BGP
              Encodings and Procedures for Multicast in MPLS/BGP IP
              VPNs", RFC 6514, DOI 10.17487/RFC6514, February 2012,
              <https://www.rfc-editor.org/info/rfc6514>.

   [RFC6625]  Rosen, E., Ed., Rekhter, Y., Ed., Hendrickx, W., and R.
              Qiu, "Wildcards in Multicast VPN Auto-Discovery Routes",
              RFC 6625, DOI 10.17487/RFC6625, May 2012,
              <https://www.rfc-editor.org/info/rfc6625>.




Zhang, et al.             Expires 27 April 2023                [Page 14]

Internet-Draft              Multicast Scaling               October 2022


   [RFC7060]  Napierala, M., Rosen, E., and IJ. Wijnands, "Using LDP
              Multipoint Extensions on Targeted LDP Sessions", RFC 7060,
              DOI 10.17487/RFC7060, November 2013,
              <https://www.rfc-editor.org/info/rfc7060>.

   [RFC7524]  Rekhter, Y., Rosen, E., Aggarwal, R., Morin, T.,
              Grosclaude, I., Leymann, N., and S. Saad, "Inter-Area
              Point-to-Multipoint (P2MP) Segmented Label Switched Paths
              (LSPs)", RFC 7524, DOI 10.17487/RFC7524, May 2015,
              <https://www.rfc-editor.org/info/rfc7524>.

   [RFC7716]  Zhang, J., Giuliano, L., Rosen, E., Ed., Subramanian, K.,
              and D. Pacella, "Global Table Multicast with BGP Multicast
              VPN (BGP-MVPN) Procedures", RFC 7716,
              DOI 10.17487/RFC7716, December 2015,
              <https://www.rfc-editor.org/info/rfc7716>.

   [RFC7761]  Fenner, B., Handley, M., Holbrook, H., Kouvelas, I.,
              Parekh, R., Zhang, Z., and L. Zheng, "Protocol Independent
              Multicast - Sparse Mode (PIM-SM): Protocol Specification
              (Revised)", STD 83, RFC 7761, DOI 10.17487/RFC7761, March
              2016, <https://www.rfc-editor.org/info/rfc7761>.

   [RFC7988]  Rosen, E., Ed., Subramanian, K., and Z. Zhang, "Ingress
              Replication Tunnels in Multicast VPN", RFC 7988,
              DOI 10.17487/RFC7988, October 2016,
              <https://www.rfc-editor.org/info/rfc7988>.

   [RFC8279]  Wijnands, IJ., Ed., Rosen, E., Ed., Dolganow, A.,
              Przygienda, T., and S. Aldrin, "Multicast Using Bit Index
              Explicit Replication (BIER)", RFC 8279,
              DOI 10.17487/RFC8279, November 2017,
              <https://www.rfc-editor.org/info/rfc8279>.

   [RFC8556]  Rosen, E., Ed., Sivakumar, M., Przygienda, T., Aldrin, S.,
              and A. Dolganow, "Multicast VPN Using Bit Index Explicit
              Replication (BIER)", RFC 8556, DOI 10.17487/RFC8556, April
              2019, <https://www.rfc-editor.org/info/rfc8556>.

Authors' Addresses

   Zhaohui Zhang
   Juniper Networks
   Email: zzhang@juniper.net


   Rishabh Parekh
   Cisco



Zhang, et al.             Expires 27 April 2023                [Page 15]

Internet-Draft              Multicast Scaling               October 2022


   Email: riparekh@cisco.com


   Hooman Bidgoli
   Nokia
   Email: hooman.bidgoli@nokia.com


   Zheng Zhang
   ZTE
   Email: zhang.zheng@zte.com.cn








































Zhang, et al.             Expires 27 April 2023                [Page 16]