NVO3 working groupInternet Engineering Task Force (IETF) A. GhanwaniInternet DraftRequest for Comments: 8293 DellIntended status:Category: Informational L. DunbarExpires: November 8, 2018ISSN: 2070-1721 M. McBride Huawei V. Bannai Google R. Krishnan DellOctober 23,December 2017 A Framework for Multicast in Network VirtualizationOverlays draft-ietf-nvo3-mcast-framework-11 Status of this Memoover Layer 3 Abstract ThisInternet-Draft is submitteddocument provides a framework for supporting multicast traffic infull conformance witha network that uses Network Virtualization over Layer 3 (NVO3). Both infrastructure multicast and application-specific multicast are discussed. It describes theprovisions of BCP 78various mechanisms that can be used for delivering such traffic as well as the data plane andBCP 79. This Internet-Draft is submitted in full conformance withcontrol plane considerations for each of theprovisionsmechanisms. Status ofBCP 78 and BCP 79.This Memo This documentmay not be modified, and derivative works of it mayis notbe created, except to publish it asanRFC and to translateInternet Standards Track specification; itinto languages other than English. Internet-Drafts are working documentsis published for informational purposes. This document is a product of the Internet Engineering Task Force(IETF), its areas, and its working groups. Note that other groups may also distribute working documents as Internet- Drafts. Internet-Drafts are draft documents valid for a maximum(IETF). It represents the consensus ofsix monthsthe IETF community. It has received public review andmay be updated, replaced, or obsoletedhas been approved for publication byotherthe Internet Engineering Steering Group (IESG). Not all documentsatapproved by the IESG are a candidate for anytime. It is inappropriate to use Internet-Drafts as reference material or to cite them other than as "work in progress." The listlevel of Internet Standard; see Section 2 of RFC 7841. Information about the currentInternet-Drafts can be accessed at http://www.ietf.org/ietf/1id-abstracts.txt The liststatus ofInternet-Draft Shadow Directories canthis document, any errata, and how to provide feedback on it may beaccessedobtained athttp://www.ietf.org/shadow.html This Internet-Draft will expire on November 8, 2016.https://www.rfc-editor.org/info/rfc8293. Copyright Notice Copyright (c) 2017 IETF Trust and the persons identified as the document authors. All rights reserved. This document is subject to BCP 78 and the IETF Trust's Legal Provisions Relating to IETF Documents(http://trustee.ietf.org/license-info)(https://trustee.ietf.org/license-info) in effect on the date of publication of this document. Please review these documents carefully, as they describe your rights and restrictions with respect to this document. Code Components extracted from this document must include Simplified BSD License text as described in Section 4.e of the Trust Legal Provisions and are provided without warranty as described in the Simplified BSD License.Abstract This document provides a framework of supporting multicast traffic in a network that uses Network Virtualization Overlays (NVO3). Both infrastructure multicast and application-specific multicast are discussed. It describes the various mechanisms that can be used for delivering such traffic as well as the data plane and control plane considerations for each of the mechanisms.Table of Contents 1.Introduction...................................................3Introduction . . . . . . . . . . . . . . . . . . . . . . . . 2 1.1. Infrastructuremulticast..................................3Multicast . . . . . . . . . . . . . . . . 3 1.2.Application-specific multicast............................4 1.3. Terminology clarification.................................4Application-Specific Multicast . . . . . . . . . . . . . 3 2.Acronyms.......................................................4Terminology and Abbreviations . . . . . . . . . . . . . . . . 3 3. MulticastmechanismsMechanisms innetworks that use NVO3.................5Networks That Use NVO3 . . . . . . . 4 3.1. Nomulticast support......................................6Multicast Support . . . . . . . . . . . . . . . . . . 5 3.2. Replication at thesource NVE.............................7Source NVE . . . . . . . . . . . . . . 6 3.3. Replication at amulticast service node...................9Multicast Service Node . . . . . . . . . 8 3.4. IPmulticastMulticast in theunderlay.............................10Underlay . . . . . . . . . . . . . . 9 3.5. Otherschemes............................................12Schemes . . . . . . . . . . . . . . . . . . . . . . 11 4. SimultaneoususeUse ofmore than one mechanism...................12More Than One Mechanism . . . . . . . . . 11 5. Otherissues..................................................13Issues . . . . . . . . . . . . . . . . . . . . . . . . 11 5.1.Multicast-agnostic NVEs..................................13Multicast-Agnostic NVEs . . . . . . . . . . . . . . . . . 11 5.2. Multicastmembership managementMembership Management for DC withVMs..........13VMs . . . . . 12 6.Summary.......................................................14 7.SecurityConsiderations.......................................14 8.Considerations . . . . . . . . . . . . . . . . . . . 12 7. IANAConsiderations...........................................14Considerations . . . . . . . . . . . . . . . . . . . . . 12 8. Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . 12 9.References....................................................14References . . . . . . . . . . . . . . . . . . . . . . . . . 12 9.1. NormativeReferences.....................................14References . . . . . . . . . . . . . . . . . . 12 9.2. InformativeReferences...................................15 10. Acknowledgments..............................................16References . . . . . . . . . . . . . . . . . 13 Acknowledgments . . . . . . . . . . . . . . . . . . . . . . . . . 15 Authors' Addresses . . . . . . . . . . . . . . . . . . . . . . . 15 1. Introduction Networkvirtualization using OverlaysVirtualization over Layer 3(NVO3)[RFC7365](NVO3) [RFC7365] is a technology that is used to address issues that arise in building large,multitenantmulti- tenant data centers (DCs) that make extensive use of server virtualization [RFC7364]. This document provides a framework for supporting multicasttraffic,traffic in a network that usesNetwork Virtualization using Overlays over Layer 3 (NVO3).NVO3. Both infrastructure multicast andapplication- specificapplication-specific multicast are considered. It describesthevariousmechanismsmechanisms, and the considerations of each of them, that can be used for delivering such traffic in networks that use NVO3. The reader is assumed to be familiar with the terminology and concepts as defined in the NVO3 Frameworkdocument[RFC7365] and NVO3 Architecturedocument [RFC8014].[RFC8014] documents. 1.1. InfrastructuremulticastMulticast Infrastructure multicastis a capability needed byrefers to networkingservices,services that require multicast or broadcast delivery, such as Address Resolution Protocol (ARP), Neighbor Discovery (ND), Dynamic Host Configuration Protocol (DHCP), multicast Domain Name Server (mDNS),etc. RFC3819 Sectionetc., some of which are described in Sections 5 and 6have detailed description for someofthe infrastructure multicastRFC 3819 [RFC3819]. It is possible to provide solutions for these services that do not involve multicast in the underlay network.InFor example, in the case of ARP/ND, anetwork virtualization authorityNetwork Virtualization Authority (NVA) can be used for distributing themappings ofIP address toMACMedia Access Control (MAC) address mappings to allnetwork virtualization edgesof the Network Virtualization Edges (NVEs).The NVEsAn NVE can then trap ARPRequest/NDRequest and/or ND Neighbor Solicitation messages from theTSs (Tenant System)Tenant Systems (TSs) that are attached to it and respond to them, thereby eliminating the needtofor the broadcast/multicast of such messages. In the case of DHCP, the NVE can be configured to forward these messages usinga helper function.the DHCP relay function [RFC2131]. Ofcoursecourse, it is possible to support all of these infrastructure multicast protocols natively if the underlay provides multicast transport. However, even in the presence of multicast transport, it may be beneficial to use the optimizations mentioned above to reduce the amount of such traffic in the network. 1.2. Application-Specific Multicast Application-specific multicastApplication-specifictraffic refers to multicast trafficare originatedthat originates and is consumed by user applications.TheSeveral such applications are described elsewhere [DC-MC]. Application-specificmulticast, which canmulticast may be either Source-Specific Multicast (SSM) or Any-Source Multicast(ASM)[RFC3569],(ASM) [RFC3569] and has the following characteristics: 1. Receiver hosts are expected to subscribe to multicast content using protocols such as IGMP [RFC3376] (IPv4) orMLDMulticast Listener Discovery (MLD) [RFC2710] (IPv6). Multicast sources and listenersparticipantparticipate in these protocols using addresses that are in theTenant SystemTS address domain. 2. Thelistset of multicast listeners for each multicast groupismay not be known in advance. Therefore, it may not be possible or practical for an NVA to get the list of participants for each multicast group ahead of time.1.3. Terminology clarification2.Acronyms &Terminology and Abbreviations In this document, the terms host,tenant system (TS)Tenant System (TS), andvirtual machineVirtual Machine (VM) are used interchangeably to represent an end station that originates or consumes data packets. ASM: Any-Source Multicast IGMP: Internet Group Management Protocol LISP: Locator/ID Separation Protocol MSN: Multicast Service Node RLOC: Routing Locator NVA: Network Virtualization Authority NVE: Network Virtualization Edge NVGRE: Network Virtualization using GRE PIM: Protocol-Independent Multicast SSM: Source-Specific Multicast TS: TenantsystemSystem VM: Virtual Machine VN: Virtual Network VTEP:VxLANVXLAN Tunnel EndPointsPoint VXLAN: Virtual eXtensible LAN 3. MulticastmechanismsMechanisms innetworks that useNetworks That Use NVO3 In NVO3 environments, traffic between NVEs is transported using an encapsulation such asVirtual eXtensible Local Area Network (VXLAN) [RFC7348,VXLAN-GPE],VXLAN [RFC7348] [VXLAN-GPE], Network VirtualizationUsingusing Generic Routing Encapsulation (NVGRE) [RFC7637], Geneve [Geneve], Generic UDP Encapsulation(GUE)[GUE], etc. What makes networks using NVO3 different fromanyothernetworknetworks is that some NVEs, especiallythe NVENVEs implementedon server,in servers, might not supportPIM or other nativeregular multicastmechanisms. They might just encapsulateprotocols such as PIM. Instead, the only capability they may support would be that of encapsulating data packets from VMs with an outer unicast header. Therefore, it is important for networks using NVO3 to have mechanisms to support multicast as a network capability for NVEs, to map multicast traffic from VMs (users/applications) to an equivalent multicast capability inside the NVE, or to figure out the outer destination address if NVE does not support native multicast(e.g.(e.g., PIM) or IGMP.Besides the need to support ARP and ND, there are several applications that require the support of multicast and/or broadcast in data centers [DC-MC].With NVO3, there are many possible ways that multicast may be handled in such networks. We discuss some of the attributes of the following four methods: 1. No multicastsupport.support 2. Replication at the sourceNVE.NVE 3. Replication at a multicast servicenode.node 4. IP multicast in theunderlay.underlay These methods are briefly mentioned in the NVO3 Framework [RFC7365] and NVO3architectureArchitecture [RFC8014]document.documents. This document provides more details about the basic mechanisms underlying each of these methods and discusses the issues and trade-offs of each. We note that other methods are also possible, such as [EDGE-REP], but we focus on the above four because they are the most common. It is worth noting that when selecting a multicastreplication strategy,mechanism, it is useful to consider theinteraction withimpact of these on any multicast congestion control mechanisms that applications may be using to obtain the desired system dynamics. In addition,for multicast we followthe same rules forECNExplicit Congestion Notification (ECN) would apply to multicast traffic being encapsulated, asany non-multicastfor unicast trafficwould and be in conformance with the appropriate encap draft [RFC6040][RFC6040]. 3.1. Nomulticast supportMulticast Support In this scenario, there is no support whatsoever for multicast traffic when using the overlay. This method can only work if the following conditions are met: 1. All of the application traffic in the network is unicasttraffictraffic, and the only multicast/broadcast traffic is from ARP/ND protocols. 2. An NVA is used by all of the NVEs to determine the mapping of a givenTenant System's (TS's) MAC/IPTS's MAC and IP address toits NVE.the NVE that it is attached to. In other words, there is nodata planedata-plane learning. Address resolution requests via ARP/ND that are issued by the TSs must be resolved by the NVE that they are attached to. With this approach, it is not possible to support application- specific multicast. However, certain multicast/broadcast applicationssuch as DHCPcan be supported without multicast; for example, DHCP, which can be supported by use ofa helperDHCP relay functionin the NVE.[RFC2131]. The main drawback of this approach, even for unicast traffic, is that it is not possible to initiate communication with a TS for which a mapping to an NVE does not already existinat the NVA. This is a problem in the case where the NVE is implemented in a physical switch and the TS is a physical end station that has not registered with the NVA. 3.2. Replication at thesourceSource NVE With this method, the overlay attempts to provide a multicast service without requiring any specific support from the underlay, other than that of a unicast service. A multicast or broadcast transmission is achieved by replicating the packet at the sourceNVE,NVE and making copies, one for each destination NVE that the multicast packet must be sent to. For this mechanism to work, the source NVE must know, a priori, the IP addresses of all destination NVEs that need to receive the packet. For the purpose of ARP/ND, this would involve knowing the IP addresses of all the NVEs that have TSs in thevirtual network (VN)VN of the TS that generated the request. For the support of application-specific multicast traffic, a method similar to that of receiver-sites registration for a particular multicastgroupgroup, described in[LISP-Signal-Free][LISP-Signal-Free], can be used. The registrations from differentreceiver-sitesreceiver sites can be merged at the NVA, which can construct a multicastreplication-listreplication list inclusive of all NVEs to which receivers for a particular multicast group are attached. Thereplication-listreplication list for each specific multicast group is maintained by the NVA.Note: Using LISP-signal-freeNote that using receiver-sites registration does not necessarily mean thehead-end (i.e. NVE)source NVE must do replication. If themapping database (i.e. NVA)NVA indicates that multicast packets are encapsulated to multicastRLOCs,service nodes, then thereiswould be no replicationhappeningat the NVE. The receiver-sites registration is achieved by egress NVEs performingtheIGMP/MLD snooping to maintain the state for which attached TSs have subscribed to a given IP multicast group. When the members of a multicast group are outside the NVO3 domain, it is necessary for NVO3 gateways to keep track of the remote members of each multicast group. The NVEs and NVO3 gateways then communicate with the multicast groups that are of interest to the NVA. If the membership is not communicated to the NVA, and if it is necessary to preventhostsTSs attached to an NVE that have not subscribed to a multicast group from receiving the multicast traffic, the NVE would need to maintain multicast group membership information. In the absence of IGMP/MLD snooping, the traffic would be delivered to all TSs that are part of the VN. Inmulti-homingmultihoming environments, i.e., in those where a TS is attached to more than one NVE, the NVA would be expected to provide information to all of the NVEs under its control about all of the NVEs to which such a TS is attached. The ingress NVE can then choose any one of those NVEs as the egressNVEsNVE for the data frames destined towards the multi-homed TS. This method requires multiple copies of the same packet to all NVEs that participate in the VN. If, for example, a tenant subnet is spread across 50 NVEs, the packet would have to be replicated 50 times at the source NVE. Obviously, this approach creates more traffic to the network that can cause congestion when the network load is high. This also creates an issue with the forwarding performance of the NVE. Note that this method is similar to what was used in Virtual Private LAN Service (VPLS) [RFC4762] prior to support ofMulti-ProtocolMultiprotocol Label Switching (MPLS) multicast [RFC7117]. While there are some similarities between MPLS Virtual Private Network (VPN) and NVO3, there are some key differences:-o The attachment from Customer Edge (CE) to Provider Edge (PE)attachmentin VPNs is somewhat static, whereas in a DC that allows VMs to migrate anywhere, the TS attachment to NVE is much more dynamic.-o The number of PEs to which a single VPN customer is attached in an MPLS VPN environment is normally far less than the number of NVEs to which a VN's VMs are attached in a DC. When a VPN customer has multiple multicast groups, "Multicast VPN" [RFC6513] combines all those multicast groups within each VPN client to one single multicast group in the MPLS (or VPN) core. The result is that messages from any of the multicast groups belonging to one VPN customer will reach all the PE nodes of the client. In other words, any messages belonging to any multicast groups under customer X will reach all PEs of the customer X. When the customer X is attached to only a handful of PEs, the use of this approach does not result in an excessivewastagewaste of bandwidth in the provider's network. In a DC environment, a typicalserver/hypervisor basedhypervisor-based virtual switch may only support on the order of 10's of VMs (as of this writing). A subnet with N VMs may be, in the worst case, spread across NvSwitches.virtual switches (vSwitches). Using an "MPLS VPN multicast" approach in such a scenario would require the creation of aMulticastmulticast group in the core in order forthisthe VN to reach all N NVEs. If only a small percentage of this client's VMs participate inapplication specificapplication-specific multicast, a great number of NVEs will receive multicast traffic that is not forwarded to any of their attached VMs, resulting in a considerablewastagewaste of bandwidth. Therefore, theMulticastmulticast VPN solution may not scale in a DC environment with the dynamic attachment ofVirtual NetworksVNs to NVEs and a greater number of NVEs for eachvirtual network.VN. 3.3. Replication at amulticast service nodeMulticast Service Node With this method, all multicast packets would be sent using a unicast tunnel encapsulation from the ingress NVE to amulticast service nodeMulticast Service Node (MSN). The MSN, in turn, would create multiple copies of the packet and would deliver a copy, using a unicast tunnel encapsulation, to each of the NVEs that are part of the multicast group for which the packet is intended. This mechanism is similar to that used by the Asynchronous Transfer Mode (ATM) Forum's LAN Emulation (LANE) specification [LANE]. The MSN is similar to theRP (Rendezvous Point)Rendezvous Point (RP) inPIM SM,Protocol Independent Multicast - Sparse Mode (PIM-SM), but different in that the user data trafficareis carried by the NVO3 tunnels. The following arethepossible ways for the MSN to get the membership information for each multicast group:-o The MSN can obtain this membership information from the IGMP/MLD report messages sent by TSs in response to IGMP/MLD query messages from the MSN. The IGMP/MLD query messages are sent from the MSN to the NVEs, which then forward the query messages to TSs attached to them. An IGMP/MLD querymessagesmessage sent out by the MSN to an NVE is encapsulated with the MSN address in the outer IP source address field and the address of the NVE in the outer IP destination address field.TheAn encapsulated IGMP/MLD querymessagesmessage also has aVNID for avirtual network (VN) identifier (corresponding to the VN that the TSs belong to) in the outer header and a multicast address in the inner IP destination address field. Upon receiving the encapsulated IGMP/MLD query message, the NVE establishes a mapping for "MSN address"<->to "multicast address", decapsulates the received encapsulated IGMP/MLD message, andmulticastmulticasts the decapsulated query message to the TSs that belong to the VNunder theattached to that NVE.AAn IGMP/MLD report message sent by a TS includes the multicast address and the address of the TS. With the proper "MSNAddress" <-> "Multicast-Address"address" to "multicast address" mapping, the NVEs can encapsulate all multicast data framestocontaining the"Multicast-Address""multicast address" with the address of the MSN in the outer IP destination address field.-o The MSN can obtain the membership information from the NVEs that have the capability to establish multicast groups by snooping native IGMP/MLD messages(p.s.(note that the communication must be specific to the multicastaddresses),addresses) or by having the NVA obtain the information from theNVEs,NVEs and in turn have MSN communicate with the NVA. This approach requires additional protocol between MSN and NVEs. Unlike the method described in Section 3.2, there is no performance impact at the ingress NVE, nor are there any issues with multiple copies of the same packet from the source NVE to theMulticast Service Node.MSN. However, there remain issues with multiple copies of the same packet on links that are common to the paths from the MSN to each of the egress NVEs. Additional issues that are introduced with this method include the availability of the MSN, methods to scale the services offered by the MSN, and thesub-optimalitysuboptimality of the delivery paths. Finally, the IP address of the source NVE must be preserved in packet copies created at the multicast service node ifdata planedata-plane learning is in use. This could create problems if IP source addressreverse path forwardingReverse Path Forwarding (RPF) checks are in use. 3.4. IPmulticastMulticast in theunderlayUnderlay In this method, the underlay supports IP multicast and the ingress NVE encapsulates the packet with the appropriate IP multicast address in the tunnel encapsulation header for delivery to the desired set of NVEs. The protocol in the underlay could be any variant ofProtocol Independent Multicast (PIM),PIM, orprotocol dependenta protocol-dependent multicast, such as [ISIS-Multicast]. If an NVE connects to its attached TSs via a Layer 2 network, there are multiple ways for NVEs to support theapplication specificapplication-specific multicast:-o The NVE only supports the basic IGMP/MLD snooping function,letwhile theTSs routers handling"TS routers" handle theapplication specificapplication-specific multicast. This scheme doesn't utilize the underlay IP multicast protocols.-Instead routers, which are themselves TSs attached to the NVE, would handle multicast protocols for the application-specific multicast. We refer to such routers as TS routers. o The NVE can act as a pseudo multicast router for the directly attachedVMsTSs and supportproperthe mapping ofIGMP/MLD'sIGMP/MLD messages to the messages needed by the underlay IP multicast protocols. With this method, there are none of the issues with the methods described in Sections3.2.3.2 and 3.3 with respect to scaling and congestion. Instead, there are other issues described below. WithPIM Sparse Mode (PIM-SM),PIM-SM, the number of flows required would be (n*g), where n is the number of source NVEs that source packets for the group, and g is the number of groups. Bidirectional PIM(BIDIR- PIM)(BIDIR-PIM) would offer better scalability with the number of flows required being g. Unfortunately, many vendors still do not fully support BIDIR or have limitations on its implementation.RFC6831[RFC6831]has good descriptiondescribes the use ofusingSSM as an alternative toBIDIR ifBIDIR, provided that theVTEP/NVE devicesNVEs have a way to learn of each other's IPaddressaddresses so that theycouldcan join all of the SSMSPT'sShortest Path Trees (SPTs) to create/maintain an underlay SSM IPMulticastmulticast tunnel solution. In the absence of any additionalmechanism, e.g.mechanism (e.g., using an NVA for addressresolution,resolution), for optimal delivery, there would have to be a separate group for eachtenant,VN for infrastructure multicast plus a separate group for each application-specific multicast address(used for multicast applications)within a tenant.Additional considerations areAn additional consideration is that only the lower 23 bits of the IP address (regardless of whether IPv4 or IPv6 is in use) are mapped to the outer MAC address, and if there is equipment that prunes multicasts at Layer 2, there will be some aliasing. Finally, a mechanism to efficiently provision such addresses for each group would be required. There are additional optimizationswhichthat are possible, but they come with their own restrictions. For example, a set of tenants may be restricted to some subset ofNVEsNVEs, and they could all share the same outer IP multicast group address.This howeverThis, however, introduces a problem ofsub-optimalsuboptimal delivery (even if a particular tenant within the group of tenants doesn't have a presence on one of the NVEswhichthat another one does, the multicast packets would still be delivered to that NVE). It also introduces an additional network management burden to optimize which tenants should be part of the same tenant group (based on the NVEs they share), which somewhat dilutes the value proposition of NVO3which is to(to completely decouple the overlay and physical network design allowing complete freedom of placement of VMs anywhere within thedata center.DC). Multicast schemes such asBIER (BitBit Indexed ExplicitReplication) [BIER-ARCH]Replication (BIER) [RFC8279] may be able to provide optimizations by allowing the underlay network to provide optimum multicast delivery without requiring routers in the core of the network to maintain per- multicast group state. 3.5. OtherschemesSchemes There are still other mechanisms that may be used that attempt to combine some of the advantages of the above methods by offering multiple replication points, each with a limited degree of replication [EDGE-REP]. Such schemes offer a trade-off between the amount of replication at an intermediate node(e.g.(e.g., router) versus performing all of the replication at the source NVE or all of the replication at a multicast service node. 4. SimultaneoususeUse ofmore than one mechanismMore Than One Mechanism While the mechanisms discussed in the previous section have been discussed individually, it is possible for implementations to rely on more than one of these. For example, the method of Section 3.1 could be used for minimizing ARP/ND, while at the same time, multicast applications may be supported by one, or acombination of,combination, of the other methods. For small multicast groups, the methods of source NVE replication or the use of a multicast service node may be attractive, while for larger multicast groups, the use of multicast in the underlay may be preferable. 5. OtherissuesIssues 5.1.Multicast-agnosticMulticast-Agnostic NVEs Some hypervisor-based NVEs do not process or recognize IGMP/MLDframes; i.e.frames, i.e., those NVEs simply encapsulate the IGMP/MLD messages in the same way as they do for regular data frames. By default,TSsa TS router periodically sends IGMP/MLD query messages to all the hosts in the subnet to trigger the hosts that are interested in the multicast stream to send back IGMP/MLD reports. In order for the MSN to get the updated multicast group information, the MSN can also send the IGMP/MLD query message comprising aclient specificclient-specific multicastaddress,address encapsulated in an overlay header to all of the NVEs to which the TSs in the VN are attached. However, the MSN may not always be aware of theclient specificclient-specific multicast addresses. In order to perform multicast filtering, the MSN has to snoop the IGMP/MLD messages between TSs and their corresponding routers to maintain the multicast membership. In order for the MSN to snoop the IGMP/MLD messages between TSs and their router, the NVA needs to configure the NVE to send copies of the IGMP/MLD messages to the MSN in addition to the default behavior of sending them to the TSs' routers;e.g.e.g., the NVA has to inform the NVEs to encapsulate data frames withDAthe Destination Address (DA) being 224.0.0.2(destination address(DA of IGMP report) to the TSs' router and MSN. This process is similar to "Source Replication" described in Section 3.2, except the NVEs only replicate the message to the TSs' router and MSN. 5.2. Multicastmembership managementMembership Management for DC with VMs Fordata centersDCs with virtualized servers, VMs can be added,deleteddeleted, or moved very easily. When VMs are added,deleteddeleted, or moved, the NVEs to which the VMs are attached are changed. When a VM is deleted from an NVE or a new VM is added to an NVE, the VM management system should notify the MSN to send the IGMP/MLD query messages to the relevant NVEs (as described in Section3.3),3.3) so that the multicast membership can be updated promptly. Otherwise, if there are changes of VMs attachment toNVEs, withinNVEs (within the duration of the configured default time interval that the TSs routers use for IGMP/MLDqueries,queries), multicast data may not reach the VM(s) that moved. 6. Security Considerations This document does not introduce any new security considerations beyond what is described in the NVO3 Architecture document [RFC8014]. 7. IANA Considerations This document does not require any IANA actions. 8. Summary This document has identified various mechanisms for supportingapplication specificapplication-specific multicast in networks that use NVO3. It highlights the basics of each mechanism and some of the issues with them. As solutions are developed, the protocols would need to consider the use of thesemechanismsmechanisms, andco-existencecoexistence may be a consideration. It also highlights some of the requirements for supporting multicast applications in an NVO3 network.7. Security Considerations This draft does not introduce any new security considerations beyond what is described n NVO3 Architecture (RFC8014). 8. IANA Considerations This document requires no IANA actions. RFC Editor: Please remove this section before publication.9. References 9.1. Normative References [RFC3376]Cain B. et al.Cain, B., Deering, S., Kouvelas, I., Fenner, B., and A. Thyagarajan, "Internet Group Management Protocol, Version 3", RFC 3376, DOI 10.17487/RFC3376, October2002.2002, <https://www.rfc-editor.org/info/rfc3376>. [RFC6513] Rosen,E. et al.,E., Ed. and R. Aggarwal, Ed., "Multicast inMPLS/BGPMPLS/ BGP IP VPNs", RFC 6513, DOI 10.17487/RFC6513, February2012.2012, <https://www.rfc-editor.org/info/rfc6513>. [RFC7364] Narten,T. et al.,T., Ed., Gray, E., Ed., Black, D., Fang, L., Kreeger, L., and M. Napierala, "Problemstatement:Statement: Overlays fornetwork virtualization",Network Virtualization", RFC 7364, DOI 10.17487/RFC7364, October2014.2014, <https://www.rfc-editor.org/info/rfc7364>. [RFC7365] Lasserre,M. et al.,M., Balus, F., Morin, T., Bitar, N., and Y. Rekhter, "Framework fordata centerData Center (DC)network virtualization",Network Virtualization", RFC 7365, DOI 10.17487/RFC7365, October2014.2014, <https://www.rfc-editor.org/info/rfc7365>. [RFC8014]Narten,Black, D., Hudson, J., Kreeger, L., Lasserre, M., and T.et al.," AnNarten, "An Architecture forOverlay NetworksData-Center Network Virtualization over Layer 3 (NVO3)",RFC8014, Dec. 2016.RFC 8014, DOI 10.17487/RFC8014, December 2016, <https://www.rfc-editor.org/info/rfc8014>. 9.2. Informative References [RFC2131] Droms, R., "Dynamic Host Configuration Protocol", RFC 2131, DOI 10.17487/RFC2131, March 1997, <https://www.rfc-editor.org/info/rfc2131>. [RFC2710]S. Deering et al,Deering, S., Fenner, W., and B. Haberman, "Multicast Listener Discovery (MLD) for IPv6",Oct 1999.RFC 2710, DOI 10.17487/RFC2710, October 1999, <https://www.rfc-editor.org/info/rfc2710>. [RFC3569]S.Bhattacharyya, S., Ed., "An Overview of Source-Specific Multicast (SSM)", RFC 3569, DOI 10.17487/RFC3569, July2003.2003, <https://www.rfc-editor.org/info/rfc3569>. [RFC3819]P. Harn et al.,Karn, P., Ed., Bormann, C., Fairhurst, G., Grossman, D., Ludwig, R., Mahdavi, J., Montenegro, G., Touch, J., and L. Wood, "Advice for Internet Subnetwork Designers", BCP 89, RFC 3819, DOI 10.17487/RFC3819, July2004.2004, <https://www.rfc-editor.org/info/rfc3819>. [RFC4762] Lasserre, M., Ed. andKompella,V.(Eds.),Kompella, Ed., "Virtual Private LAN Service (VPLS)usingUsing Label Distribution Protocol (LDP)signaling,"Signaling", RFC 4762, DOI 10.17487/RFC4762, January2007.2007, <https://www.rfc-editor.org/info/rfc4762>. [RFC6040]B.Briscoe, B., "Tunnelling of Explicit Congestion Notification",Nov 2010.RFC 6040, DOI 10.17487/RFC6040, November 2010, <https://www.rfc-editor.org/info/rfc6040>. [RFC6831] Farinacci,D. et al.,D., Meyer, D., Zwiebel, J., and S. Venaas, "The Locator/IDSeperationSeparation Protocol (LISP) for Multicast Environments",Jan, 2013.RFC 6831, DOI 10.17487/RFC6831, January 2013, <https://www.rfc-editor.org/info/rfc6831>. [RFC7117] Aggarwal,R. et al.,R., Ed., Kamite, Y., Fang, L., Rekhter, Y., and C. Kodeboniya, "Multicast inVPLS,"Virtual Private LAN Service (VPLS)", RFC 7117, DOI 10.17487/RFC7117, February2014.2014, <https://www.rfc-editor.org/info/rfc7117>. [RFC7348] Mahalingam,M. et al., " VirtualM., Dutt, D., Duda, K., Agarwal, P., Kreeger, L., Sridhar, T., Bursell, M., and C. Wright, "Virtual eXtensible Local Area Network (VXLAN): A Framework for Overlaying Virtualized Layer 2 Networks over Layer 3 Networks", RFC 7348, DOI 10.17487/RFC7348, August2014. [RFC7365] M. Lasserre, et al. "Framework for Data Center (DC) Network Virtualization", Oct 2014.2014, <https://www.rfc-editor.org/info/rfc7348>. [RFC7637]Garg P.Garg, P., Ed. andWang,Y.(Eds.),Wang, Ed., "NVGRE: NetworkVvirtualization usingVirtualization Using Generic Routing Encapsulation", RFC 7637, DOI 10.17487/RFC7637, September2015. [BIER-ARCH]2015, <https://www.rfc-editor.org/info/rfc7637>. [RFC8279] Wijnands,IJ. (Ed.) et al.,IJ., Ed., Rosen, E., Ed., Dolganow, A., Przygienda, T., and S. Aldrin, "MulticastusingUsing Bit Index ExplicitReplication," <draft-ietf-bier-architecture-03>, January 2016.Replication (BIER)", RFC 8279, DOI 10.17487/RFC8279, November 2017, <https://www.rfc-editor.org/info/rfc8279>. [DC-MC] McBride, M. andLui, H.,H. Liu, "Multicast in thedata center overview," <draft-mcbride-armd-mcast-overview-02>, workData Center Overview", Work inprogress,Progress, draft-mcbride-armd-mcast- overview-02, July 2012. [EDGE-REP]MarquesMarques, P., Fang, L., Winkworth, D., Cai, Y., and P.et al.,Lapukhov, "Edge multicast replication for BGP IPVPNs," <draft-marques-l3vpn-mcast-edge-01>, workVPNs.", Work inprogress,Progress, draft-marques-l3vpn-mcast-edge-01, June 2012. [Geneve] Gross,J. andJ., Ganga,I. (Eds.),I., and T. Sridhar, "Geneve: Generic Network Virtualization Encapsulation",<draft-ietf-nvo3-geneve- 01>, workWork inprogress, January 2016.Progress, draft-ietf-nvo3-geneve-05, September 2017. [GUE] Herbert,T. et al.,T., Yong, L., and O. Zia, "Generic UDP Encapsulation",<draft- ietf-nvo3-gue-02>, workWork inprogress, December 2015.Progress, draft-ietf-intarea-gue- 04, May 2017. [ISIS-Multicast] Yong,L. et al., "ISISL., Weiguo, H., Eastlake, D., Qu, A., Hudson, J., and U. Chunduri, "IS-IS Protocol ExtensionforFor Building Distribution Trees",<draft-yong-isis-ext-4- distribution-tree-03>, workWork inprogress,Progress, draft-yong-isis- ext-4-distribution-tree-03, October 2014. [LANE]"LAN emulation over ATM," TheATM Forum, "LAN Emulation Over ATM: Version 1.0", ATM Forum Technical Committee, af-lane-0021.000, January 1995. [LISP-Signal-Free] Moreno, V. and D. Farinacci,D.,"Signal-Free LISP Multicast",<draft-ietf-lisp-signal-free-multicast-01>, workWork inprogress, April 2016.Progress, draft-ietf-lisp-signal-free-multicast- 07, November 2017. [VXLAN-GPE] Maino, F., Kreeger,L.L., andElzur,U.(Eds.),Elzur, "Generic Protocol Extension for VXLAN",<draft-ietf-nvo3-vxlan-gpe-02>, workWork inprogress, April 2016. 10.Progress, draft-ietf-nvo3- vxlan-gpe-05, October 2017. Acknowledgments Many thanks are due to Dino Farinacci, Erik Nordmark, Lucy Yong, Nicolas Bouliane, Saumya Dikshit, Joe Touch, Olufemi Komolafe, and MatthewBocci,Bocci for their valuable comments and suggestions.This document was prepared using 2-Word-v2.0.template.dot.Authors' Addresses Anoop Ghanwani Dell Email: anoop@alumni.duke.edu Linda Dunbar Huawei Technologies 5340 Legacy Drive, Suite 1750 Plano, TX75024, USA75024 United States of America Phone: (469) 277 5840 Email: ldunbar@huawei.com Mike McBride Huawei Technologies Email: mmcbride7@gmail.com Vinay Bannai Google Email: vbannai@gmail.com Ram Krishnan Dell Email: ramkri123@gmail.com