Network functions virtualization (NFV) is the process of decoupling network functions from hardware and running them on a software platform. It is a different although could be a complementary approach to software-defined networking (SDN) for network management. While both manage networks, they rely on different methods. While SDN separates the control and forwarding planes to offer a centralized view of the network, NFV primarily focuses on optimizing the network services themselves. It is a key part of cloud computing, data center networking, SD-WAN, and many others.
NFV began when service providers attempted to speed up deployment of new network services in order to advance their revenue and growth plans. Developers found that proprietary hardware-based appliances limited their ability to achieve these goals. They looked to standard IT virtualization technologies and found that virtualizing network functions accelerated service dynamics and provisioning.
HyperCX NFV is designed to easily provide VNFs (Virtual Network Functions) to HyperCX cloud users. It is not designed to be a competitor of advanced, carrier-grade NFV architectures like MANO, or be feature-rich like software firewalls like PfSense. Instead, simplicity and easy of use are key design goals while maintaining a good performance for small and medium cloud deployments.
HyperCX NFV has been pre-loaded with the following features:
HA support using VRRP protocol through keepalived.
Routing among connected Virtual Networks.
VPN support leveraging OpenVPN.
Support for masquerade on private networks.
L4 and L7 load balancing features.
Site to Site (S2S) VPN support leveraging StronSwang.
All of these features can be used at the same time on a single instance, or only one per HyperCX NFV instance. Next it will be shown how to deploy a basic NFV appliance and after that each feature will be described.
NOTE: HyperCX NFV is deployed and managed like a Virtual Router. In this document, the term Virtual Router or vRouter can also be used to refer to HyperCX NFV instances.
Obtaining the NFV appliance
NFV appliances are available since HyperCX 10.5.2 and can easily be obtained on the marketplace. The appliance name on the marketplace is HyperCX_NFV and the version can be seen on the details. More information on how to download it can be found here.
Upgrading the NFV appliance
The version of the HyperCX NFV appliance that is being used on the cluster can be seen when login in to any VM instance via SSH or VNC. To check the version of the appliance in the marketplace, simply open the appliance under Apps tab and check the details.
If the market's version is newer, you might want to upgrade your current version. In order to do this, simply download the new appliance from the market and destroy your old local appliance (template + images). Note that you will not be able to destroy a template if VMs are running from it.
Deploying an NFV appliance
Easy of use is one of the core principles around HyperCX NFV. An instance can be created in standalone mode (no support for HA) or in HA mode. The chosen deployment strategy should be based on how critical these services will be. Note that HyperCX public clouds and HyperCX Bento clusters use redundancy on all the major components, providing full HA infrastructures, but a server failure (which is an isolated event since every hardware component is proactively monitored) will bring a downtime of 6-8 minutes to all the instances that were running on that server. This is the time that will take the system to detect the failure, validate it is not a false positive and migrate all the instances.
HyperCX NFV appliances are deployed and managed from Instances --> Virtual Routers.
Deploying in standalone mode
To deploy a single, non-HA, instance of HyperCX NFV create a new Virtual Router appliance.
Leave in blank the Keepalive parameters since they only make sense for HA deployments.
Add all the networks that will require service from HyperCX NFV. Note that these networks will be routed among them. This appliance, like every HyperCX virtual instance, will use the gateway from the first NIC (or vNET) defined. This is very important when configuring the network interfaces. For example, if you intend that the internal (private) networks are routed to the internet (most common case) the first Virtual Network should use a public virtual network defined with a public IP and a gateway.
Select your HyperCX NFV template and set 1 under Number of VM instances.
Deploying in HA mode
For HA deployment a few parameters can be leveraged, which made no sense on standalone instances:
- Keepalive service ID: Defines Keepalived virtual_router_id parameter. This parameter will be the same among all the VM instances of the appliance. If left blank, the vRouter ID will be used (recommended).
- Keepalive password: Defines a common password for all the instances to communicate via VRRP. If left blank, no password will be used and any other Keepalive instance will be able to join the cluster without authentication.
When selecting the networks, a floating IP will be needed on each NIC that will leverage HA. In order to do this, insert the IP under Force IPv4 (or v6) and click Floating IP. This setting will provide an IP obtained from HyperCX's IPAM to each instance, and the defined IP will also be used on the master keepalived instance. A floating IP can be used on each different virtual network, but depending on the use case, not all the virtual networks actually need a floating IP. This will be covered in more detail later on.
If Floating IP is not selected, several instances will be deployed with a single IP on the specific NIC, no floating IP will be configured on that NIC so Active-Passive HA will not be supported on the selected interface. If no IP is specified but Floating IP option is checked, the floating IP will be automatically assigned from HyperCX internal IPAM.
Under Number of VM instances, select 2 or more. No more than 2 instances are needed or recommended.
After the virtual router is deployed, you can get information like the version number and all the enabled VNFs and their configuration by logging in to any VM instance. A message similar to this will appear:
_ _ ______ __ _ _ _______ __ | | | |_ _ _ __ ___ _ __ / ___\ \/ / | \ | | ___\ \ / / | |_| | | | | '_ \ / _ \ '__| | \ / | \| | |_ \ \ / / | _ | |_| | |_) | __/ | | |___ / \ | |\ | _| \ V / |_| |_|\__, | .__/ \___|_| \____/_/\_\ |_| \_|_| \_/ |___/|_| VERSION: 1.0 =========================VRRP======================== Interface ETH0 using floating IP: Interface ETH1 using floating IP: 10.10.0.180 Keepalived ENABLED =========================MASQUERADE======================== IPV4 MASQUERADE ENABLED VIA ETH0 IPV6 MASQUERADE DISABLED =========================LOAD_BALANCER======================== LOAD BALANCER ENABLED. Load Balancer backend 1: 10.10.0.123 Load Balancer backend 2: 10.10.0.124 Load Balancer backend 3: 10.10.0.125 Load Balancer backend 4: 10.10.0.32 -------------------------------------------------- LOAD BALANCER AUTHENTICATION ENABLED Load Balancer monitoring portal URL: http://server_address:8989/stats Load Balancer monitoring portal user: admin Load Balancer monitoring portal password: password -------------------------------------------------- =========================OPENVPN======================== OpenVPN ENABLED. Created OpenVPN account for user user1 Created OpenVPN account for user user2 Detected internal network: 10.10.0.21 255.255.0.0 Detected internal network: 192.168.80.0 255.255.255.0 Last login: Tue Feb 4 23:45:14 2020 from 126.96.36.199
Virtual Network Functions
Next, every VNF supported by HyperCX NFV will be explained.
This VNF provide Internet access to the internal Virtual Network. If the first Virtual Network defined on the virtual router uses a public IP, masquerade will be automatically enabled through this interface. No extra configuration is needed. HyperCX leverages iptables for the masquerade VNF.
Masquerade VNF will require that every VM uses the vRouter's IP as gateway. For HA deployments, a floating IP must be used on each internal network, but it is optional for the public network. The VMs inside the virtual networks will use the active vRouter's public IP as source IP for the masquerade.
L4 Load Balancing
HyperCX uses HAProxy to provide load balancing features. This is used to balance all kind of applications, except thoses that use https, since it is not supported to handle the certificate. If this VNF is used, HAProxy will be configured with the following considerations:
HAProxy will work in reverse-proxy mode, the backend servers see its IP address as their client address. This is easy to configure and manage since the backends will not need to configure the load balancer as gateway, and the client will not get a response from a different IP to which it made the request. This is sometimes annoying when the client's IP address is expected in server logs, so some monitoring features or sessions based on source IPs would be limited.
To overcome the previous limitation, the well-known HTTP header "X-Forwarded-For" is added to all requests sent to the server. This header contains a value representing the client's IP address. Since this header is always appended at the end of the existing header list, the server must be configured to always use the last occurrence of this header only. This is something easy to implement on the backend application.
Client requests will be maintained to a single server. This is useful for web applications that rely on sessions.
Backends are monitored via tcp-checks. This means that, as long as the port is open and listening for connections, the backend will be marked as healthy and will receive requests.
Load Balancing VNF does not require that the backend VMs use the vRouter's IP as gateway.
There are two main strategies regarding the virtual networks that will be used:
The vRouter will use at least two virtual networks. The frontend will listen into one while the backends are located on the second one. A good case scenario is to use a public IP on the first vNet and any IP on a private network where the backends are located. This will allow to publish a web app to the Internet without the need of an extra firewall performing NAT features.
The vrouter will use a single virtual network. On this configuration, the frontend listens on the same virtual network where the backends are located. This is easier to setup and a common scenario could be a frontend located on Network X that connects to a few API servers located on the same network via the load balancer.
For HA deployments, a floating IP must be used on the frontend network and requests must be directed to the floating IP only. Configuring a floating IP on the backend's network is not necessary.
Only two parameters are mandatory to enable the load balancing VNF. During instantiation, under Custom attributes. If both of the following attributes are filled, the load balancer will be enabled.
IPs of the backend servers separated by spaces: Set the IP of every backend server you intend to use separated by an space. Ex: 10.10.0.123 10.10.0.124 10.10.0.125
Listening port for frontend and backends: Set the port that will be used by the backends. Only one port is supported at this time. This port will also be used by the frontend to listen for incoming requests.
HAProxy supports a monitoring web page that shows information of all the backends. This feature is disabled by default. In order to enable it, these two parameters must be filled:
Username for the load balancer's management portal
Password for the load balancer's management portal
After inserting the previous information, the management portal can be reached using the following URL:
HyperCX NFV provides VPNs based on OpenVPN. Some considerations on this VPN are as follows:
- CA certificate is the same on all HyperCX NFV appliances. User's authentication is only performed via password via a custom authentication plugin developed by Virtalus.
VPN VNF will require that every VM uses the vRouter's IP as gateway. For HA deployments, a floating IP must be used on the public network and each internal network. This floating IP is the one that must be defined on the .ovpn configuration file.
In order to enable the VPN VNF, a public network with a public IP must be set up as the first Virtual Network (just like the masquerade VNF). Besides of this, a configuration parameter is mandatory:
- User and password for each VPN user: This will configure all the users that will be able to use the VPN and their credentials. Each user will be configured with it's password separated by a colon in the format: "username:password". Several user/password pairs can be inserted separated by spaces. Ex: user1:pass1 user2:pass2 user3:pass3
After the appliance is running, you can get the .ovpn configuration files under /root/vpn_config_files on any instance. Alternatively, you can copy the following configuration file:
dev tun cipher AES-256-CBC persist-key persist-tun status openvpn-status.log verb 3 explicit-exit-notify 1 client auth-user-pass auth-nocache remote $PUBLIC_IPV4_ADDRESS 1194 udp <ca> -----BEGIN CERTIFICATE----- MIIDQjCCAiqgAwIBAgIUKHe8HaisIp9iK7BoWrF3PrWET7MwDQYJKoZIhvcNAQEL BQAwEzERMA8GA1UEAwwIQ2hhbmdlTWUwHhcNMjAwMjAzMDUxNTI1WhcNMzAwMTMx MDUxNTI1WjATMREwDwYDVQQDDAhDaGFuZ2VNZTCCASIwDQYJKoZIhvcNAQEBBQAD ggEPADCCAQoCggEBANZWy4vULsjfqsbUuj+S7oJ51EO1rhBC18cIa4ehgJegUGH7 Sz3BOE9uDVBy585pIwdta3KAkxV6rgdxEo23CfoUn+ibHoZtri3FUZJ+Rur5vyNF jGj1AI8GDbAW7K11rhSvfLneNY+Ia/6/uJG+Wa28zmM1scC9u37PrWpFOXIfxCdY tGe9sfkVx4yRbKNsNAZb14D6HEHqoK0F0dMqFnJeh6cERI6X77eq2QwPKqXzmqPg PClkdohJnQ7Gg5Ac4LDUx8Zk/QUXWT/yPXC1NgAqVMGBiiXhG5nfgvef2W+WX7og iFeU+RKXS1eJBru72mzEl2u5UJZP1VsAzsVmiykCAwEAAaOBjTCBijAdBgNVHQ4E FgQUlq6eCqIWaGAEIrmbv+orFDm5hKMwTgYDVR0jBEcwRYAUlq6eCqIWaGAEIrmb v+orFDm5hKOhF6QVMBMxETAPBgNVBAMMCENoYW5nZU1lghQod7wdqKwin2IrsGha sXc+tYRPszAMBgNVHRMEBTADAQH/MAsGA1UdDwQEAwIBBjANBgkqhkiG9w0BAQsF AAOCAQEAUQI/iQcCCHSkiQD9qUGozgsoHXMPPkm5xeDnjVXny+Ei3lCgrUJC/oqG xPgja3WTODuOzUhFphRZQKqVxRgjVWRf8dOTJx3+KLVseYumFQ11yaZ/MBoTicMz sjk9p8DqIrfU8x9cy720x7NQCaHvQlAushxnefXuzlkqWyeXmBJnPOM9hcxJc5/G kPruTlKWmSBBg7qYEsy6CifBsVbN+6gZHwsgmkhyxQ/j7/h8t2gva4d11XWXtVWj LyA0HCZw6iQSIowFnwWPFv+WckV4WWdC/ERlR95nBsPWCHzOsKtnkpEJjOHiBWh1 LzKPHae4zKnB4sNyRj/hEa5Xa/nNwg== -----END CERTIFICATE----- </ca>
Modify the public IPv4 address.
There are two basic options to harden the VPN security. In both cases, this requires that the user generates it's own PKI and copies some files to every vRouter instance.
Generating a PKI
This is a simple example to build a PKI. This does not need to be executed on any virtual router instance (although it is an option). The goal is to generate the required files and save them.
Download and decompress easy-rsa from this link.
./easyrsa --pki-dir=/etc/openvpn/pki --batch init-pki ./easyrsa --pki-dir=/etc/openvpn/pki --batch build-ca nopass ./easyrsa --pki-dir=/etc/openvpn/pki --batch gen-req HyperCX nopass ./easyrsa --pki-dir=/etc/openvpn/pki --batch sign-req server HyperCX openssl dhparam -out /etc/openvpn/dh2048.pem 2048 openvpn --genkey --secret /etc/openvpn/ta.key
Custom CA certificate
The easiest option would be to replace the CA certificate while keep relying on the user/password authentication. This will maintain all the server configurations and only the new CA certificate is required. Replace the server's /etc/openvpn/pki/ca.crt with the newly generated certificate. Client's configuration will also need to modify the included certificate.
Custom PKI and certificate based authentication
This is the recommended option, and requires the entire pki to be replaced. Also, the client's keys will need to be generated:
./easyrsa --pki-dir=/etc/openvpn/pki --batch gen-req client nopass ./easyrsa --pki-dir=/etc/openvpn/pki --batch sign-req client client
A new OpenVPN server and client configuration file are also required.
HyperCX NFV provides StrongSwan based Site-to-Site (S2S) VPNs. This VPN comes preconfigured and only needs a few parameters to be configured.
- All the preconfigured parameters on the appliances must be the same as in the remote site for the VPN to work.
- For the VPN to be available, three parameters are required: Remote Site Public IP, Remote Site Networks, Pre-shared Key
- Each VM will be required to use the vRouter's IP as gateway.
- For HA deployments, a floating IP must be used on the public network and on each internal network.
Since HyperCX NFV is based on its simplicity, there are some VPN parameters that are preconfigured as follows:
- Key Exchange version: Auto
- Authentication method: Mutual PSK
- Encryption algorithm: aes256
- Hash: sha256
- DH Group: 2(1024)
- Lifetime (Seconds): 28800
- Protocol: ESP
- Encryption algorithm: aes256
- Hash Algorithm: sha256
- PFS key group: 2(1024)
- Lifetime: 3600
To enable S2S VPN, a public network with a public IP must be configured as the first Virtual Network (just like the OpenVPN VNF). In addition, three parameters are mandatory:
- Remote Site Public IP: This will be used as the edge point at the remote site.
- Remote Site Networks: These are the networks from the remote site that will be able to access the local private networks and vice versa. Multiple networks saparated by spaces can be inserted. Ex: 10.10.10.0/24 10.10.20.0/24
- Pre-shared Key: This will be used to authenticate the VPN on each side. Must be the same on both sites.
There is another (optional) parameter that could be configured to monitor the status of the VPN, the Auto Ping Host. This option is for the Local Site to monitor the other edge, it must be a host at the Remote Site with the ping allowed.
L7 Load Balancer
HyperCX NFV provides Layer 7 balancing capabilities based on NGINX. This is used to balance websites behind an https certificate. Here, NGINX will be configured with the following considerations:
- NGINX will work as reverse proxy, the backend servers see its IP as their client address. This is easy to configure and manage since the backends will not need to configure the load balancer as gateway, and the client will not get a response from a different IP to which it made the request. It is very similiar to the other Load Balancer that HyperCX NFV provide.
- Client requests will be maintained to a single server.
- The SSL certificate will be maintained to a single server, the HyperCX NFV appliance. This avoids to keep the certificate in all the backends.
- Backends are monitored. This means that, as long as the backends is listening for connections, it will be marked as healthy and will receive requests.
L7 Load Balancer does not require that the backends use the vRouter's IP as gateway. The considerations here are similar to the previously mentioned L4 Load Balancing VNF.
These VNF will automatically configure an SSL certificate for the HTTPS endpoint based on letsencrypt. For this to work, a DNS entry must be configured on your DNS servers pointing Site URL to the public IP these vrouter will have. You can force a specific IP from the address pool on the NIC options.
Only two parameters are mandatory to enable the load balancer. During instantiation, under Custom attributes. If both of the following attributes are filled, the load balancer will be enabled.
- IPs of the backends separated by spaces: Set the IP of every backend server you intend to use separated by an space. Ex: 10.10.0.123 10.10.0.124 10.10.0.125
- Site url: Set the url of the website you intend to use. Ex: demo.example.com.
All http requests will be automatically redirected to https.
L7 Load Balancer is not compatible with HyperCX NFV HA. Works only with standalone deployments.
After the virtual router based on the HyperCX NFV template is deployed, any configuration from any previous VNF must be configured directly on the VM instances. First, you need to identify all the VM instances that belong to the virtual router instance. This can be seen by selecting the virtual router and, inside the virtual router instance, clicking on the VMs tab.
To change the config in each VM it is necessary to go into the Conf tab from the VM and click Update Configurations. The required configurations can be found under Context --> Custom vars.