Skip to main content

Fabric Aggregate Manager Handlers

Project description

PyPI

AMHandlers

Aggregate Manager

An aggregate manager controls access to the substrate components. It controls some set of infrastructure resources in a particular site consisting of a set of servers, storage units, network elements or other components under common ownership and control. AMs inform brokers about available resources by passing to the resource advertisement information models. AMs may be associated with more than one broker and the partitioning of resources between brokers is the decision left to the AM. Oversubscription is possible, depending on the deployment needs. FABRIC enables a substrate provider to outsource resource arbitration and calendar scheduling to a broker. By delegating resources to the broker, the AM consents to the broker’s policies, and agrees to try to honor reservations issued by the broker if the user has authorization on the AM.

Besides, common code each AM type has resource-specific modules that determine its resource allocation behavior (Resource Management Policy) and the specific actions it takes to provision a sliver (Resource Handler). Both plugins are invoked by AM common core code based on the resource type or type of request being considered.

Handlers

The AM up calls a handler interface to setup and teardown each sliver. Resource handlers perform any substrate-specific configuration actions needed to implement slivers. The handler interface includes a probe method to poll the current status of a sliver, and modify to adjust attributes of a sliver.

Handlers are registered and selected by resource type. Each handler invocation executes in an independent thread, so handlers may block for slow configuration actions. Handlers are invoked through a class called HandlerProcessor, which can invoke an interpreter for a handler scripting language. The handlers will be written in the Python scripting language.

Each handler implements 3 basic operation types for a resource:

  • Create - provision a resource
    • Example - create a VM or a bare-metal node, or a network connection
  • Delete - un-provision a resource
    • Undo the create above
  • Modify - modify the state of a resource
    • Modify a property of the VM, or a network connection (e.g. change bandwidth)

Each operation can have subcommands and parameters that determine the details of the actions taken, some of which are discussed below. These parameters help ‘stitch’ multiple slivers together. A canonical example is the passing of network information from the handler provisioning the network to the handler provisioning a compute node so that the compute node ends up with correct network configuration (e.g. attached to a correct VLAN). Specific parameters for operations are a matter of convention between the resource management policy and the plugin.

Handlers receive the parameters as part of the provisioning workflow (sequence of redeem operations) executed by the Orchestrator on the AMs. They can also pass information back to the Orchestrator about reserved resources as part of the standard exchange of messages between AM and Orchestrator during the provisioning.

Playbooks

Handlers use Ansible Playbooks for provisioning.

Interface and design

Class Diagram

Configuration

AM Handlers require following configuration to be setup:

  • Inventory information for the headnode
  • Handler config file (VM Handler example depicted below)

VM Handler Config File

VM Handler config file can be found fabric_am/config/vm_handler_config.yml. It describes the Playbook location and names for specific operations.

playbooks:
  location: /etc/fabric/actor/playbooks
  inventory_location: /etc/fabric/actor/playbooks/inventory
  VM: head_vm_provisioning.yml
  GPU: worker_pci_provisioning.yml
  SmartNIC: worker_pci_provisioning.yml
  SharedNIC: worker_pci_provisioning.yml
  FPGA: worker_pci_provisioning.yml
  NVME: worker_pci_provisioning.yml

PCI Devices support

PCI devices are passed to the VMs using PCI pass through. FABRIC also support SRIOV functions which are also passed on to the VMs using PCI pass through. In order to handle the hairpin connections on SRIOV, VEPA mode is enabled on the underlying bridge for the dedicated card used to host the virtual functions.

[kissel@uky-w1 ~]$ sudo bridge link set dev ens6f1 hwmode vepa

What is VEPA?

In the standard mode the software upgrade to the VEB in the hypervisor simply forces each VM frame out to the external switch regardless of destination. This causes no change for destination MAC addresses external to the host, but for destinations within the host (another VM in the same VLAN) it forces that traffic to the upstream switch which forwards it back instead of handling it internally, called a hairpin turn.) It's this hairpin turn that causes the requirement for the upstream switch to have updated firmware, typical STP behavior prevents a switch from forwarding a frame back down the port it was received on. The firmware update allows the negotiation between the physical host and the upstream switch of a VEPA port which then allows this hairpin turn.

VEPA simply forces VM traffic to be handled by an external switch. This allows each VM frame flow to be monitored managed and secured with all of the tools available to the physical switch. This does not provide any type of individual tunnel for the VM, or a configurable switchport but does allow for things like flow statistic gathering, ACL enforcement, etc. Basically we're just pushing the MAC forwarding decision to the physical switch and allowing that switch to perform whatever functions it has available on each transaction. The drawback here is that we are now performing one ingress and egress for each frame that was previously handled internally. This means that there are bandwidth and latency considerations to be made. Functions like Single Root I/O Virtualization (SR/IOV) and Direct Path I/O can alleviate some of the latency issues when implementing this. Like any technology there are typically trade offs that must be weighed. In this case the added control and functionality should outweigh the bandwidth and latency additions.

More details about VEPA can be found here

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

fabric-am-handlers-1.7.1.tar.gz (239.0 kB view details)

Uploaded Source

Built Distribution

fabric_am_handlers-1.7.1-py3-none-any.whl (134.9 kB view details)

Uploaded Python 3

File details

Details for the file fabric-am-handlers-1.7.1.tar.gz.

File metadata

  • Download URL: fabric-am-handlers-1.7.1.tar.gz
  • Upload date:
  • Size: 239.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: python-requests/2.28.2

File hashes

Hashes for fabric-am-handlers-1.7.1.tar.gz
Algorithm Hash digest
SHA256 615c71d8a48f45e4355d9bedaeb39bf0d11cfb4d66680c26073593d82ad88a3c
MD5 6169f36a6b6f7d18693b6505e990a360
BLAKE2b-256 aa1b7046e34da180f7e59e82984c200f8af67a27c418f49659a5c2c689aa0fa6

See more details on using hashes here.

File details

Details for the file fabric_am_handlers-1.7.1-py3-none-any.whl.

File metadata

File hashes

Hashes for fabric_am_handlers-1.7.1-py3-none-any.whl
Algorithm Hash digest
SHA256 5dd6aef86b7939f90098b46000fe3ba5b72a56b773a90045b91de4ce95836729
MD5 ccdd8d4d491cd2cd8306312048ff45ff
BLAKE2b-256 df05e394bc6744e8dc08557e6954ea790f4c2aa60066b66987f5539a35fa69f4

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page