- General Workflow
- play_arrow Apstra GUI
- play_arrow Design
- Logical Devices (Datacenter Design)
- Interface Maps (Datacenter Design)
- Rack Types (Datacenter Design)
- Templates (Datacenter Design)
- Config Templates (Freeform Design)
- play_arrow Configlets (Datacenter Design)
- play_arrow Property Sets (Datacenter Design)
- TCP/UDP Port Aliases (Datacenter Design)
- Tags (Design)
- play_arrow Devices
- Device Configuration Lifecycle
- play_arrow Managed Devices
- play_arrow System Agents
- play_arrow Pristine Configuration
- play_arrow Telemetry
- play_arrow Apstra ZTP
- play_arrow Resources Introduction
- play_arrow Freeform Reference Design
- Create / Delete Freeform Blueprint
- Freeform Blueprint Summary and Dashboard
- Topology (Freeform)
- play_arrow Systems (Freeform)
- Device Context (Freeform)
- play_arrow Links (Freeform)
- play_arrow Resource Management
- play_arrow Config Templates (Freeform Blueprint)
- Import Device Profile (Freeform)
- play_arrow Property Sets (Freeform Blueprints)
- play_arrow Tags (Freeform Blueprint)
- Tasks - Staged (Freeform)
- play_arrow Active
- Commit Blueprint
- Time Voyager
- play_arrow Analytics
- Configure Auto-Enabled Dashboards
- Instantiate Predefined Dashboard
- Create Analytics Dashboard
- Edit / Delete Dashboard
- Anomalies (Analytics)
- Widgets Overview
- Create Anomaly Heat Map Widget
- Create Stage Widget
- Edit / Delete Widget
- Probes
- Instantiate Predefined Probe
- Create Probe
- Import / Export Probe
- Edit / Delete Probe
- play_arrow Providers (External Systems)
- play_arrow Platform
- play_arrow User/Role Management (Platform)
- play_arrow Security (Platform)
- Syslog Configuration (Platform)
- Receivers (Platform)
- Global Statistics (Platform)
- Event Log (Platform)
- play_arrow Apstra VM Clusters
- play_arrow Developers (Platform)
- play_arrow Juniper Technical Support
- Favorites & User
- play_arrow Apstra Server Management
- Monitor Apstra Server via CLI
- Restart Apstra Server
- Reset Apstra Server VM Password
- Reinstall Apstra Server
- Apstra Database Overview
- Back up Apstra Database
- Restore Apstra Database
- Reset Apstra Database
- Migrate Apstra Database
- Replace SSL Certificate on Apstra Server with Signed One
- Replace SSL Certificate on Apstra Server with Self-Signed One
- Change Apstra Server Hostname
- Apstra CLI Utility
- play_arrow Guides
- play_arrow References
- play_arrow Apstra Feature Matrix
- Qualified Devices and NOS Versions
- NOS Upgrade Paths (Devices)
- play_arrow Predefined Dashboards (Analytics)
- Device Environmental Health Summary Dashboard (New in 4.1.2)
- Device Health Summary Dashboard
- Device Telemetry Health Summary Dashboard (New in 4.1.2)
- Drain Validation Dashboard
- Throughput Health MLAG Dashboard
- Traffic Trends Dashboard
- Virtual Infra Fabric Health Check Dashboard
- Virtual Infra Redundancy Check Dashboard
- play_arrow Predefined Probes (Analytics)
- BGP Session Flapping Probe
- Bandwidth Utilization Probe
- Critical Services: Utilization, Trending, Alerting Probe
- Device Environmental Checks Probe (New in 4.1.2)
- Device System Health Probe
- Device Telemetry Health Probe
- Device Traffic Probe
- Drain Traffic Anomaly Probe
- ECMP Imbalance (External Interfaces) Probe
- ECMP Imbalance (Fabric Interfaces) Probe
- ECMP Imbalance (Spine to Superspine Interfaces) Probe
- ESI Imbalance Probe
- EVPN Host Flapping Probe
- EVPN VXLAN Type-3 Route Validation Probe
- EVPN VXLAN Type-5 Route Validation Probe
- External Routes Probe
- Hot/Cold Interface Counters (Fabric Interfaces) Probe
- Hot/Cold Interface Counters (Specific Interfaces) Probe
- Hot/Cold Interface Counters (Spine to Superspine Interfaces) Probe
- Hypervisor and Fabric LAG Config Mismatch Probe (Virtual Infra)
- Hypervisor and Fabric VLAN Config Mismatch Probe (Virtual Infra)
- Hypervisor MTU Mismatch Probe (Virtual Infra NSX-T Only)
- Hypervisor MTU Threshold Check Probe (Virtual Infra)
- Hypervisor Missing LLDP Config Probe (Virtual Infra)
- Hypervisor Redundancy Checks Probe (Virtual Infra)
- Interface Flapping (Fabric Interfaces) Probe
- Interface Flapping (Specific Interfaces) Probe
- Interface Flapping (Specific Interfaces) Probe
- Interface Policy 802.1x Probe
- LAG Imbalance Probe
- Leafs Hosting Critical Services: Utilization, Trending, Alerting Probe
- Link Fault Tolerance in Leaf and Access LAGs Probe
- MLAG Imbalance Probe
- Multiagent Detector Probe
- Optical Transceivers Probe
- Packet Discard Percentage Probe
- Spine Fault Tolerance Probe
- Total East/West Traffic Probe
- VMs without Fabric Configured VLANs Probe (Virtual Infra)
- VXLAN Flood List Validation Probe
- play_arrow Probe Processors (Analytics)
- Processor: Accumulate
- Processor: Average
- Processor: Comparison
- Processor: EVPN Type 3
- Processor: EVPN Type 5
- Processor: Extensible Service Data Collector
- Processor: Generic Graph Collector
- Processor: Generic Service Data Collector
- Processor: Interface Counters
- Processor: Logical Operator
- Processor: Match Count
- Processor: Match Percentage
- Processor: Match String
- Processor: Max
- Processor: Min
- Processor: Periodic Average
- Processor: Range
- Processor: Ratio
- Processor: Service Data Collector
- Processor: Set Comparison
- Processor: Set Count
- Processor: Standard Deviation
- Processor: State
- Processor: Subtract
- Processor: Sum
- Processor: System Utilization
- Processor: Time in State
- Processor: Traffic Monitor
- Processor: Union
- Processor: VXLAN Floodlist
- Configlet Examples (Design)
- Apstra-CLI Commands
- Apstra EVPN Support Addendum
- Apstra Server Configuration File
- Agent Configuration File (Devices)
- Graph
- Juniper Apstra Technology Preview
NSX-T Integration
VMware NSX-T Integration Overview
You can integrate NSX-T with Apstra software to help deploy fabric VLANs that are needed for deploying NSX-T in the data center or for providing connectivity between NSX-T overlay networks and fabric underlay networks. You can accelerate NSX-T deployments by making sure the fabric is ready in terms of LAG, MTU and VLAN configuration as per NSX-T transport node requirements. This feature also helps network operators with fabric visibility in terms of seeing all the NSX-T VMs, VM ports, and physical gateway ports. NSX-T integration helps identify issues on the fabric and on the virtual infrastructure. It eliminates manual config validation tasks between the NSX-T nodes side and the ToR switches.
When NSX-T VM is attached into VLAN Transport, VM query shows TOR switch/interface information together. When NSX-T VM is attached into Overlay Transport, VM query doesn't show TOR switch/interface information. Be sure to add ESXi host in generic systems, not external generic systems.
As of Apstra version 4.1.2, you can create Virtual Infra Managers for NSX-T Manager version 3.2.x using DVS mode. You can also add multiple Virtual Infra Managers per blueprint. This is useful when you have multiple NSX-T Managers or multiple vCenter Servers hosted in the same fabric blueprint. You'll need to provide the vCenter compute managers information (address and credentials) when you add the NSX-T Virtual Infra.
Supported Versions
Apstra version 4.1.2: VMware NSX-T Manager version 3.2.x
Apstra version 4.1.1 and 4.1.0: VMware NSX-T version 3.0.2
Limitations
- NSX-T Edge VM migration is supported only within a rack. Attempting to migrate between racks results in BGP disruption. You can migrate the NSX-T Edge VM from the ESXi host connected to leaf pair (that is, ToR-Leaf and ToR-Right) to the other ESXi host which is connected to single leaf with the rack.
(Apstra versions 4.1.1 and 4.1.0 only) Having more than one NSX-T virtual infra in a blueprint is not supported. We recommend only one virtual infra per blueprint.
- (Apstra versions 4.1.1 and 4.1.0 only) NSX-T integration does not support DVS port group with VLAN-type trunking.
Enable NSX-T Integration
We recommend that you create a user profile dedicated to managing NSX-T integration activities.
Virtual Infrastructure Visibility
When you've successfully integrated NSX-T, you have visibility of NSX-T VMs and transport nodes in the virtual infrastructure. You can query the status of the VMware fabric health.
To see a list of the VMs connected to the hypervisor, navigate to the dashboard and
scroll to fabric health for VMware option.
You can also query VMs that are hosted on hypervisors connected to ToR leaf devices.
From the blueprint, navigate to Active > Query > VMs.
VMs include the following details:
Parameter | Description |
---|---|
VM Name | The Virtual Machine name which is hosted on NSX managed hypervisor. |
Hosted On | The ESXi host on which Virtual Machine is hosted. |
Hypervisor Hostname | The hypervisor hostname on which Virtual Machine is hosted and is connected to the leaf TORs in a fabric. |
Hypervisor Version | The software version of OS running on the hypervisor. |
VM IP | The IP address as reported by NSX-T after the installation of VM tools. If the IP address is not available this field is empty. Apstra displays VM IP if the IP address is available on installation VM tools on the VM. |
Leaf:Interface | System ID for the interface on the leaf to which ESXi host is connected and on which VM resides. |
Port Group Name:VLAN ID | The VLAN ID which NSX-T port groups are using. Overlay VM to VM traffic in a NSX-T enabled Data Center tunnels between transport nodes over this Virtual network. |
MAC Addresses | MAC address of the VM connected to the Apstra Fabric. |
Virtual Infra address | IP address of the NSX-T infra added to a Blueprint |
To search for nodes in the physical topology that have VMs, navigate to
Active > Physical and select Has
VMs? from the Nodes drop-down list.
If the VM is moved from one Transport node to another in NSX-T it can be visualized
in Apstra under Active > Physical > Nodes > Generic System
(Node_name). Select the VMs tab as shown
below:
Validate Virtual Infra Integration
You can validate virtual infra with intent-based analytics. Apstra validates BGP
session towards NSX-T Edge. If BGP neighborship in NSX-T Manager is deleted, then
respective anomalies are displayed in the Apstra dashboard.
Two predefined analytics dashboards (as listed below) are available that instantiate predefined virtual infra probes.
Virtual Infra Fabric Health Check Dashboard
- Hypervisor MTU Mismatch Probe
- Hypervisor MTU Threshold Check Probe
- Hypervisor & Fabric LAG Config Mismatch Probe
- Hypervisor & Fabric VLAN Config Mismatch Probe
- Hypervisor Missing LLDP Config Probe
- VMs without Fabric Configured VLANs Probe
Virtual Infra Redundancy Check Dashboard
For more information, see Analytics dashboard and Instantiate Predefined Probe.
Disable Virtual Infra Integration
To disable virtual infra integrations, delete them from the blueprint and external systems.
- From the blueprint, navigate to Staged > Virtual > Virtual Infra and click the Delete button for the virtual infra to disable.
- Click Uncommitted (top menu) and commit the deletion.
- From the left navigation menu, navigate to External Systems > Virtual Infra Managers and click the Delete button for the virtual infra to disable.