Data-center fabrics are scaling beyond “just 10G”: east-west microbursts from containers, NVMe-over-TCP, and AI pipelines mean your Data Center Switches need deep buffers, lossless options, and clean scale to 100/200/400G.
The NS-S6500 family from Network-Switch delivers exactly that same-class Leaf Switch and Spine Switch building blocks with 25G access and 100G uplinks, plus new QSFP56 200G and QSFP-DD 400G uplinks, shipped as your product.
We can customize everything you touch: bezel and logo, labels/packaging, and day-0 software (VLAN plan, QoS/AAA, Syslog/SNMP, LACP, templates).
The portfolio spans fixed 25G leaf with 8 × 100G uplinks, 10G/25G access with 100G uplinks, 64-port 100G systems for leaf/spine fan-in, and next-gen boxes offering 200G access + 400G uplinks or 48×100G + 8×400G with hot-swap PSUs/fans and EVPN-VXLAN/M-LAG for modern fabrics. Public product pages and datasheets confirm the port maps and, for the 400G class, headline performance and ZR/ZR+ optics options.
Product Overview
Where each sub-series fits?
- NS-S6510 Switches (25G/100G class): 48 × SFP28 25G leaf + 8 × QSFP28 100G uplinks, plus variants with 32 × 100G or a compact box-type modular system that scales to 96 × 25G or 32 × 100G via plug-in cards, ideal for ToR TOR Switch roles or compact aggregation.
- NS-S6520 Switches (high-fan-in 100G): dense 64 × 100G QSFP28 for leaf/spine aggregation or small spine layers where you want maximum 100G density in 2RU class hardware with redundant PSUs/fans.
- NS-S6580 Switches (200G/400G class): two fixed systems, 48 × 100G + 8 × 400G (QSFP-DD) and 24 × 200G (QSFP56) + 8 × 400G—for AI/HPC pods, fat-tree spines, or DCI, including 400G ZR/ZR+ optics up to ~120 km and split modes (200G → 2 × 100G).
Because these are Network-Switch branded equivalents, we can ship with your logo and all your defaults baked in LACP/MLAG policies, MSTP edge guards, CoPP/CPU-protect rules, QoS/AAA order, SNMPv3/Syslog, login banners, and per-port “roles” (server/storage/vHost/OOB).
Models Lineup at a Glance
NS model names mirror same-class porting. Specs below reflect the public series pages and datasheets.
Model (NS) | Downlinks (Leaf Ports) | Uplinks | Fabric & Performance | Typical Roles | Notes |
NS-S6510-48VS8CQ | 48 × SFP28 25G | 8 × QSFP28 100G | 25G leaf with 100G fan-in; 32 MB buffer & advanced scheduling for bursts | ToR Leaf Switch; campus/DC 25G access | Cross-device LAG (M-LAG/VSU) and VXLAN noted on public page. |
NS-S6510-32CQ | 32 × QSFP28 100G | — | 100G DCI/aggregation leaf; lossless Ethernet for RDMA | 100G ToR or aggregation in HPC/AI | Lossless E2E RDMA and telemetry highlighted. |
NS-S6510-4C (modular) | up to 96 × 25G (via cards) | or 32 × 100G | 2U Modular Switches (box-type) | Flexible leaf/aggregation | Supports M6500-24VS2CQ / M6500-08CQ cards. |
NS-S6520-64CQ | 64 × QSFP28 100G | — | Dense 100G; hot-swap PSUs (1+1) & fans | Leaf/spine aggregation; compact Spine Switch | 64×100G confirmed on spec pages. |
NS-S6580-48CQ8QC | 48 × 100G | 8 × 400G (QSFP-DD) | Up to 16.0 Tbps / 5,350 Mpps per public listings | Spine/aggregation; AI/HPC pods | 400G uplinks; redundancy (5+1 fans, 1+1 PSUs) noted by listings. |
NS-S6580-24DC8QC | 24 × 200G (QSFP56) | 8 × 400G (QSFP-DD) | 200G access; 400G ZR/ZR+ to ~120 km; 200G split to 2×100G | Spine/aggregation or 200G leaf | Official page details 200G/400G modes and ZR/ZR+. |
Fabric & Software
- EVPN-VXLAN overlays for scalable L2 domains and any-to-any L3 gateways at the leaf; ideal when workloads migrate across racks/zones. (Supported across 25G/100G classes.)
- M-LAG / VSU de-stacking for active-active upstreams and millisecond-class failovers without spanning-tree loops.
- Lossless Ethernet and buffer scheduling for RDMA flows (RoCEv2) where required—especially on the 100G leaf/aggregation SKUs.
- Routing & policy: dual-stack IPv4/IPv6 at line rate; static, OSPF/IS-IS/BGP (image-dependent) so you can run routed leaf/aggregation with edge policy. (Series datasheets list protocol sets.)
- Ops & serviceability: hot-swap PSUs/fans, dual-image/boot, telemetry (sFlow/streaming), optical power readouts, and Web/CLI/SSH with SNMP/Syslog integration.
Deployment Scenarios
25G leaf with 100G spine (NS-S6510-48VS8CQ)
Standardize on 25G server NICs with 8 × 100G northbound. Use two or four 100G links in M-LAG toward the aggregation pair; VXLAN at the leaf keeps mobility and tenant policy clean during host moves. 32 MB buffer and smarter scheduling absorb microbursts from builds and backup windows.
Compact 100G aggregation or DCI edge (NS-S6510-32CQ)
Where you terminate multiple 25G leafs or run DCI, a 32 × 100G box provides non-blocking fan-in and lossless Ethernet for RDMA—useful in HPC and NVMe-over-TCP designs.
Box-type modular leaf/agg (NS-S6510-4C)
Start with 96 × 25G and later re-card to 32 × 100G as uplinks scale. Box-type Chassis Switch behavior (field-replaceable I/O modules) gives capacity flexibility without a full-size chassis.
Dense 100G fan-in for small spines (NS-S6520-64CQ)
Use 64 × 100G to aggregate many leafs in two rack units. Redundant PSUs/fans keep maintenance simple; map QSFP28 breakouts where needed.
Next-gen 200G/400G pod or spine (NS-S6580-24DC8QC / 48CQ8QC)
For AI/HPC or greenfield cores, pick 24 × 200G + 8 × 400G, with 200G→2×100G split to reuse existing 100G optics and 400G ZR/ZR+ up to ~120 km for campus-to-DC interconnect. Or select 48 × 100G + 8 × 400G to super-charge a 100G fabric while seeding 400G Uplinks. Headline fabric numbers (up to 16.0 Tbps / 5,350 Mpps) give you planning headroom.
Optics & Cabling
Port Type | Typical Modules / Cables | Common Use | Notes |
QSFP-DD 400G | SR8/DR4/FR4/LR4, ZR/ZR+ for long span | Spine uplinks, DCI | ZR/ZR+ interconnect supported to ~120 km on the 200G/400G class. |
QSFP56 200G | 200G SR4/FR4; breakout 2×100G | 200G leaf or aggregation | 200G port split to 2×100G documented on the 24DC8QC class. |
QSFP28 100G | SR4/DR/FR/LR; breakout 4×25G | 100G spine/agg and leaf uplinks | Dense 100G systems (32CQ/64CQ) use this across the board. |
SFP28 25G | SR/LR; DAC/AOC | Server access (25G leaf) | 48VS8CQ class lists 25G access with 100G uplinks. |
NS vs. Fixed-Brand Spine-Leaf
Aspect | NS-S6510 / NS-S6520 / NS-S6580 | Fixed-Brand OEM | Other Third-Party Resellers |
Performance & Ports | Same-class hardware: 25G leaf + 100G Uplinks, 64×100G, and 200G/400G with QSFP56/QSFP-DD | OEM-grade, but features may hinge on licenses/optics | Mixed; may be re-badged/refurb |
Fabric & Overlays | EVPN-VXLAN, M-LAG/VSU, telemetry | Similar, often tied to proprietary toolchains | Feature parity varies by firmware |
Customization | Exterior (logo), labels; preloaded VLAN/QoS/AAA, LACP, ERPS/edge guards, ACL baselines | Branding/defaults fixed; limited templateing | Cosmetic only; config after arrival |
Serviceability | Hot-swap PSUs/fans, dual-image; box-type Modular Switches option | Similar FRUs but ecosystem-locked SKUs common | FRU sourcing inconsistent |
Pricing & TCO | Typically ~30% under OEM; ships ready-to-deploy | Premium list + license add-ons | Slightly under OEM; fewer guarantees |
Lead Time | Custom SKUs; factory pre-config & burn-in | Standard SKUs | Variable, long tails |
Operations, Visibility & Security
- Zero-touch day-0: We preload images and configs—VLAN plan (e.g., 10-User / 20-App / 30-Storage), LACP/MLAG policies on uplinks, MSTP edge guards, CoPP/NFPP, SNMPv3/Syslog, and port roles—so every DC Switch or Campus Core Switch boots consistently.
- Faster troubleshooting: Web/CLI/SSH, SNMP, Syslog/telemetry, optical power readouts, and dual-image recovery keep change windows short; hot-swap PSUs/fans lower the blast radius.
- Security & QoS defaults: DHCP Snooping + IP Source Guard + DAI, CPU-protect, and 802.1p/DSCP queueing patterns tuned for voice/video and storage flows. (Protocol suites listed in series docs.)
FAQs
Q1: Which NS-S6500 models support 400G?
A: NS-S6580-48CQ8QC and NS-S6580-24DC8QC provide 8 × 400G QSFP-DD uplinks; the 24DC8QC adds 24 × 200G access and 400G ZR/ZR+ for ~120 km interconnect.
Q2: Does the 200G platform work with existing 100G optics?
A: Yes, each 200G (QSFP56) port can split into 2 × 100G, easing staged upgrades.
Q3: What’s the headline performance on the 48×100G + 8×400G box?
A: Public listings cite 16.0 Tbps switching and 5,350 Mpps forwarding, with 1+1 PSUs and 5+1 fans.
Q4: Do the 25G leaf models support active-active uplinks?
A: Yes, M-LAG/VSU (de-stacking) is documented for the 25G leaf class, enabling active-active upstreams.
Q5: Is there a “Chassis Switch” option in this family?
A: The NS-S6510-4C is a 2U modular box that accepts I/O cards (up to 96 × 25G or 32 × 100G)—a compact alternative to full chassis systems.
Conclusion
Whether you’re refreshing a 100G fabric, introducing 200G/400G Uplinks, or standardizing 25G leaves, the NS-S6510 Switches, NS-S6520 Switches, and NS-S6580 Switches from Network-Switch cover the span from TOR Switch roles to Spine Switch cores.
You get standards-aligned optics (QSFP28, QSFP56, QSFP-DD), modern EVPN-VXLAN fabrics with M-LAG resiliency, and factory pre-configuration under your brand. That means lower rollout friction today, and a cleaner path to tomorrow’s bandwidth.
Did this article help you or not? Tell us on Facebook and LinkedIn . We’d love to hear from you!