Hi,

I’ve got a network setup consisting of 3 proxmox servers of which two have a 10Gbe interface. All interfaces are RJ45 copper ports.

So far I’ve got two qnap switches with 2 10Gbe and 5 2.5Gbe portss. They are connected to each other via one of the 10Gbe ports, and one of the servers on the other one.

This setup is working flawlessly, iperf measurements show 9.4gbit/s in both directions.

Recently I tried to increase my network for future expansion and bought a trendnet switch with 5 10Gbe ports.

Weird problems occurred and as they were also described in some amazon reviews I returned the switch and bought another one, this time a 4port 10Gbe switch from ubiquity.

Again there are problems. This time one direction seems to be at 10gbit/s as expected while the other direction (between the same two servers) is limited to 1gbit/s. The connection to the third server shows the same problem (one direction 2.5gbit/s and the other one limited to 1gbit/s).

All wires are connected correctly and as there is no problem on the qnap switches I do believe that there is no hardware problem on the cable/NIC side.

However I’m not sure if this is this just bad luck or some deep network problem I don’t understand.

Maybe someone has an idea?

  • bigredgiraffe@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    Oh yeah for sure, every time I’m like “it can’t be spanning tree” it is spanning tree. Do you mean copper vs fiber? LC connectors can carry a variety of speeds but generally yeah I try to use fiber or DAC cables which are shielded wherever I can.