while screenshot shows awesome resolution of your screen it does not tell exactly what is happening.
First of - if you are running something in KVM, you can do just the same on host system, to see if that is load issue or something different since packets for guest are manged by host, then by guest and then by host again (if traffic is passing through the guest)
Also, where you are generating the traffic? on the same host adding even more load or somewhere else?
What is purpose of this test?
as different configurations have different possible implementations, while you can have 1000 clients and can have 1000 simple queues for each of them, you also can do the same just having queues for user groups, like starter pack queue, normal queue and high bandwidth queue and all of them using PCQ to manage load. And there will be huge difference in latency having 1000 or 3 queues in configuration.
I too thought they were optomised in V6
I will explain this, i apologize it was not so clear
At this moment all of the kvm's are not in production, they have absolutly no config other than an IP address assigned from the Host machine, they don't even have any queues, i'm explaining this in case its assumed that we want to see the load of each individual KVM, as there might be a separate load there via other reasons.
I could provide more screenshots, but i can assure you that each KVM is not 100% with in it self.
How ever the Host System has the cpu usage very high, because well from what i can understand
the traffic generator which is on the Host system is generating the packets, and sending to its own local bridge, which is the bridge that has the subnet containing the KVM's.
I simply targeted the broadcast domain, and the KVM's are showing the traffic response to the Host machine perfectly on the interface menu.
Queue tree's 1 for each "virtual Ethernet Interface" the queues respond quite clearly
But with Simple Queues, it has very strange behavior, which is on display, if you look closely, you can see that there is strangely
Some queues that have a high amount of traffic at that "instance", but the issue is, im not even sending any data, when i took that screenshot.
So if i was to cut to the chase of the issue, the last say 10 or so queues are stuck in motion the numbers do not change, UNLESS
i fiddle around and disable and enable the say 0-20 on and off, then ! and only then ! do the last 10 or so queues the figures/numbers change very quickly.
I do hope that has explained it more, the purpose of my test was, to bench test a Dell server, i want to test it stability, research an development you could say, I want to provide this as a solution in out data center, and this is the makings to see of this is up to standard and stable.