I understand their explanations but for what a benchmark serves that ignores my configurations and also freezes the order, clarify that only happened to me with the 1080ti. If I already know that I do not benchmark if you respect my changes, but there is so little information.
The benchmark is to make those control points, but I should also respect my configuration, I want to give me the result of my configuration. What is the usefulness of a car's Bencmark intensity that, in doing so, fails in some algorithms due to excessive intensity? It has no use.
You say that a process by card is more efficient than a process that includes all the cards to mine. but there are problems with this. It requires MUCH more cpu, and the lack of specialized miners makes the result of HA low.
Example I with the version of Alexis suitable in X17 doing manual, I get 120-125 mhs, an average of 20 mhs per card, and it is a whole process.
If I do it with Trupovt or klaust and from HA, each card only reaches 15-16 mhs at most, although each card has its own process and requires much more CPU, the overall result is much lower than using a more appropriate miner and a process for all cards. That makes 5mhs lost by 6 cards = 30 MHS, I lose a lot of power comparing HA with the only process in the right miner.
The GPU0 card is slower although I use RCP, windows remote desktop, which does not use the gpu. And it does not happen to me when I launch a process with all the miners directly.
I have made modifications to the benchmarking process so that it will use custom intensity settings to help in situations where the default intensity setting may be causing issues on certain GPUs. I'm in the process of testing these changes now. Unfortunately, the testing is taking a little longer than anticipated because certain algorithms were causing other problems when benchmarked individually rather than as part of the miner's default benchmark routine. However, I still plan on releasing 1.8.1 later today - but it may be tomorrow for you considering the difference in time zones.
CPU usage depends greatly on the type of CPU that you are using. However, on contemporary CPUs, including budget Ryzen 5s, each miner process uses only .5% of the CPU. All my mining rigs are running CPU-intensive software such as Folding@Home while mining and not incurring any performance penalties from running each GPU in its own process. Even on a system with eight cards, the combined CPU usage of all those miners should be less than 10% of the system total. It may be higher on older CPUs and admittedly the RAM usage is higher using separate miner processes than a single process. However, that is part of the trade-off of being able to run the most profitable algorithm for each device (which is important in systems that use different types of GPUS) and increased fault-tolerance. Towards the end of your video you show how that one GPU couldn't benchmark but the other GPUs were mining. That would be impossible to do if all the cards shared the same miner process.
I have done tests of certain miners, such as Tpruvot that show an increase in hash rate during actual mining - not benchmarking - for certain algorithms when running in separate processes. For example, even on a system with only two cards, there is an increase of 1 MH/s per card in Lyra2v2 when each GPU is run in its own miner process compared to when they are mining the same pool in a single process. However, this depends on the algorithm as similar testing showed only a slight increase in Neoscrypt performance.
As for your GPU0 issue, even though you are using RDP, Windows is still using the local machine's graphics card to render the desktop before sending the image to the remote connection as described here:
https://msdn.microsoft.com/en-us/library/aa383015(v=vs.85).aspx. In fact, in terms of mining performance, using RDP may be worse than if you had a monitor connected directly to the machine because Microsoft is using its own RDP display driver - which is probably not tuned for performance - and not the NVIDIA device driver. This decrease may not be as evident when you are running all the cards in a single process due to the way the mining software calculates averages for each card.
In regards to Alexis and other miners, I will point out that Hash Auger only uses the mining software's accepted hash rate and not the total hash rate for more accurate pricing estimates. More often than not, the total hash rate is at least 10% higher than the accepted hash rate because it includes the hash rate of all the low-difficulty jobs that were run when the miner and pool began coordinating work. You can see the total hash rate in the miner output window, but Hash Auger only displays the accepted rate and then uses that rate to revise benchmarks.
My concern with Alexis and other older variants of CCMiner is that they have not been updated in a long time and may have comptability issues with newer versions of the NVIDIA device driver. Early versions of Hash Auger included the Nanashi miner which was slightly faster in some algorithms than Tpruvot. However, it caused system freezes for users that had more than a few GPUs in their systems. While Alexis may work fine on your system, as a developer, I have to maintain reliability and stability for the entire user base. Often, that means not including miners that are no longer maintained or that some significant stability issues.