CPU Utilization issue
kmecpp opened this issue ยท 1 comments
With Pterodactyl you can allocate a certain number of cpu cores to each server. For some reason this confuses the Spark CPU monitor which consistently thinks the system CPU load is out of the number of cores available to the docker the server is running in and not the real number of cores available to the system.
This leads to a very high reported utilization (usually consistently 100% but the server wasn't that active when I took the screenshots) but the actual server utilization is much lower
Running on a 4 core 8 thread system so load average numbers should get up to 4 or 8.
Spark gets this info from the OperatingSystemMXBean.
If the values are incorrect, then that's probably a JVM issue - not much I can do to fix, sorry!