-
Notifications
You must be signed in to change notification settings - Fork 31
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
N^2 pbench data being collected #73
Comments
It only creates label
for fio test we have below organization once pbench tools are registered. I think it is not recursive @akrzos @chaitanyaenr opinions?
|
Hi @portante can you provide us the evidence so we can investigate ourselves? My first impression is this should not be occurring as logically this code simply prints only hostnames of nodes determined by labeling for starting/stopping/copying of tool data as @ekuric has pointed out in his example. |
@akrzos, @ekuric, yes the output shown above of the contents of the As for evidence, I have one from April 8th that is from our internal server (the URL I can't share here), but the output from a BAD tools directory looks like:
It should look like:
This results from having the "remote@*" files in the tool directories on all hosts, and so To prevent this, using Since you can't today rely on |
So this must be from the old svt version of tooling because openshift-scale/workloads didn't exist until late may. (See first commit - fa15713 ) This is further illustrated by the labels beginning with "svt_" in the example. The labels from a workload run here will resemble:
Lastly configmaps mounted are read only:
Thus this condition should never occur since the configmap would overwrite what the controller is writing. |
I see this on other more recent runs as well:
|
This shows tools proc-interrupts and proc-vmstat both of which we do not register with this tooling (nor ever have) but the svt tooling does. This result also looks to be from openshift on openstack which no-one from our specific team was running at that time period. Maybe @smalleni or someone from the shiftstack team might have been running at this time. Can you try reproducing this with the actual tooling? |
There is evidence to suggest that because of lines 22 - 33 of
master/workloads/templates/workload-baseline-script-cm.yml.j2
all nodes are copy tool data from all other nodes to their local file system. This seems to result in the actual controller having multiple unexpected sub-directories in stored tool data containing the resulting data.If lines 22-33 are only executed on the pbench controller so that only that
tools-default
file has the remote hosts listed, this would not happen.This is resulting in unexpectedly large tar balls of pbench data being collected.
The text was updated successfully, but these errors were encountered: