-
Notifications
You must be signed in to change notification settings - Fork 40.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Optimize random string generator to avoid multiple locks & use bit-masking #53720
Conversation
/test pull-kubernetes-kubemark-e2e-gce-big |
/test pull-kubernetes-kubemark-e2e-gce-big |
ea81420
to
2452824
Compare
/test pull-kubernetes-kubemark-e2e-gce-big |
1 similar comment
/test pull-kubernetes-kubemark-e2e-gce-big |
2452824
to
9d18c41
Compare
/test pull-kubernetes-kubemark-e2e-gce-big |
9d18c41
to
a867f2a
Compare
/test pull-kubernetes-kubemark-e2e-gce-big |
a867f2a
to
f55f81a
Compare
/test pull-kubernetes-kubemark-e2e-gce-big |
@jsafrane would be able to answer. |
0d913d8
to
42c0314
Compare
42c0314
to
c279a53
Compare
Why are we using a global static utility function that needs a global static lock? That seems... bad. We should be initializing a random source for whatever components need it and let them do their locking. The volume controller should only be contending with itself, in which case just give it a seed per worker (which should be easy). |
@smarterclayton I filed #53888 for the issue you stated. I'd prefer having that change in a separate PR as there are multiple places where the change needs to be made (i.e replace global RNG references with |
I agree - let's not try to solve all the problems of random library as a resolving scalability regression. /lgtm |
/approve |
@wojtek-t Also added the benchmark (though I'm still trying to make it run locally). Could you re-lgtm? |
/lgtm |
[APPROVALNOTIFIER] This PR is APPROVED This pull-request has been approved by: shyamjvs, wojtek-t Associated issue: 53327 The full list of commands accepted by this bot can be found here.
Needs approval from an approver in each of these OWNERS Files:
You can indicate your approval by writing |
Results of benchmark: without this PR:
with this PR:
|
Cool, but that's something I expected. |
/test all [submit-queue is verifying that this PR is safe to merge] |
Automatic merge from submit-queue. If you want to cherry-pick this change to another branch, please follow the instructions here. |
Changing it is fine, I just want to make sure we don't continue the pattern
(i.e. if we cut ourselves on a sharp corner, we should fix the corner, not
put a sign up that says "don't touch"). This should be the last time
someone regresses on this by accidentally using this package.
…On Fri, Oct 13, 2017 at 9:11 AM, Kubernetes Submit Queue < ***@***.***> wrote:
Merged #53720 <#53720>.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
<#53720 (comment)>,
or mute the thread
<https://github.com/notifications/unsubscribe-auth/ABG_p4JF90soU7iaTu7vWfegXOXIr9Nuks5sr2GSgaJpZM4P1ems>
.
|
Agree. I just wanted to fix regression ASAP and this seemed to fix it (and already existed). |
Automatic merge from submit-queue. UPSTREAM: 53989: Remove repeated random string generations in scheduler volume predicate @sjenning @smarterclayton Though the upstream PR 53793 has been backported to kube 1.7 branch (53884). I am not sure if we have a plan for another origin rebase to latest kube 1.7, and if we would want to wait for that. So this backports following 3 PRs: kubernetes/kubernetes#53793 kubernetes/kubernetes#53720 (partial) kubernetes/kubernetes#53989
Ref #53327
We recently started seeing a 50% decrease in scheduling throughput (for e.g in kubemark-500 scale job) and turns out #53135 introduced it.
The reason is this call to create a random 32-length string.
From the code of the
rand
utility (which is being heavily used throughout the system for randomizing object names), I noticed following performance issues:rand.Intn()
each of which does a lock+unlock operation on the RNG.. while just 1 lock+unlock operation is enough for alllen()
(we're making n of those).. while we can just use a const string (immutable) which will make len directly available as a cached constant (yes, go does it!)This PR is making the above fixes. I'll try to add some benchmarking to measure the difference (as @wojtek-t suggested).
/cc @kubernetes/sig-scalability-misc @kubernetes/sig-scheduling-bugs @kubernetes/sig-api-machinery-misc @wojtek-t @smarterclayton