-
Notifications
You must be signed in to change notification settings - Fork 12.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Experiment with Uint8Array to reduce memory consumption of sourcemap generation. #43987
Conversation
@typescript-bot perf test this |
Heya @DanielRosenwasser, I've started to run the perf test suite on this PR at c36edb3. You can monitor the build here. Update: The results are in! |
@DanielRosenwasser Here they are:Comparison Report - master..43987
System
Hosts
Scenarios
Developer Information: |
Co-authored-by: David Michon <[email protected]>
@typescript-bot perf test this |
Heya @DanielRosenwasser, I've started to run the perf test suite on this PR at 070005f. You can monitor the build here. Update: The results are in! |
@DanielRosenwasser Here they are:Comparison Report - master..43987
System
Hosts
Scenarios
Developer Information: |
Looks like this around an 8% win in emit time for compiling the old compiler-with-unions codebase. |
@typescript-bot pack this |
Heya @DanielRosenwasser, I've started to run the tarball bundle task on this PR at 070005f. You can monitor the build here. |
Hey @DanielRosenwasser, I've packed this into an installable tgz. You can install it for testing by referencing it in your
and then running |
@typescript-bot perf test this |
Heya @DanielRosenwasser, I've started to run the perf test suite on this PR at 9c42104. You can monitor the build here. Update: The results are in! |
@DanielRosenwasser Here they are:Comparison Report - master..43987
System
Hosts
Scenarios
Developer Information: |
Ahhhh.... Doesn't look like a lot of reduction (-718k (- 0.38%)), but at the sacrifice of readability. The main change I noticed was to reduce the allocation on mapping. May we just set mapping as a array and join it when we need the value? |
A few notes: The Monaco and TFS scenarios do not emit source maps by default, and the material-ui test does not emit anything, which is why we're not seeing changes for those scenarios. Compiler-Unions has source maps enabled by default, and Angular has inline source maps enabled, so those are the only scenarios that will really show any differences. This leads me to a few thoughts about emit performance:
|
@rbuckton It is possible to get some of the gains by appending character codes to an ordinary array, and running it through String.fromCharCode >=1024 characters at a time and concatenating the results. That approach works on ES5, but will retain some of the allocation overhead. Probably worth a separate perf test. I know that keeping data UTF-8 encoded throughout the entire pipeline was one of the tricks that ESBuild uses to help with its performance; it'd be interesting to see what effect it has in ECMAScript. |
@DanielRosenwasser , would be useful to run the allocator tracking tool you ran on the baseline again to have a comparison |
@typescript-bot perf test this |
Heya @DanielRosenwasser, I've started to run the perf test suite on this PR at 1b4a431. You can monitor the build here. Update: The results are in! |
@dmichon-msft @amcasey I've tried integrating the same batching trick, though not the "commit on every 1024 characters generated" approach, so let's see how this works out. |
@DanielRosenwasser Here they are:Comparison Report - master..43987
System
Hosts
Scenarios
Developer Information: |
src/compiler/sourcemap.ts
Outdated
function toJSON(): RawSourceMap { | ||
commitPendingMapping(); | ||
const mappings = (lastMappings ??= serializeMappings(mappingsPos)); |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This seems sketchy. Do we guarantee that no new mappings will ever be added after calling toJSON()
?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The new version which just uses your 1024-byte buffer trick no longer needs a cached value for mappings
.
@typescript-bot perf test this |
Heya @DanielRosenwasser, I've started to run the perf test suite on this PR at f850a85. You can monitor the build here. Update: The results are in! |
@DanielRosenwasser Here they are:Comparison Report - master..43987
System
Hosts
Scenarios
Developer Information: |
I'm guessing the memory bump is from allocating a fresh buffer for each source map generator, which for small files may be larger than the previous string allocations? |
@typescript-bot perf test this |
Heya @DanielRosenwasser, I've started to run the perf test suite on this PR at aa732cf. You can monitor the build here. Update: The results are in! |
@DanielRosenwasser Here they are:Comparison Report - master..43987
System
Hosts
Scenarios
Developer Information: |
Looks like in Node 10 and 12, using an ASCII decoding reduces overall compile time by 100%. 😅 |
@typescript-bot perf test this |
Heya @DanielRosenwasser, I've started to run the perf test suite on this PR at 25e15fb. You can monitor the build here. Update: The results are in! |
@DanielRosenwasser Here they are:Comparison Report - master..43987
System
Hosts
Scenarios
Developer Information: |
This was fun! ultimately, @dmichon-msft's approach is way simpler and pretty much just as fast, so I'll defer to #44031. |
I was playing around with pprof the other day, and took a heap profile of the compiler while compiling itself:
Something that I noticed was a function called
addMapping
causing a lot of allocations.I realized the allocations might be coming from a combination of repeated string concatenations and allocations of individual mappings. One approach to use an array and join the results seemed to have mixed results.
I then realized that since every mapping falls within the ASCII range, we can just throw every calculated character code into a byte array and decode it at the end.
One downside of this is that the current implementation requires a
TextDecoder
which isn't available on all runtimes - Node.js 10 supports it, but it's under theutil
module. Any farther back than that, and you're hosed.My approach is pretty naïve, but it already shows a sizable improvement on the emit time for compiler-with-unions - probably since we're generating a single massive file.
I think there's still room for more wins such as:
Fixes #43999.