Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Pre-allocate parsedPatterns and cache some calculations #371

Merged
merged 6 commits into from
Mar 6, 2025

Conversation

DanielRosenwasser
Copy link
Member

For paths of CompilerOptions, we currently split the entries into a set of strings to do a quick lookup, and a (shorter) list of patterns to walk on. Later on, exact matches or any match with the longest prefix (breaking any ties by ordering in paths, see #369 and #370 for fixes around that behavior).

A lot of this work was done in microsoft/TypeScript#59048, but that PR also cached the creation of these parsedPatterns. tsgo currently does not cache these at all, and on a fairly large internal codebase, this results in a significant amount of user time (13.7 seconds below) spent in tryParsePatterns:

tryParsePatterns spends 13.7 seconds of user time

This PR reintroduces the cache for parsedPatterns for paths in CompilerOptions, and places it on the Resolver's caches.

This PR also does an extra walk of paths to provide a size hint. When uncached, this actually added time at the cost of more allocation. It may be worth experimenting with removing the extra walk.

@DanielRosenwasser DanielRosenwasser marked this pull request as ready for review February 20, 2025 00:32
@@ -18,6 +19,7 @@ type caches struct {
typeReferenceDirectiveCache *resolutionCache[*ResolvedTypeReferenceDirective]
packageJsonInfoCache *packagejson.InfoCache
resolvedTypeReferenceDirectiveLookupLocations map[*ResolvedTypeReferenceDirective]*LookupLocations
parsedPatternsCache map[*collections.OrderedMap[string, []string]]parsedPatterns
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I can't say I super like holding onto these maps like this, but I guess we'll probably throw these objects away periodically.

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Wait, we don't even need this to be a map; isn't tryParsePatternsCached only ever called on r.compilerOptions.Paths, so we could just sync.Once it.

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I haven’t exactly worked out how I want to do this, but eventually there will be a module resolution cache that persists across program updates and/or is shared between different programs.

Copy link

@Copilot Copilot AI left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

PR Overview

This PR improves module resolution performance by caching the parsed patterns used for path matching. Key changes include:

  • Replacing the direct call to tryParsePatterns with a cached version (tryParsePatternsCached) in the Resolver.
  • Updating the parsedPatterns structure and its initialization to use core.Set instead of collections.OrderedSet.
  • Adding a new cache field for parsed patterns in the Resolver's caches and minor documentation updates in the core set implementation.

Reviewed Changes

File Description
internal/compiler/module/resolver.go Introduces a caching mechanism for parsedPatterns and refactors the pattern parsing logic.
internal/core/set.go Adds a comment to NewSetWithSizeHint for clarity.
internal/compiler/module/cache.go Adds the parsedPatternsCache field to the caches struct.

Copilot reviewed 3 out of 3 changed files in this pull request and generated 1 comment.

Tip: Copilot code review supports C#, Go, Java, JavaScript, Markdown, Python, Ruby and TypeScript, with more languages coming soon. Learn more

if r.parsedPatternsCache == nil {
r.parsedPatternsCache = make(map[*collections.OrderedMap[string, []string]]parsedPatterns)
}
r.caches.parsedPatternsCache[paths] = pathPatterns
Copy link
Preview

Copilot AI Feb 26, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

There is an inconsistency in caching: the code first checks r.parsedPatternsCache but then assigns the new value to r.caches.parsedPatternsCache. Consider using a single cache field consistently to avoid unexpected behavior.

Suggested change
r.caches.parsedPatternsCache[paths] = pathPatterns
r.parsedPatternsCache[paths] = pathPatterns

Copilot is powered by AI, so mistakes are possible. Review output carefully before use.

@DanielRosenwasser DanielRosenwasser force-pushed the parsedPatternsOptimization branch from 6b825a0 to 1bf7ca3 Compare March 4, 2025 22:16
@@ -4,6 +4,13 @@ type Set[T comparable] struct {
M map[T]struct{}
}

// NewSetWithSizeHint creates a new Set with a hint for the number of elements it will contain.
func NewSetWithSizeHint[T comparable](hint int) *Set[T] {
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Probably just do Set[T] here, since it's always dereferenced at its use.

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

But everyone else is doing *core.Set (everyone else is one function)

Copy link
Member

@jakebailey jakebailey left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

With the one change, seems okay, but will defer to @andrewbranch.

@DanielRosenwasser DanielRosenwasser added this pull request to the merge queue Mar 6, 2025
Merged via the queue into main with commit d3fe2f5 Mar 6, 2025
17 checks passed
@jakebailey jakebailey deleted the parsedPatternsOptimization branch March 7, 2025 18:37
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants