Fix some server/watch rebuild issues
Two issues:
1. Fixe potential edit-loop in server/watch mode (see below)
2. Drain the cache eviction stack before we start calculating the change set. This should allow more fine grained rebuilds for bigger sites and/or in low memory situations.
The fix in 6c68142cc1
wasn't really fixing the complete problem.
In Hugo we have some steps that takes more time than others, one example being CSS building with TailwindCSS.
The symptom here is that sometimes when you:
1. Edit content or templates that does not trigger a CSS rebuild => Snappy rebuild.
2. Edit stylesheet or add a CSS class to template that triggers a CSS rebuild => relatively slow rebuild (expected)
3. Then back to content editing or template edits that should not trigger a CSS rebuild => relatively slow rebuild (not expected)
This commit fixes this by pulling the dynacache GC step up and merge it with the cache buster step.
Fixes #13316
This commit is contained in:
parent
778f0d9002
commit
db28695ff5
6 changed files with 59 additions and 51 deletions
16
cache/dynacache/dynacache.go
vendored
16
cache/dynacache/dynacache.go
vendored
|
@ -176,11 +176,12 @@ func (c *Cache) ClearMatching(predicatePartition func(k string, p PartitionManag
|
|||
}
|
||||
|
||||
// ClearOnRebuild prepares the cache for a new rebuild taking the given changeset into account.
|
||||
func (c *Cache) ClearOnRebuild(changeset ...identity.Identity) {
|
||||
// predicate is optional and will clear any entry for which it returns true.
|
||||
func (c *Cache) ClearOnRebuild(predicate func(k, v any) bool, changeset ...identity.Identity) {
|
||||
g := rungroup.Run[PartitionManager](context.Background(), rungroup.Config[PartitionManager]{
|
||||
NumWorkers: len(c.partitions),
|
||||
Handle: func(ctx context.Context, partition PartitionManager) error {
|
||||
partition.clearOnRebuild(changeset...)
|
||||
partition.clearOnRebuild(predicate, changeset...)
|
||||
return nil
|
||||
},
|
||||
})
|
||||
|
@ -479,7 +480,12 @@ func (p *Partition[K, V]) clearMatching(predicate func(k, v any) bool) {
|
|||
})
|
||||
}
|
||||
|
||||
func (p *Partition[K, V]) clearOnRebuild(changeset ...identity.Identity) {
|
||||
func (p *Partition[K, V]) clearOnRebuild(predicate func(k, v any) bool, changeset ...identity.Identity) {
|
||||
if predicate == nil {
|
||||
predicate = func(k, v any) bool {
|
||||
return false
|
||||
}
|
||||
}
|
||||
opts := p.getOptions()
|
||||
if opts.ClearWhen == ClearNever {
|
||||
return
|
||||
|
@ -525,7 +531,7 @@ func (p *Partition[K, V]) clearOnRebuild(changeset ...identity.Identity) {
|
|||
// Second pass needs to be done in a separate loop to catch any
|
||||
// elements marked as stale in the other partitions.
|
||||
p.c.DeleteFunc(func(key K, v V) bool {
|
||||
if shouldDelete(key, v) {
|
||||
if predicate(key, v) || shouldDelete(key, v) {
|
||||
p.trace.Log(
|
||||
logg.StringFunc(
|
||||
func() string {
|
||||
|
@ -601,7 +607,7 @@ type PartitionManager interface {
|
|||
adjustMaxSize(addend int) int
|
||||
getMaxSize() int
|
||||
getOptions() OptionsPartition
|
||||
clearOnRebuild(changeset ...identity.Identity)
|
||||
clearOnRebuild(predicate func(k, v any) bool, changeset ...identity.Identity)
|
||||
clearMatching(predicate func(k, v any) bool)
|
||||
clearStale()
|
||||
}
|
||||
|
|
4
cache/dynacache/dynacache_test.go
vendored
4
cache/dynacache/dynacache_test.go
vendored
|
@ -147,13 +147,13 @@ func TestClear(t *testing.T) {
|
|||
|
||||
c.Assert(cache.Keys(predicateAll), qt.HasLen, 4)
|
||||
|
||||
cache.ClearOnRebuild()
|
||||
cache.ClearOnRebuild(nil)
|
||||
|
||||
// Stale items are always cleared.
|
||||
c.Assert(cache.Keys(predicateAll), qt.HasLen, 2)
|
||||
|
||||
cache = newTestCache(t)
|
||||
cache.ClearOnRebuild(identity.StringIdentity("changed"))
|
||||
cache.ClearOnRebuild(nil, identity.StringIdentity("changed"))
|
||||
|
||||
c.Assert(cache.Keys(nil), qt.HasLen, 1)
|
||||
|
||||
|
|
|
@ -1123,6 +1123,9 @@ func (h *HugoSites) resolveAndClearStateForIdentities(
|
|||
l logg.LevelLogger,
|
||||
cachebuster func(s string) bool, changes []identity.Identity,
|
||||
) error {
|
||||
// Drain the cache eviction stack to start fresh.
|
||||
h.Deps.MemCache.DrainEvictedIdentities()
|
||||
|
||||
h.Log.Debug().Log(logg.StringFunc(
|
||||
func() string {
|
||||
var sb strings.Builder
|
||||
|
@ -1163,17 +1166,32 @@ func (h *HugoSites) resolveAndClearStateForIdentities(
|
|||
}
|
||||
|
||||
// The order matters here:
|
||||
// 1. Handle the cache busters first, as those may produce identities for the page reset step.
|
||||
// 1. Then GC the cache, which may produce changes.
|
||||
// 2. Then reset the page outputs, which may mark some resources as stale.
|
||||
// 3. Then GC the cache.
|
||||
if cachebuster != nil {
|
||||
if err := loggers.TimeTrackfn(func() (logg.LevelLogger, error) {
|
||||
ll := l.WithField("substep", "gc dynacache cachebuster")
|
||||
h.dynacacheGCCacheBuster(cachebuster)
|
||||
return ll, nil
|
||||
}); err != nil {
|
||||
return err
|
||||
if err := loggers.TimeTrackfn(func() (logg.LevelLogger, error) {
|
||||
ll := l.WithField("substep", "gc dynacache")
|
||||
|
||||
predicate := func(k any, v any) bool {
|
||||
if cachebuster != nil {
|
||||
if s, ok := k.(string); ok {
|
||||
return cachebuster(s)
|
||||
}
|
||||
}
|
||||
return false
|
||||
}
|
||||
|
||||
h.MemCache.ClearOnRebuild(predicate, changes...)
|
||||
h.Log.Trace(logg.StringFunc(func() string {
|
||||
var sb strings.Builder
|
||||
sb.WriteString("dynacache keys:\n")
|
||||
for _, key := range h.MemCache.Keys(nil) {
|
||||
sb.WriteString(fmt.Sprintf(" %s\n", key))
|
||||
}
|
||||
return sb.String()
|
||||
}))
|
||||
return ll, nil
|
||||
}); err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
// Drain the cache eviction stack.
|
||||
|
@ -1238,23 +1256,6 @@ func (h *HugoSites) resolveAndClearStateForIdentities(
|
|||
return err
|
||||
}
|
||||
|
||||
if err := loggers.TimeTrackfn(func() (logg.LevelLogger, error) {
|
||||
ll := l.WithField("substep", "gc dynacache")
|
||||
|
||||
h.MemCache.ClearOnRebuild(changes...)
|
||||
h.Log.Trace(logg.StringFunc(func() string {
|
||||
var sb strings.Builder
|
||||
sb.WriteString("dynacache keys:\n")
|
||||
for _, key := range h.MemCache.Keys(nil) {
|
||||
sb.WriteString(fmt.Sprintf(" %s\n", key))
|
||||
}
|
||||
return sb.String()
|
||||
}))
|
||||
return ll, nil
|
||||
}); err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
|
|
|
@ -27,7 +27,6 @@ import (
|
|||
|
||||
"github.com/bep/logg"
|
||||
"github.com/gohugoio/hugo/bufferpool"
|
||||
"github.com/gohugoio/hugo/cache/dynacache"
|
||||
"github.com/gohugoio/hugo/deps"
|
||||
"github.com/gohugoio/hugo/hugofs"
|
||||
"github.com/gohugoio/hugo/hugofs/files"
|
||||
|
@ -828,6 +827,11 @@ func (h *HugoSites) processPartialFileEvents(ctx context.Context, l logg.LevelLo
|
|||
addedContentPaths []*paths.Path
|
||||
)
|
||||
|
||||
var (
|
||||
addedOrChangedContent []pathChange
|
||||
changes []identity.Identity
|
||||
)
|
||||
|
||||
for _, ev := range eventInfos {
|
||||
cpss := h.BaseFs.ResolvePaths(ev.Name)
|
||||
pss := make([]*paths.Path, len(cpss))
|
||||
|
@ -854,6 +858,13 @@ func (h *HugoSites) processPartialFileEvents(ctx context.Context, l logg.LevelLo
|
|||
if err == nil && g != nil {
|
||||
cacheBusters = append(cacheBusters, g)
|
||||
}
|
||||
|
||||
if ev.added {
|
||||
changes = append(changes, identity.StructuralChangeAdd)
|
||||
}
|
||||
if ev.removed {
|
||||
changes = append(changes, identity.StructuralChangeRemove)
|
||||
}
|
||||
}
|
||||
|
||||
if ev.removed {
|
||||
|
@ -865,11 +876,6 @@ func (h *HugoSites) processPartialFileEvents(ctx context.Context, l logg.LevelLo
|
|||
}
|
||||
}
|
||||
|
||||
var (
|
||||
addedOrChangedContent []pathChange
|
||||
changes []identity.Identity
|
||||
)
|
||||
|
||||
// Find the most specific identity possible.
|
||||
handleChange := func(pathInfo *paths.Path, delete, isDir bool) {
|
||||
switch pathInfo.Component() {
|
||||
|
@ -1063,18 +1069,6 @@ func (h *HugoSites) processPartialFileEvents(ctx context.Context, l logg.LevelLo
|
|||
|
||||
resourceFiles := h.fileEventsContentPaths(addedOrChangedContent)
|
||||
|
||||
defer func() {
|
||||
// See issue 13316.
|
||||
h.MemCache.DrainEvictedIdentitiesMatching(func(ki dynacache.KeyIdentity) bool {
|
||||
for _, c := range changes {
|
||||
if c.IdentifierBase() == ki.Identity.IdentifierBase() {
|
||||
return true
|
||||
}
|
||||
}
|
||||
return false
|
||||
})
|
||||
}()
|
||||
|
||||
changed := &WhatChanged{
|
||||
needsPagesAssembly: needsPagesAssemble,
|
||||
identitySet: make(identity.Identities),
|
||||
|
|
|
@ -33,6 +33,9 @@ const (
|
|||
|
||||
// GenghisKhan is an Identity everyone relates to.
|
||||
GenghisKhan = StringIdentity("__genghiskhan")
|
||||
|
||||
StructuralChangeAdd = StringIdentity("__structural_change_add")
|
||||
StructuralChangeRemove = StringIdentity("__structural_change_remove")
|
||||
)
|
||||
|
||||
var NopManager = new(nopManager)
|
||||
|
|
|
@ -95,6 +95,10 @@ func (c *Client) Concat(targetPath string, r resource.Resources) (resource.Resou
|
|||
}
|
||||
|
||||
idm := c.rs.Cfg.NewIdentityManager("concat")
|
||||
|
||||
// Re-create on structural changes.
|
||||
idm.AddIdentity(identity.StructuralChangeAdd, identity.StructuralChangeRemove)
|
||||
|
||||
// Add the concatenated resources as dependencies to the composite resource
|
||||
// so that we can track changes to the individual resources.
|
||||
idm.AddIdentityForEach(identity.ForEeachIdentityProviderFunc(
|
||||
|
|
Loading…
Add table
Reference in a new issue