d9d6e2b00e
How do you decide you need to purge your cache? Relying on runtime.ReadMemStats sucks for two reasons. First, it's a stop-the-world call, which is pretty bad in general and down right stupid for a supposedly concurrent-focused package. Secondly, it only tells you the total memory usage, but most time you really want to limit the amount of memory the cache itself uses. Since there's no great way to determine the size of an object, that means users need to supply the size. One way is to make it so that any cached item satisfies a simple interface which exposes a Size() method. With this, we can track how much memory is set put and a delete releases. But it's hard for consumers to know how much memory they're taking when storing complex object (the entire point of an in-process cache is to avoid having to serialize the data). Since any Size() is bound to be a rough guess, we can simplify the entire thing by evicting based on # of items. This works really bad when items vary greatly in size (an HTTP cache), but in a lot of other cases it works great. Furthermore, even for an HTTP cache, given enough values, it should average out in most cases. Whatever. This improve performance and should improve the usability of the cache. It is a pretty big breaking change though.
62 lines
1.6 KiB
Go
62 lines
1.6 KiB
Go
package ccache
|
|
|
|
import (
|
|
"github.com/karlseguin/gspec"
|
|
"strconv"
|
|
"testing"
|
|
"time"
|
|
)
|
|
|
|
func TestCacheGCsTheOldestItems(t *testing.T) {
|
|
spec := gspec.New(t)
|
|
cache := New(Configure().ItemsToPrune(10))
|
|
for i := 0; i < 500; i++ {
|
|
cache.Set(strconv.Itoa(i), i, time.Minute)
|
|
}
|
|
cache.gc()
|
|
spec.Expect(cache.Get("9")).ToBeNil()
|
|
spec.Expect(cache.Get("10").(int)).ToEqual(10)
|
|
}
|
|
|
|
func TestCachePromotedItemsDontGetPruned(t *testing.T) {
|
|
spec := gspec.New(t)
|
|
cache := New(Configure().ItemsToPrune(10).GetsPerPromote(1))
|
|
for i := 0; i < 500; i++ {
|
|
cache.Set(strconv.Itoa(i), i, time.Minute)
|
|
}
|
|
time.Sleep(time.Millisecond * 10) //run the worker once to init the list
|
|
cache.Get("9")
|
|
time.Sleep(time.Millisecond * 10)
|
|
cache.gc()
|
|
spec.Expect(cache.Get("9").(int)).ToEqual(9)
|
|
spec.Expect(cache.Get("10")).ToBeNil()
|
|
spec.Expect(cache.Get("11").(int)).ToEqual(11)
|
|
}
|
|
|
|
func TestCacheTrackerDoesNotCleanupHeldInstance(t *testing.T) {
|
|
spec := gspec.New(t)
|
|
cache := New(Configure().ItemsToPrune(10).Track())
|
|
for i := 0; i < 10; i++ {
|
|
cache.Set(strconv.Itoa(i), i, time.Minute)
|
|
}
|
|
item := cache.TrackingGet("0")
|
|
time.Sleep(time.Millisecond * 10)
|
|
cache.gc()
|
|
spec.Expect(cache.Get("0").(int)).ToEqual(0)
|
|
spec.Expect(cache.Get("1")).ToBeNil()
|
|
item.Release()
|
|
cache.gc()
|
|
spec.Expect(cache.Get("0")).ToBeNil()
|
|
}
|
|
|
|
func TestCacheRemovesOldestItemWhenFull(t *testing.T) {
|
|
spec := gspec.New(t)
|
|
cache := New(Configure().MaxItems(5).ItemsToPrune(1))
|
|
for i := 0; i < 7; i++ {
|
|
cache.Set(strconv.Itoa(i), i, time.Minute)
|
|
}
|
|
time.Sleep(time.Millisecond * 10)
|
|
spec.Expect(cache.Get("0")).ToBeNil()
|
|
spec.Expect(cache.Get("1")).ToBeNil()
|
|
spec.Expect(cache.Get("2").(int)).ToEqual(2)
|
|
}
|