Skip to content
/ recache Public
forked from jfbosch/recache

A high performance, in-memory, concurrency safe, transparent generic key / value cache - basically, it makes some stuff a lot faster

License

Notifications You must be signed in to change notification settings

bakasa/recache

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

95 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

ReCache

There are lots of caching options out there, but if you need something really simple that you can get running within 5 minutes, without any external dependencies, this might just be for you.

Why?

Because sometimes an in-memory cache is really the better option over an out-of-process, external cache service such as redis, memcached, and the like.

The skinny on Features

  • Simple
  • Fast
  • Generic
  • Transparent
  • Concurrency Safe
  • Max Size Indicator
  • Expiry Configuration
  • Auto Flush Stale Entries
  • Optionally Self-Refreshing

What does it do?

Quite simply, it caches any .Net value in memory against a key. This makes for really, really fast retrieval of that value when asking for it via the key. Both the key and the value are generic types, so the cache can be of any types you would like. e.g. Cache<int, string>, or Cache<string, MyFunkyClass>.

The cache is transparent. This means you set it up with a loader func, and when you hit the cache the first time, before it has been cached, it will invoke your provided loader func to go and fetch the value from anywhere, in any way you would like. It then caches the value, and returns it. Next time you hit the same key, providing the value is not yet stale as per your configured expiry timeout, you will get the same value back, but just much, much faster than the first time.

There is also a self-refreshing version. e.g. SelfRefreshingCache<int, string>. Once a key is in the cache, it will go and refresh itself on a configurable schedule. This means that the clients of the cache, after the very first load, will never see the performance hit again of fetching the fresh values through the loader func from the data source. It works on a generational model. While generation 2 is being fetch asynchronously in the background, the cache will keep serving up generation 1 to any clients that hit that key. The moment generation 2 is available, it replaces generation 1. In this way, the cache stays fresh, but there is no occasional delay as a client hits an expired cache entry.

It is built on ConcurrentDictionary<TKey, TValue>, so it is concurrency / thread safe.

Does it work?

Yes it works exceptionally well. We have it running in production, caching anything from lookups to search results to templates to types for dependency injection. It has given us incredible performance improvements. Once you start using it for one thing, you will probably find other places to plug it in as well.

How to get it?

Nuget, of course: https://www.nuget.org/packages/ReCache/

Samples

The below test can be viewed and run in the related Test project.

// This func is to fetch the value from the data source. In this case it 
// is just a random generator, but this could be anything like a web service or database.
// The parameter should be the same type as your cache key type, and the return value
// should be the same type as your cache's value type.
Func<int, string> loaderFunction = (key) =>
{
	int rnd = new Random().Next();
	return string.Format("value{0}", rnd);
};

var cacheOptions = new CacheOptions
{
	CacheItemExpiry = TimeSpan.FromSeconds(1),
	FlushInterval = TimeSpan.FromSeconds(5),
	MaximumCacheSizeIndicator = 100
};
var cache = new Cache<int, string>(
	cacheOptions, loaderFunction);


// Record the first hit values for 100 keys
var firstHitValues = new List<string>();
for (int key = 0; key < 100; key++)
{
	var value = cache.GetOrLoad(key);
	firstHitValues.Add(value);
}

// Assert that 50 subsequent hits return the same values for all 100 keys.
for (int i = 0; i < 50; i++)
{
	for (int key = 0; key < 100; key++)
	{
		var value = cache.GetOrLoad(key);
		Assert.AreEqual(firstHitValues[key], value);
	}
}

// Wait a while so that all cache entries expire as per the configured options
Thread.Sleep(1000);

// Assert that the next hit returns fresh values. i.e. not equal to the first hit values. 
for (int key = 0; key < 100; key++)
{
	var value = cache.GetOrLoad(key);
	Assert.AreNotEqual(firstHitValues[key], value);
}

About

A high performance, in-memory, concurrency safe, transparent generic key / value cache - basically, it makes some stuff a lot faster

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • C# 99.3%
  • Smalltalk 0.7%