Cache guide
The Cache module provides your Cherrycake application with a standardized interface to implement caching and shared memory mechanisms into your App by connecting to multiple external cache providers.
Cache providers
To use cache in your application, you must first define a cache provider. Each cache provider connects your app with a different cache mechanism or server, that means your app can make use of different caches at the same time.
Cache providers are configured in the Cache configuration file /config/Cache.config.php
. For example, if you want to use a simple but fast APCu cache provider, your Cache configuration file would look like this:
Note we called our cache provider
fast
You can configure more than one cache providers. For example, let's add a Redis cache provider too, called huge
:
Since some functionalities of Cherrycake make use of caching mechanisms, there is a default cache provider called
engine
that uses an APCu cache provider. This provider is always defined, no matter what you setup in yourCache.config.php
file.
Modules that depend on Cache
Some Core modules make use of Cache by their own, like Database and Patterns. Those modules always accept a configuration key to tell them the name of the cache provider to use, as defined in your Cache.config.php
How long cached objects stay in cache?
The data you store in cache, as well as the objects cached by modules like Database and Patterns, are normally persistent between requests, and it's the job of the cache mechanism to keep them there for as long as possible, or until their TTL expiration time arrives.
However, you can not rely on a cache or shared memory system as a persistent way of storing information. Generally, cached objects are deleted when the server restarts, when it runs out of memory, or if it implements certain cache eviction policies, like removing all cached objects that are too old, or that haven't been accessed too much.
To store information in a persistent way, use Database.
Other ways of using shared memory
Some specific cache providers like Redis implement other useful ways of working with shared memory, like Lists, Queues and Pools.
This methods are specially suited for high performance operations like the storage of events in a high traffic scenario, or the intermediate storage of data that needs to be accessed extremely fast, lots of times per second.
Last updated