Cache
Cache
This filter was introduced in MariaDB MaxScale 2.1.
Table of Contents
Overview
From MaxScale version 2.2.11 onwards, the cache filter is no longer considered experimental. The following changes to the default behaviour have also been made:
- The default value of
cached_data
is nowthread_specific
(used to beshared
). - The default value of
selects
is nowassume_cacheable
(used to beverify_cacheable
).
The cache filter is a simple cache that is capable of caching the result of SELECTs, so that subsequent identical SELECTs are served directly by MaxScale, without the queries being routed to any server.
By default the cache will be used and populated in the following circumstances:
- There is no explicit transaction active, that is, autocommit is used,
- there is an explicitly read-only transaction (that is,
START TRANSACTION READ ONLY
) active, or - there is a transaction active and no statement that modifies the database has been performed.
In practice, the last bullet point basically means that if a transaction has
been started with BEGIN
, START TRANSACTION
or START TRANSACTION READ
WRITE
, then the cache will be used and populated until the first UPDATE
,
INSERT
or DELETE
statement is encountered.
That is, in default mode the cache effectively causes the system to behave
as if the isolation level would be READ COMMITTED
, irrespective of what
the isolation level of the backends actually is.
The default behaviour can be altered using the configuration parameter cache_in_transactions.
By default it is assumed that all SELECT
statements are cacheable, which
means that also statements like SELECT LOCALTIME
are cached. Please check
selects for how to change the default behaviour.
Limitations
All of these limitations may be addressed in forthcoming releases.
Invalidation
Currently there is no cache invalidation, apart from time-to-live.
Prepared Statements
Resultsets of prepared statements are not cached.
Security
The cache is not aware of grants.
The implication is that unless the cache has been explicitly configured who the caching should apply to, the presence of the cache may provide a user with access to data he should not have access to.
Please read the section Security for more detailed information.
Configuration
The cache is simple to add to any existing service. However, some experimentation may be required in order to find the configuration settings that provide the maximum benefit.
[Cache] type=filter module=cache hard_ttl=30 soft_ttl=20 rules=... ... [Cached-Routing-Service] type=service ... filters=Cache
Each configured cache filter uses a storage of its own. That is, if there are two services, each configured with a specific cache filter, then, even if queries target the very same servers the cached data will not be shared.
Two services can use the same cache filter, but then either the services should use the very same servers or a completely different set of servers, where the used table names are different. Otherwise there can be unintended sharing.
Filter Parameters
The cache filter has no mandatory parameters but a range of optional ones.
Note that it is advisable to specify max_size
to prevent the cache from
using up all memory there is, in case there is very litte overlap among the
queries.
storage
The name of the module that provides the storage for the cache. That
module will be loaded and provided with the value of storage_options
as
argument. For instance:
storage=storage_inmemory
The default is storage_inmemory
.
See Storage for what storage modules are available.
storage_options
A comma separated list of arguments to be provided to the storage module,
specified in storage
, when it is loaded. Note that the needed arguments
depend upon the specific module. For instance,
storage_options=storage_specific_option1=value1,storage_specific_option2=value2
hard_ttl
Hard time to live; the maximum amount of time the cached
result is used before it is discarded and the result is fetched from the
backend (and cached). See also soft_ttl
below.
hard_ttl=60s
The default value is 0s
, which means no limit.
The duration can be specified as explained here. If no explicit unit has been specified, the value is interpreted as seconds in MaxScale 2.4. In subsequent versions a value without a unit may be rejected.
soft_ttl
Soft time to live; the amount of time - in seconds - the cached result is
used before it is refreshed from the server. When soft_ttl
has passed, the
result will be refreshed when the first client requests the value.
However, as long as hard_ttl
has not passed, all other clients requesting
the same value will use the result from the cache while it is being fetched
from the backend. That is, as long as soft_ttl
but not hard_ttl
has passed,
even if several clients request the same value at the same time, there will be
just one request to the backend.
soft_ttl=60s
The default value is 0
, which means no limit. If the value of soft_ttl
is
larger than hard_ttl
it will be adjusted down to the same value.
The duration can be specifed as explained here. If no explicit unit has been specified, the value is interpreted as seconds in MaxScale 2.4. In subsequent versions a value without a unit may be rejected.
max_resultset_rows
Specifies the maximum number of rows a resultset can have in order to be stored in the cache. A resultset larger than this, will not be stored.
max_resultset_rows=1000
The default value is 0
, which means no limit.
max_resultset_size
Specifies the maximum size of a resultset, for it to be stored in the cache. A resultset larger than this, will not be stored. The size can be specified as described here.
max_resultset_size=128Ki
The default value is 0
, which means no limit.
Note that the value of max_resultset_size
should not be larger than the
value of max_size
.
max_count
The maximum number of items the cache may contain. If the limit has been reached and a new item should be stored, then an older item will be evicted.
Note that if cached_data
is thread_specific
then this limit will be
applied to each cache separately. That is, if a thread specific cache
is used, then the total number of cached items is #threads * the value
of max_count
.
max_count=1000
The default value is 0
, which means no limit.
max_size
The maximum size the cache may occupy. If the limit has been reached and a new item should be stored, then some older item(s) will be evicted to make space. The size can be specified as described here.
Note that if cached_data
is thread_specific
then this limit will be
applied to each cache separately. That is, if a thread specific cache
is used, then the total size is #threads * the value of max_size
.
max_size=100Mi
The default value is 0
, which means no limit.
rules
Specifies the path of the file where the caching rules are stored. A relative path is interpreted relative to the data directory of MariaDB MaxScale.
rules=/path/to/rules-file
cached_data
An enumeration option specifying how data is shared between threads. The allowed values are:
shared
: The cached data is shared between threads. On the one hand it implies that there will be synchronization between threads, on the other hand that all threads will use data fetched by any thread.thread_specific
: The cached data is specific to a thread. On the one hand it implies that no synchonization is needed between threads, on the other hand that the very same data may be fetched and stored multiple times.
cached_data=shared
Default is thread_specific
. See max_count
and max_size
what implication
changing this setting to shared
has.
selects
An enumeration option specifying what approach the cache should take with
respect to SELECT
statements. The allowed values are:
assume_cacheable
: The cache can assume that allSELECT
statements, without exceptions, are cacheable.verify_cacheable
: The cache can not assume that allSELECT
statements are cacheable, but must verify that.
selects=verify_cacheable
Default is assume_cacheable
. In this case, all SELECT
statements are
assumed to be cacheable and will be parsed only if some specific rule
requires that.
If verify_cacheable
is specified, then all SELECT
statements will be
parsed and only those that are safe for caching - e.g. do not call any
non-cacheable functions or access any non-cacheable variables - will be
subject to caching.
If verify_cacheable
has been specified, the cache will not be used in
the following circumstances:
- The
SELECT
uses any of the following functions:BENCHMARK
,CONNECTION_ID
,CONVERT_TZ
,CURDATE
,CURRENT_DATE
,CURRENT_TIMESTAMP
,CURTIME
,DATABASE
,ENCRYPT
,FOUND_ROWS
,GET_LOCK
,IS_FREE_LOCK
,IS_USED_LOCK
,LAST_INSERT_ID
,LOAD_FILE
,LOCALTIME
,LOCALTIMESTAMP
,MASTER_POS_WAIT
,NOW
,RAND
,RELEASE_LOCK
,SESSION_USER
,SLEEP
,SYSDATE
,SYSTEM_USER
,UNIX_TIMESTAMP
,USER
,UUID
,UUID_SHORT
. - The
SELECT
accesses any of the following fields:CURRENT_DATE
,CURRENT_TIMESTAMP
,LOCALTIME
,LOCALTIMESTAMP
- The
SELECT
uses system or user variables.
Note that parsing all SELECT
statements carries a significant performance
cost. Please read performance for more details.
cache_in_transactions
An enumeration option specifying how the cache should behave when there are active transactions:
never
: When there is an active transaction, no data will be returned from the cache, but all requests will always be sent to the backend. The cache will be populated inside explicitly read-only transactions. Inside transactions that are not explicitly read-only, the cache will be populated until the first non-SELECT statement.read_only_transactions
: The cache will be used and populated inside explicitly read-only transactions. Inside transactions that are not explicitly read-only, the cache will be populated, but not used until the first non-SELECT statement.all_transactions
: The cache will be used and populated inside explicitly read-only transactions. Inside transactions that are not explicitly read-only, the cache will be used and populated until the first non-SELECT statement.
cache_in_transactions=never
Default is all_transactions
.
The values read_only_transactions
and all_transactions
have roughly the
same effect as changing the isolation level of the backend to read_committed
.
debug
An integer value, using which the level of debug logging made by the cache can be controlled. The value is actually a bitfield with different bits denoting different logging.
0
(0b00000
) No logging is made.1
(0b00001
) A matching rule is logged.2
(0b00010
) A non-matching rule is logged.4
(0b00100
) A decision to use data from the cache is logged.8
(0b01000
) A decision not to use data from the cache is logged.16
(0b10000
) Higher level decisions are logged.
Default is 0
. To log everything, give debug
a value of 31
.
debug=31
enabled
Specifies whether the cache is initially enabled or disabled.
enabled=false
Default is true
.
The value affects the initial state of the MaxScale user variables using which the behaviour of the cache can be modified at runtime. Please see Runtime Configuration for details.
Runtime Configuration
@maxscale.cache.populate
Using the variable @maxscale.cache.populate
it is possible to specify at
runtime whether the cache should be populated or not. Its initial value is
the value of the configuration parameter enabled
. That is, by default the
value is true
.
The purpose of this variable is make it possible for an application to decide statement by statement whether the cache should be populated.
SET @maxscale.cache.populate=true; SELECT a, b FROM tbl; SET @maxscale.cache.populate=false; SELECT a, b FROM tbl;
In the example above, the first SELECT
will always be sent to the
server and the result will be cached, provided the actual cache rules
specifies that it should be. The second SELECT
may be served from the
cache, depending on the value of @maxscale.cache.use
(and the cache
rules).
The value of @maxscale.cache.populate
can be queried
SELECT @maxscale.cache.populate;
but only after it has been explicitly set once.
@maxscale.cache.use
Using the variable @maxscale.cache.use
it is possible to specify at
runtime whether the cache should be used or not. Its initial value is
the value of the configuration parameter enabled
. That is, by default the
value is true
.
The purpose of this variable is make it possible for an application to decide statement by statement whether the cache should be used.
SET @maxscale.cache.use=true; SELECT a, b FROM tbl; SET @maxscale.cache.use=false; SELECT a, b FROM tbl;
The first SELECT
will be served from the cache, providing the rules
specify that the statement should be cached, the cache indeed contains
the result and the date is not stale (as specified by the TTL).
If the data is stale, the SELECT
will be sent to the server and
the cache entry will be updated, irrespective of the value of
@maxscale.cache.populate
.
If @maxscale.cache.use
is true
but the result is not found in the
cache, and the result is subsequently fetched from the server, the
result will not be added to the cache, unless
@maxscale.cache.populate
is also true
.
The value of @maxscale.cache.use
can be queried
SELECT @maxscale.cache.use;
but only after it has explicitly been set once.
@maxscale.cache.soft_ttl
Using the variable @maxscale.cache.soft_ttl
it is possible at runtime
to specify in seconds what soft ttl should be applied. Its initial
value is the value of the configuration parameter soft_ttl
. That is,
by default the value is 0.
The purpose of this variable is make it possible for an application to decide statement by statement what soft ttl should be applied.
set @maxscale.cache.soft_ttl=600; SELECT a, b FROM unimportant; set @maxscale.cache.soft_ttl=60; SELECT c, d FROM important;
When data is SELECT
ed from the unimportant table unimportant
, the data
will be returned from the cache provided it is no older than 10 minutes,
but when data is SELECT
ed from the important table important
, the
data will be returned from the cache provided it is no older than 1 minute.
Note that @maxscale.cache.hard_ttl
overrules @maxscale.cache.soft_ttl
in the sense that if the former is less that the latter, then soft ttl
will, when used, be adjusted down to the value of hard ttl.
The value of @maxscale.cache.soft_ttl
can be queried
SELECT @maxscale.cache.soft_ttl;
but only after it has explicitly been set once.
@maxscale.cache.hard_ttl
Using the variable @maxscale.cache.hard_ttl
it is possible at runtime
to specify in seconds what hard ttl should be applied. Its initial
value is the value of the configuration parameter hard_ttl
. That is,
by default the value is 0.
The purpose of this variable is make it possible for an application to decide statement by statement what hard ttl should be applied.
Note that as @maxscale.cache.hard_ttl
overrules @maxscale.cache.soft_ttl
,
is is important to ensure that the former is at least as large as the latter
and for best overall performance that it is larger.
set @maxscale.cache.soft_ttl=600, @maxscale.cache.hard_ttl=610; SELECT a, b FROM unimportant; set @maxscale.cache.soft_ttl=60, @maxscale.cache.hard_ttl=65; SELECT c, d FROM important;
The value of @maxscale.cache.hard_ttl
can be queried
SELECT @maxscale.cache.hard_ttl;
but only after it has explicitly been set once.
Client Driven Caching
With @maxscale.cache.populate
and @maxscale.cache.use
is it possible
to make the caching completely client driven.
Provide no rules
file, which means that all SELECT
statements are
subject to caching and that all users receive data from the cache. Set
the startup mode of the cache to disabled.
[TheCache] type=filter module=cache enabled=false
Now, in order to mark statements that should be cached, set
@maxscale.cache.populate
to true
, and perform those SELECT
s.
SET @maxscale.cache.populate=true; SELECT a, b FROM tbl1; SELECT c, d FROM tbl2; SELECT e, f FROM tbl3; SET @maxscale.cache.populate=false;
Note that those SELECT
s must return something in order for the
statement to be marked for caching.
After this, the value of @maxscale.cache.use
will decide whether
or not the cache is considered.
SET @maxscale.cache.use=true; SELECT a, b FROM tbl1; SET @maxscale.cache.use=false;
With @maxscale.cache.use
being true
, the cache is considered
and the result returned from there, if not stale. If it is stale,
the result is fetched from the server and the cached entry is updated.
By setting a very long TTL it is possible to prevent the cache from ever considering an entry to be stale and instead manually cause the cache to be updated when needed.
UPDATE tbl1 SET a = ...; SET @maxscale.cache.populate=true; SELECT a, b FROM tbl1; SET @maxscale.cache.populate=false;
Rules
The caching rules are expressed as a JSON object or as an array of JSON objects.
There are two decisions to be made regarding the caching; in what circumstances should data be stored to the cache and in what circumstances should the data in the cache be used.
Expressed in JSON this looks as follows
{ store: [ ... ], use: [ ... ] }
or, in case an array is used, as
[ { store: [ ... ], use: [ ... ] }, { ... } ]
The store
field specifies in what circumstances data should be stored to
the cache and the use
field specifies in what circumstances the data in
the cache should be used. In both cases, the value is a JSON array containg
objects.
If an array of rule objects is specified, then, when looking for a rule that
matches, the store
field of each object are evaluated in sequential order
until a match is found. Then, the use
field of that object is used when
deciding whether data in the cache should be used.
When to Store
By default, if no rules file have been provided or if the store
field is
missing from the object, the results of all queries will be stored to the
cache, subject to max_resultset_rows
and max_resultset_size
cache filter
parameters.
By providing a store
field in the JSON object, the decision whether to
store the result of a particular query to the cache can be controlled in
a more detailed manner. The decision to cache the results of a query can
depend upon
- the database,
- the table,
- the column, or
- the query itself.
Each entry in the store
array is an object containing three fields,
{ "attribute": <string>, "op": <string> "value": <string> }
where,
- the attribute can be
database
,table
,column
orquery
, - the op can be
=
,!=
,like
orunlike
, and - the value a string.
If op is =
or !=
then value is used as a string; if it is like
or unlike
, then value is interpreted as a pcre2 regular expression.
Note though that if attribute is database
, table
or column
, then
the string is interpreted as a name, where a dot .
denotes qualification
or scoping.
The objects in the store
array are processed in order. If the result
of a comparison is true, no further processing will be made and the
result of the query in question will be stored to the cache.
If the result of the comparison is false, then the next object is processed. The process continues until the array is exhausted. If there is no match, then the result of the query is not stored to the cache.
Note that as the query itself is used as the key, although the following queries
select * from db1.tbl
and
use db1; select * from tbl
target the same table and produce the same results, they will be cached separately. The same holds for queries like
select * from tbl where a = 2 and b = 3;
and
select * from tbl where b = 3 and a = 2;
as well. Although they conceptually are identical, there will be two cache entries.
Note that if a column has been specified in a rule, then a statement will match irrespective of where that particular column appears. For instance, if a rule specifies that the result of statements referring to the the column a should be cached, then the following statement will match
select a from tbl;
and so will
select b from tbl where a > 5;
Qualified Names
When using =
or !=
in the rule object in conjunction with database
,
table
and column
, the provided string is interpreted as a name, that is,
dot (.
) denotes qualification or scope.
In practice that means that if attribute is database
then value may
not contain a dot, if attribute is table
then value may contain one
dot, used for separating the database and table names respectively, and
if attribute is column
then value may contain one or two dots, used
for separating table and column names, or database, table and column names.
Note that if a qualified name is used as a value, then all parts of the
name must be available for a match. Currently Maria DB MaxScale may not
always be capable of deducing in what table a particular column is. If
that is the case, then a value like tbl.field
may not necessarily
be a match even if the field is field
and the table actually is tbl
.
Implication of the default database
If the rules concerns the database
, then only if the statement refers
to no specific database, will the default database be considered.
Regexp Matching
The string used for matching the regular expression contains as much information as there is available. For instance, in a situation like
use somedb; select fld from tbl;
the string matched against the regular expression will be somedb.tbl.fld
.
Examples
Cache all queries targeting a particular database.
{ "store": [ { "attribute": "database", "op": "=", "value": "db1" } ] }
Cache all queries not targeting a particular table
{ "store": [ { "attribute": "table", "op": "!=", "value": "tbl1" } ] }
That will exclude queries targeting table tbl1 irrespective of which database it is in. To exclude a table in a particular database, specify the table name using a qualified name.
{ "store": [ { "attribute": "table", "op": "!=", "value": "db1.tbl1" } ] }
Cache all queries containing a WHERE clause
{ "store": [ { "attribute": "query", "op": "like", "value": ".*WHERE.*" } ] }
Note that that will actually cause all queries that contain WHERE anywhere, to be cached.
When to Use
By default, if no rules file have been provided or if the use
field is
missing from the object, all users may be returned data from the cache.
By providing a use
field in the JSON object, the decision whether to use
data from the cache can be controlled in a more detailed manner. The decision
to use data from the cache can depend upon
- the user.
Each entry in the use
array is an object containing three fields,
{ "attribute": <string>, "op": <string> "value": <string> }
where,
- the attribute can be
user
, - the op can be
=
,!=
,like
orunlike
, and - the value a string.
If op is =
or !=
then value is interpreted as a MariaDB account
string, that is, %
means indicates wildcard, but if op is like
or
unlike
it is simply assumed value is a pcre2 regular expression.
For instance, the following are equivalent:
{ "attribute": "user", "op": "=", "value": "'bob'@'%'" } { "attribute": "user", "op": "like", "value": "bob@.*" }
Note that if op is =
or !=
then the usual assumptions apply,
that is, a value of bob
is equivalent with 'bob'@'%'
. If like
or unlike is used, then no assumptions apply, but the string is
used verbatim as a regular expression.
The objects in the use
array are processed in order. If the result
of a comparison is true, no further processing will be made and the
data in the cache will be used, subject to the value of ttl
.
If the result of the comparison is false, then the next object is processed. The process continues until the array is exhausted. If there is no match, then data in the cache will not be used.
Note that use
is relevant only if the query is subject to caching,
that is, if all queries are cached or if a query matches a particular
rule in the store
array.
Examples
Use data from the cache for all users except admin
(actually 'admin'@'%'
),
regardless of what host the admin
user comes from.
{ "use": [ { "attribute": "user", "op": "!=", "value": "admin" } ] }
Security
As the cache is not aware of grants, unless the cache has been explicitly configured who the caching should apply to, the presence of the cache may provide a user with access to data he should not have access to.
Suppose there is a table access
that the user alice has access to,
but the user bob does not. If bob tries to access the table, he will
get an error as reply:
MySQL [testdb]> select * from access; ERROR 1142 (42000): SELECT command denied to user 'bob'@'localhost' for table 'access'
If we now setup caching for the table, using the simplest possible rules file, bob will get access to data from the table, provided he executes a select identical with one alice has executed.
For instance, suppose the rules look as follows:
{ "store": [ { "attribute": "table", "op": "=", "value": "access" } ] }
If alice now queries the table, she will get the result, which also will be cached:
MySQL [testdb]> select * from access; +------+------+ | a | b | +------+------+ | 47 | 11 | +------+------+
If bob now executes the very same query, and the result is still in the cache, it will be returned to him.
MySQL [testdb]> select current_user(); +----------------+ | current_user() | +----------------+ | [email protected] | +----------------+ 1 row in set (0.00 sec) MySQL [testdb]> select * from access; +------+------+ | a | b | +------+------+ | 47 | 11 | +------+------+
That can be prevented, by explicitly declaring in the rules that the caching should be applied to alice only.
{ "store": [ { "attribute": "table", "op": "=", "value": "access" } ], "use": [ { "attribute": "user", "op": "=", "value": "'alice'@'%'" } ] }
With these rules in place, bob is again denied access, since queries
targeting the table access
will in his case not be served from the cache.
Storage
storage_inmemory
This simple storage module uses the standard memory allocator for storing the cached data.
storage=storage_inmemory
Example
In the following we define a cache MyCache that uses the cache storage module
storage_inmemory
and whose soft ttl is 30
seconds and whose hard ttl is
45
seconds. The cached data is shared between all threads and the maximum size
of the cached data is 50
mebibytes. The rules for the cache are in the file
cache_rules.json
.
Configuration
[MyCache] type=filter module=cache storage=storage_inmemory soft_ttl=30 hard_ttl=45 cached_data=shared max_size=50Mi rules=cache_rules.json [MyService] type=service ... filters=MyCache
cache_rules.json
The rules specify that the data of the table sbtest
should be cached.
{ "store": [ { "attribute": "table", "op": "=", "value": "sbtest" } ] }
Performance
Perhaps the most significant factor affecting the performance of the cache is
whether the statements need to be parsed or not. By default, all statements are
parsed in order to exclude SELECT
statements that use non-cacheable functions,
access non-cacheable variables or refer to system or user variables.
If it is known that no such statements are used or if it does not matter if the results are cached, that safety measure can be turned off. To do that, add the following line to the cache configuration:
[MyCache] ... selects=assume_cacheable
With that configuration, the cache itself will not cause the statements to be parsed.
But note that even with assume_cacheable
configured, a rule referring
specifically to a database, table or column will still cause the
statement to be parsed.
For instance, a simple rule like
{ "store": [ { "attribute": "database", "op": "=", "value": "db1" } ] }
cannot be fulfilled without parsing the statement.
If the rule is instead expressed using a regular expression
{ "store": [ { "attribute": "query", "op": "like", "value": "FROM db1\\..*" } ] }
then the statement will again not be parsed.
However, even if regular expression matching performance wise is cheaper than parsing, it still carries a cost. In the following is a table with numbers giving a rough picture of the relative cost of different approaches.
In the table, regexp match means that the cacheable statements were picked out using a rule like
{ "attribute": "query", "op": "like", "value": "FROM dbname" }
while exact match means that the cacheable statements were picked out using a rule like
{ "attribute": "database", "op": "=", "value": "dbname" }
The exact match rule requires all statements to be parsed.
Note that the qps figures are only indicative and that the difference under high load may be significantly greater.
selects |
Rule | qps |
---|---|---|
assume_cacheable |
none | 100 |
assume_cacheable |
regexp match | 98 |
assume_cacheable |
exact match | 60 |
verify_cacheable |
none | 60 |
verify_cacheable |
regexp match | 58 |
verify_cacheable |
exact match | 58 |
Summary
For maximum performance:
- Arrange the situation so that the default
selects=assume_cacheable
can be used, and use no rules. - If
selects=assume_cacheable
is used, use only regexp based rules. - If
selects=verify_cacheable
has been configured, non-regex based matching can be used.