X-Git-Url: https://git.saurik.com/redis.git/blobdiff_plain/5e2022bb484791b1b3099b7ce304f02b1f4a3212..c44d3b56df4ce4cae1c2e6db6397eaab9651ff7a:/redis.conf diff --git a/redis.conf b/redis.conf index fc5ac59d..c9ff26cf 100644 --- a/redis.conf +++ b/redis.conf @@ -133,7 +133,7 @@ dir ./ # happens this is the preferred way to run Redis. If instead you care a lot # about your data and don't want to that a single record can get lost you should # enable the append only mode: when this mode is enabled Redis will append -# every write operation received in the file appendonly.log. This file will +# every write operation received in the file appendonly.aof. This file will # be read on startup in order to rebuild the full dataset in memory. # # Note that you can have both the async dumps and the append only file if you @@ -141,7 +141,7 @@ dir ./ # Still if append only mode is enabled Redis will load the data from the # log file at startup ignoring the dump.rdb file. # -# The name of the append only file is "appendonly.log" +# The name of the append only file is "appendonly.aof" # # IMPORTANT: Check the BGREWRITEAOF to check how to rewrite the append # log file in background when it gets too big. @@ -255,23 +255,6 @@ vm-max-threads 4 # in terms of number of queries per second. Use 'yes' if unsure. glueoutputbuf yes -# Use object sharing. Can save a lot of memory if you have many common -# string in your dataset, but performs lookups against the shared objects -# pool so it uses more CPU and can be a bit slower. Usually it's a good -# idea. -# -# When object sharing is enabled (shareobjects yes) you can use -# shareobjectspoolsize to control the size of the pool used in order to try -# object sharing. A bigger pool size will lead to better sharing capabilities. -# In general you want this value to be at least the double of the number of -# very common strings you have in your dataset. -# -# WARNING: object sharing is experimental, don't enable this feature -# in production before of Redis 1.0-stable. Still please try this feature in -# your development environment so that we can test it better. -shareobjects no -shareobjectspoolsize 1024 - # Hashes are encoded in a special way (much more memory efficient) when they # have at max a given numer of elements, and the biggest element does not # exceed a given threshold. You can configure this limits with the following @@ -279,6 +262,26 @@ shareobjectspoolsize 1024 hash-max-zipmap-entries 64 hash-max-zipmap-value 512 +# Active rehashing uses 1 millisecond every 100 milliseconds of CPU time in +# order to help rehashing the main Redis hash table (the one mapping top-level +# keys to values). The hash table implementation redis uses (see dict.c) +# performs a lazy rehashing: the more operation you run into an hash table +# that is rhashing, the more rehashing "steps" are performed, so if the +# server is idle the rehashing is never complete and some more memory is used +# by the hash table. +# +# The default is to use this millisecond 10 times every second in order to +# active rehashing the main dictionaries, freeing memory when possible. +# +# If unsure: +# use "activerehashing no" if you have hard latency requirements and it is +# not a good thing in your environment that Redis can reply form time to time +# to queries with 2 milliseconds delay. +# +# use "activerehashing yes" if you don't have such hard requirements but +# want to free memory asap when possible. +activerehashing yes + ################################## INCLUDES ################################### # Include one or more other config files here. This is useful if you