X-Git-Url: https://git.saurik.com/redis.git/blobdiff_plain/9d65a1bbae9e59269472e8067cb2fff1e1cce24c..7493d2a0325fe33dc75317bfedf9b1c1e5b0d0b5:/TODO diff --git a/TODO b/TODO index 3d1059df..580adda6 100644 --- a/TODO +++ b/TODO @@ -1,56 +1,60 @@ -Redis TODO and Roadmap +Redis TODO +---------- -VERSION 1.1 TODO (Zsets, Integer encoding, Append only journal) -=============================================================== +WARNING: are you a possible Redis contributor? + Before implementing what is listed what is listed in this file + please drop a message in the Redis google group or chat with + antirez or pietern on irc.freenode.org #redis to check if the work + is already in progress and if the feature is still interesting for + us, and *how* exactly this can be implemented to have good changes + of a merge. Otherwise it is probably wasted work! Thank you -Most of the features already implemented for this release. The following is a list of the missing things in order to release the first beta tar.gz: +DISKSTORE TODO +============== -* For now only the last argument gets integer encoded, so make sure that: 1) every multi bulk command implemented will have the last arg that is indeed a value, and not used otherwise. 2) to explicitly call the function to encode the object in MSET and other commands where there are multiple "values". -* Man pages for SRANDMEMBER, missing Z-commands, ... -* Use strcoll() to compare objects in sorted sets, like it already happens for SORT. -* Write docs for the "STORE" operaiton of SORT. Link to the article about SORT by written by defunkt. -* Append only mode: testing and a command to rebuild the log from scratch. -* ZRANGEBYSCORE test, ZRANGEBYSCORE LIMIT option. -* Sorted sets infinity tests. -* Instead to do the lamest thing of using getDecodedObject only if the encoding - of the object != REDIS_ENCODING_RAW, hack getDecodedObject in order to just - increment the reference count if the object is already RAW-encoded. +* Check that 00/00 and ff/ff exist at startup, otherwise exit with error. +* Implement sync flush option, where data is written synchronously on disk when a command is executed. +* Implement MULTI/EXEC as transaction abstract API to diskstore.c, with transaction_start, transaction_end, and a journal to recover. +* Stop BGSAVE thread on shutdown and any other condition where the child is killed during normal bgsave. +* Fix RANDOMKEY to really do something interesting +* Fix DBSIZE to really do something interesting +* Add a DEBUG command to check if an entry is or not in memory currently -VERSION 1.2 TODO (Hash type) -============================ +* dscache.c near 236, kobj = createStringObject... we could use static obj. -* Hashes (HSET, HGET, HEXISTS, HLEN, ...). -* An utility able to export an .rdb file into a text-only JSON dump, we can't live anymore without such a tool. Probably an extension to redis-cli. +APPEND ONLY FILE +================ -VERSION 1.3 TODO (Virtual memory) -================================= +* in AOF rewirte use HMSET to rewrite small hashes instead of multiple calls + to HSET. -* Redis Virtual Memory for datasets bigger than RAM (http://groups.google.com/group/redis-db/msg/752997c7b38553cd) +OPTIMIZATIONS +============= -VERSION 1.4 TODO (Fault tollerant sharding) -=========================================== +* Avoid COW due to incrementing the dict iterators counter. +* SORT: Don't copy the list into a vector when BY argument is constant. +* Write the hash table size of every db in the dump, so that Redis can resize the hash table just one time when loading a big DB. +* Read-only mode for slaves. -* Redis-cluster, a fast intermediate layer (proxy) that implements consistent hashing and fault tollerant nodes handling. +REPORTING +========= -Interesting readings about this: +* Better INFO output with sections. - - http://ayende.com/Blog/archive/2009/04/06/designing-rhino-dht-a-fault-tolerant-dynamically-distributed-hash.aspx +RANDOM +====== -VERSION 1.5 TODO (Optimizations and latency) -============================================ +* Clients should be closed as far as the output buffer list is bigger than a given number of elements (configurable in redis.conf) +* Should the redis default configuration, and the default redis.conf, just bind 127.0.0.1? -* Lower the CPU usage. -* Lower the RAM usage everywhere possible. -* Use epool and alike to rewrite ae.c for Linux and other platforms suppporting fater-than-select() mutiplexing APIs. -* Implement an UDP interface for low-latency GET/SET operations. +KNOWN BUGS +========== -SHORT/LONG TERM RANDOM TODO ITEMS -================================= +* What happens in the following scenario: + 1) We are reading an AOF file. + 2) SETEX FOO 5 BAR + 3) APPEND FOO ZAP + What happens if between 1 and 2 for some reason (system under huge load + or alike) too many time passes? We should prevent expires while the + AOF is loading. - * SORT: Don't copy the list into a vector when BY argument is constant. - * Write the hash table size of every db in the dump, so that Redis can resize the hash table just one time when loading a big DB. - * LOCK / TRYLOCK / UNLOCK as described many times in the google group - * Replication automated tests - * BYTEDARRAY type - * zmalloc() should avoid to add a private header for archs where there is some other kind of libc-specific way to get the size of a malloced block. - * Read-only mode.