summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorantirez <antirez@gmail.com>2011-01-05 12:34:14 +0100
committerantirez <antirez@gmail.com>2011-01-05 12:34:14 +0100
commitf771dc23a0bce29d7665f001b4c2dfba3e0dc8a5 (patch)
tree855b08ee961cb153ad70ab0a0442acb076f959bc
parent6eaad66373982aa9c1c0ff6b584c542670548d0e (diff)
downloadredis-f771dc23a0bce29d7665f001b4c2dfba3e0dc8a5.tar.gz
IO performances greatly improved under high writes load
-rw-r--r--src/dscache.c189
-rw-r--r--src/redis.h4
2 files changed, 113 insertions, 80 deletions
diff --git a/src/dscache.c b/src/dscache.c
index 901c70f07..4ebca7087 100644
--- a/src/dscache.c
+++ b/src/dscache.c
@@ -107,7 +107,8 @@
*/
void spawnIOThread(void);
-int cacheScheduleIOPushJobs(int onlyloads);
+int cacheScheduleIOPushJobs(int flags);
+int processActiveIOJobs(int max);
/* =================== Virtual Memory - Blocking Side ====================== */
@@ -216,9 +217,9 @@ int cacheFreeOneEntry(void) {
* otherwise we'll use an infinite amount of memory if changes to
* the dataset are faster than I/O */
if (listLength(server.cache_io_queue) > 0) {
- cacheScheduleIOPushJobs(0);
- waitEmptyIOJobsQueue();
- processAllPendingIOJobs();
+ redisLog(REDIS_DEBUG,"--- Busy waiting IO to reclaim memory");
+ cacheScheduleIOPushJobs(REDIS_IO_ASAP);
+ processActiveIOJobs(1);
return REDIS_OK;
}
/* Nothing to free at all... */
@@ -484,62 +485,86 @@ void spawnIOThread(void) {
server.io_active_threads++;
}
-/* Wait that all the pending IO Jobs are processed */
-void waitEmptyIOJobsQueue(void) {
- while(1) {
+/* Wait that up to 'max' pending IO Jobs are processed by the I/O thread.
+ * From our point of view an IO job processed means that the count of
+ * server.io_processed must increase by one.
+ *
+ * If max is -1, all the pending IO jobs will be processed.
+ *
+ * Returns the number of IO jobs processed.
+ *
+ * NOTE: while this may appear like a busy loop, we are actually blocked
+ * by IO since we continuously acquire/release the IO lock. */
+int processActiveIOJobs(int max) {
+ int processed = 0;
+
+ while(max == -1 || max > 0) {
int io_processed_len;
lockThreadedIO();
if (listLength(server.io_newjobs) == 0 &&
listLength(server.io_processing) == 0)
{
+ /* There is nothing more to process */
unlockThreadedIO();
- return;
+ break;
}
+
+#if 0
/* If there are new jobs we need to signal the thread to
* process the next one. */
redisLog(REDIS_DEBUG,"waitEmptyIOJobsQueue: new %d, processing %d",
listLength(server.io_newjobs),
listLength(server.io_processing));
- /* FIXME: signal or not?
if (listLength(server.io_newjobs)) {
pthread_cond_signal(&server.io_condvar);
}
- */
- /* While waiting for empty jobs queue condition we post-process some
- * finshed job, as I/O threads may be hanging trying to write against
- * the io_ready_pipe_write FD but there are so much pending jobs that
- * it's blocking. */
+#endif
+
+ /* Check if we can process some finished job */
io_processed_len = listLength(server.io_processed);
unlockThreadedIO();
if (io_processed_len) {
vmThreadedIOCompletedJob(NULL,server.io_ready_pipe_read,
(void*)0xdeadbeef,0);
- /* FIXME: probably wiser to drop this sleeps. Anyway
- * the contention on the IO thread will avoid we to loop
- * too fast here. */
- usleep(1000); /* 1 millisecond */
- } else {
- /* FIXME: same as fixme above. */
- usleep(10000); /* 10 milliseconds */
+ processed++;
+ if (max != -1) max--;
}
}
+ return processed;
}
-/* Process all the IO Jobs already completed by threads but still waiting
- * processing from the main thread. */
-void processAllPendingIOJobs(void) {
- while(1) {
+void waitEmptyIOJobsQueue(void) {
+ processActiveIOJobs(-1);
+}
+
+/* Process up to 'max' IO Jobs already completed by threads but still waiting
+ * processing from the main thread.
+ *
+ * If max == -1 all the pending jobs are processed.
+ *
+ * The number of processed jobs is returned. */
+int processPendingIOJobs(int max) {
+ int processed = 0;
+
+ while(max == -1 || max > 0) {
int io_processed_len;
lockThreadedIO();
io_processed_len = listLength(server.io_processed);
unlockThreadedIO();
- if (io_processed_len == 0) return;
+ if (io_processed_len == 0) break;
vmThreadedIOCompletedJob(NULL,server.io_ready_pipe_read,
(void*)0xdeadbeef,0);
+ if (max != -1) max--;
+ processed++;
}
+ return processed;
+}
+
+void processAllPendingIOJobs(void) {
+ processPendingIOJobs(-1);
}
/* This function must be called while with threaded IO locked */
@@ -665,7 +690,7 @@ void cacheScheduleIO(redisDb *db, robj *key, int type) {
* in queue for the same key. */
if (type == REDIS_IO_LOAD && !(flags & REDIS_IO_SAVE)) {
listAddNodeHead(server.cache_io_queue, op);
- cacheScheduleIOPushJobs(1);
+ cacheScheduleIOPushJobs(REDIS_IO_ONLYLOADS);
} else {
/* FIXME: probably when this happens we want to at least move
* the write job about this queue on top, and set the creation time
@@ -675,13 +700,19 @@ void cacheScheduleIO(redisDb *db, robj *key, int type) {
}
/* Push scheduled IO operations into IO Jobs that the IO thread can process.
- * If 'onlyloads' is true only IO_LOAD jobs are processed: this is useful
- * since it's save to push LOAD IO jobs from any place of the code, while
+ *
+ * If flags include REDIS_IO_ONLYLOADS only load jobs are processed:this is
+ * useful since it's safe to push LOAD IO jobs from any place of the code, while
* SAVE io jobs should never be pushed while we are processing a command
* (not protected by lookupKey() that will block on keys in IO_SAVEINPROG
- * state. */
+ * state.
+ *
+ * The REDIS_IO_ASAP flag tells the function to don't wait for the IO job
+ * scheduled completion time, but just do the operation ASAP. This is useful
+ * when we need to reclaim memory from the IO queue.
+ */
#define MAX_IO_JOBS_QUEUE 100
-int cacheScheduleIOPushJobs(int onlyloads) {
+int cacheScheduleIOPushJobs(int flags) {
time_t now = time(NULL);
listNode *ln;
int jobs, topush = 0, pushed = 0;
@@ -699,66 +730,64 @@ int cacheScheduleIOPushJobs(int onlyloads) {
while((ln = listFirst(server.cache_io_queue)) != NULL) {
ioop *op = ln->value;
+ struct dictEntry *de;
+ robj *val;
if (!topush) break;
topush--;
- if (op->type == REDIS_IO_LOAD ||
- (!onlyloads && (now - op->ctime) >= server.cache_flush_delay))
+ if (op->type != REDIS_IO_LOAD && flags & REDIS_IO_ONLYLOADS) break;
+
+ if (!(flags & REDIS_IO_ASAP) &&
+ (now - op->ctime) < server.cache_flush_delay) break;
+
+ /* Don't add a SAVE job in the IO thread queue if there is already
+ * a save in progress for the same key. */
+ if (op->type == REDIS_IO_SAVE &&
+ cacheScheduleIOGetFlags(op->db,op->key) & REDIS_IO_SAVEINPROG)
{
- struct dictEntry *de;
- robj *val;
-
- /* Don't add a SAVE job in queue if there is already
- * a save in progress for the same key. */
- if (op->type == REDIS_IO_SAVE &&
- cacheScheduleIOGetFlags(op->db,op->key) & REDIS_IO_SAVEINPROG)
- {
- /* Move the operation at the end of the list of there
- * are other operations. Otherwise break, nothing to do
- * here. */
- if (listLength(server.cache_io_queue) > 1) {
- listDelNode(server.cache_io_queue,ln);
- listAddNodeTail(server.cache_io_queue,op);
- continue;
- } else {
- break;
- }
+ /* Move the operation at the end of the list if there
+ * are other operations, so we can try to process the next one.
+ * Otherwise break, nothing to do here. */
+ if (listLength(server.cache_io_queue) > 1) {
+ listDelNode(server.cache_io_queue,ln);
+ listAddNodeTail(server.cache_io_queue,op);
+ continue;
+ } else {
+ break;
}
+ }
- redisLog(REDIS_DEBUG,"Creating IO %s Job for key %s",
- op->type == REDIS_IO_LOAD ? "load" : "save", op->key->ptr);
+ redisLog(REDIS_DEBUG,"Creating IO %s Job for key %s",
+ op->type == REDIS_IO_LOAD ? "load" : "save", op->key->ptr);
- if (op->type == REDIS_IO_LOAD) {
- dsCreateIOJob(REDIS_IOJOB_LOAD,op->db,op->key,NULL);
+ if (op->type == REDIS_IO_LOAD) {
+ dsCreateIOJob(REDIS_IOJOB_LOAD,op->db,op->key,NULL);
+ } else {
+ /* Lookup the key, in order to put the current value in the IO
+ * Job. Otherwise if the key does not exists we schedule a disk
+ * store delete operation, setting the value to NULL. */
+ de = dictFind(op->db->dict,op->key->ptr);
+ if (de) {
+ val = dictGetEntryVal(de);
} else {
- /* Lookup the key, in order to put the current value in the IO
- * Job. Otherwise if the key does not exists we schedule a disk
- * store delete operation, setting the value to NULL. */
- de = dictFind(op->db->dict,op->key->ptr);
- if (de) {
- val = dictGetEntryVal(de);
- } else {
- /* Setting the value to NULL tells the IO thread to delete
- * the key on disk. */
- val = NULL;
- }
- dsCreateIOJob(REDIS_IOJOB_SAVE,op->db,op->key,val);
+ /* Setting the value to NULL tells the IO thread to delete
+ * the key on disk. */
+ val = NULL;
}
- /* Mark the operation as in progress. */
- cacheScheduleIODelFlag(op->db,op->key,op->type);
- cacheScheduleIOAddFlag(op->db,op->key,
- (op->type == REDIS_IO_LOAD) ? REDIS_IO_LOADINPROG :
- REDIS_IO_SAVEINPROG);
- /* Finally remove the operation from the queue.
- * But we'll have trace of it in the hash table. */
- listDelNode(server.cache_io_queue,ln);
- decrRefCount(op->key);
- zfree(op);
- pushed++;
- } else {
- break; /* too early */
+ dsCreateIOJob(REDIS_IOJOB_SAVE,op->db,op->key,val);
}
+ /* Mark the operation as in progress. */
+ cacheScheduleIODelFlag(op->db,op->key,op->type);
+ cacheScheduleIOAddFlag(op->db,op->key,
+ (op->type == REDIS_IO_LOAD) ? REDIS_IO_LOADINPROG :
+ REDIS_IO_SAVEINPROG);
+ /* Finally remove the operation from the queue.
+ * But we'll have trace of it in the hash table. */
+ listDelNode(server.cache_io_queue,ln);
+ decrRefCount(op->key);
+ zfree(op);
+ pushed++;
}
return pushed;
}
diff --git a/src/redis.h b/src/redis.h
index b15bb3703..54d82cb6b 100644
--- a/src/redis.h
+++ b/src/redis.h
@@ -125,6 +125,10 @@
#define REDIS_IO_LOADINPROG 4
#define REDIS_IO_SAVEINPROG 8
+/* Generic IO flags */
+#define REDIS_IO_ONLYLOADS 1
+#define REDIS_IO_ASAP 2
+
#define REDIS_MAX_COMPLETED_JOBS_PROCESSED 1
#define REDIS_THREAD_STACK_SIZE (1024*1024*4)