diff options
author | Liang Li <liang.z.li@intel.com> | 2015-03-23 16:32:25 +0800 |
---|---|---|
committer | Juan Quintela <quintela@redhat.com> | 2015-05-07 18:31:53 +0200 |
commit | 68ae113646dc84637359472e89669e5547dc5ee3 (patch) | |
tree | 34bb9ec1d3f2628bcdc2a953a8bff0ede9d6d375 /arch_init.c | |
parent | 98f1138902195bd9ab8a753d0ee2cf2a0a88b6e8 (diff) | |
download | qemu-68ae113646dc84637359472e89669e5547dc5ee3.tar.gz qemu-68ae113646dc84637359472e89669e5547dc5ee3.tar.bz2 qemu-68ae113646dc84637359472e89669e5547dc5ee3.zip |
migration: Add the core code for decompression
Implement the core logic of multiple thread decompression,
the decompression can work now.
Signed-off-by: Liang Li <liang.z.li@intel.com>
Signed-off-by: Yang Zhang <yang.z.zhang@intel.com>
Signed-off-by: Juan Quintela <quintela@redhat.com>
Diffstat (limited to 'arch_init.c')
-rw-r--r-- | arch_init.c | 50 |
1 files changed, 48 insertions, 2 deletions
diff --git a/arch_init.c b/arch_init.c index 129c710c3c..539011d2c6 100644 --- a/arch_init.c +++ b/arch_init.c @@ -887,6 +887,13 @@ static inline void start_compression(CompressParam *param) qemu_mutex_unlock(¶m->mutex); } +static inline void start_decompression(DecompressParam *param) +{ + qemu_mutex_lock(¶m->mutex); + param->start = true; + qemu_cond_signal(¶m->cond); + qemu_mutex_unlock(¶m->mutex); +} static uint64_t bytes_transferred; @@ -1460,8 +1467,26 @@ void ram_handle_compressed(void *host, uint8_t ch, uint64_t size) static void *do_data_decompress(void *opaque) { + DecompressParam *param = opaque; + unsigned long pagesize; + while (!quit_decomp_thread) { - /* To be done */ + qemu_mutex_lock(¶m->mutex); + while (!param->start && !quit_decomp_thread) { + qemu_cond_wait(¶m->cond, ¶m->mutex); + pagesize = TARGET_PAGE_SIZE; + if (!quit_decomp_thread) { + /* uncompress() will return failed in some case, especially + * when the page is dirted when doing the compression, it's + * not a problem because the dirty page will be retransferred + * and uncompress() won't break the data in other pages. + */ + uncompress((Bytef *)param->des, &pagesize, + (const Bytef *)param->compbuf, param->len); + } + param->start = false; + } + qemu_mutex_unlock(¶m->mutex); } return NULL; @@ -1493,6 +1518,11 @@ void migrate_decompress_threads_join(void) quit_decomp_thread = true; thread_count = migrate_decompress_threads(); for (i = 0; i < thread_count; i++) { + qemu_mutex_lock(&decomp_param[i].mutex); + qemu_cond_signal(&decomp_param[i].cond); + qemu_mutex_unlock(&decomp_param[i].mutex); + } + for (i = 0; i < thread_count; i++) { qemu_thread_join(decompress_threads + i); qemu_mutex_destroy(&decomp_param[i].mutex); qemu_cond_destroy(&decomp_param[i].cond); @@ -1509,7 +1539,23 @@ void migrate_decompress_threads_join(void) static void decompress_data_with_multi_threads(uint8_t *compbuf, void *host, int len) { - /* To be done */ + int idx, thread_count; + + thread_count = migrate_decompress_threads(); + while (true) { + for (idx = 0; idx < thread_count; idx++) { + if (!decomp_param[idx].start) { + memcpy(decomp_param[idx].compbuf, compbuf, len); + decomp_param[idx].des = host; + decomp_param[idx].len = len; + start_decompression(&decomp_param[idx]); + break; + } + } + if (idx < thread_count) { + break; + } + } } static int ram_load(QEMUFile *f, void *opaque, int version_id) |