X-Git-Url: https://gerrit.opnfv.org/gerrit/gitweb?a=blobdiff_plain;f=src%2Fceph%2Fsrc%2Fcompressor%2FAsyncCompressor.cc;fp=src%2Fceph%2Fsrc%2Fcompressor%2FAsyncCompressor.cc;h=0000000000000000000000000000000000000000;hb=7da45d65be36d36b880cc55c5036e96c24b53f00;hp=d4a77686b534f365897c1f4e07a65e10b0686ca5;hpb=691462d09d0987b47e112d6ee8740375df3c51b2;p=stor4nfv.git diff --git a/src/ceph/src/compressor/AsyncCompressor.cc b/src/ceph/src/compressor/AsyncCompressor.cc deleted file mode 100644 index d4a7768..0000000 --- a/src/ceph/src/compressor/AsyncCompressor.cc +++ /dev/null @@ -1,155 +0,0 @@ -// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*- -// vim: ts=8 sw=2 smarttab -/* - * Ceph - scalable distributed file system - * - * Copyright (C) 2015 Haomai Wang - * - * This is free software; you can redistribute it and/or - * modify it under the terms of the GNU Lesser General Public - * License version 2.1, as published by the Free Software - * Foundation. See file COPYING. - * - */ - -#include "common/dout.h" -#include "common/errno.h" -#include "AsyncCompressor.h" - -#define dout_subsys ceph_subsys_compressor -#undef dout_prefix -#define dout_prefix *_dout << "compressor " - -AsyncCompressor::AsyncCompressor(CephContext *c): - compressor(Compressor::create(c, c->_conf->async_compressor_type)), cct(c), - compress_tp(cct, "AsyncCompressor::compressor_tp", "tp_async_compr", cct->_conf->async_compressor_threads, "async_compressor_threads"), - job_lock("AsyncCompressor::job_lock"), - compress_wq(this, c->_conf->async_compressor_thread_timeout, c->_conf->async_compressor_thread_suicide_timeout, &compress_tp) { -} - -void AsyncCompressor::init() -{ - ldout(cct, 10) << __func__ << dendl; - compress_tp.start(); -} - -void AsyncCompressor::terminate() -{ - ldout(cct, 10) << __func__ << dendl; - compress_tp.stop(); -} - -uint64_t AsyncCompressor::async_compress(bufferlist &data) -{ - uint64_t id = ++job_id; - pair::iterator, bool> it; - { - Mutex::Locker l(job_lock); - it = jobs.insert(make_pair(id, Job(id, true))); - it.first->second.data = data; - } - compress_wq.queue(&it.first->second); - ldout(cct, 10) << __func__ << " insert async compress job id=" << id << dendl; - return id; -} - -uint64_t AsyncCompressor::async_decompress(bufferlist &data) -{ - uint64_t id = ++job_id; - pair::iterator, bool> it; - { - Mutex::Locker l(job_lock); - it = jobs.insert(make_pair(id, Job(id, false))); - it.first->second.data = data; - } - compress_wq.queue(&it.first->second); - ldout(cct, 10) << __func__ << " insert async decompress job id=" << id << dendl; - return id; -} - -int AsyncCompressor::get_compress_data(uint64_t compress_id, bufferlist &data, bool blocking, bool *finished) -{ - assert(finished); - Mutex::Locker l(job_lock); - unordered_map::iterator it = jobs.find(compress_id); - if (it == jobs.end() || !it->second.is_compress) { - ldout(cct, 10) << __func__ << " missing to get compress job id=" << compress_id << dendl; - return -ENOENT; - } - - retry: - auto status = it->second.status.load(); - if (status == status_t::DONE) { - ldout(cct, 20) << __func__ << " successfully getting compressed data, job id=" << compress_id << dendl; - *finished = true; - data.swap(it->second.data); - jobs.erase(it); - } else if (status == status_t::ERROR) { - ldout(cct, 20) << __func__ << " compressed data failed, job id=" << compress_id << dendl; - jobs.erase(it); - return -EIO; - } else if (blocking) { - auto expected = status_t::WAIT; - if (it->second.status.compare_exchange_strong(expected, status_t::DONE)) { - ldout(cct, 10) << __func__ << " compress job id=" << compress_id << " hasn't finished, abort!"<< dendl; - if (compressor->compress(it->second.data, data)) { - ldout(cct, 1) << __func__ << " compress job id=" << compress_id << " failed!"<< dendl; - it->second.status = status_t::ERROR; - return -EIO; - } - *finished = true; - } else { - job_lock.Unlock(); - usleep(1000); - job_lock.Lock(); - goto retry; - } - } else { - ldout(cct, 10) << __func__ << " compress job id=" << compress_id << " hasn't finished."<< dendl; - *finished = false; - } - return 0; -} - -int AsyncCompressor::get_decompress_data(uint64_t decompress_id, bufferlist &data, bool blocking, bool *finished) -{ - assert(finished); - Mutex::Locker l(job_lock); - unordered_map::iterator it = jobs.find(decompress_id); - if (it == jobs.end() || it->second.is_compress) { - ldout(cct, 10) << __func__ << " missing to get decompress job id=" << decompress_id << dendl; - return -ENOENT; - } - retry: - auto status = it->second.status.load(); - if (status == status_t::DONE) { - ldout(cct, 20) << __func__ << " successfully getting decompressed data, job id=" << decompress_id << dendl; - *finished = true; - data.swap(it->second.data); - jobs.erase(it); - } else if (status == status_t::ERROR) { - ldout(cct, 20) << __func__ << " compressed data failed, job id=" << decompress_id << dendl; - jobs.erase(it); - return -EIO; - } else if (blocking) { - auto expected = status_t::WAIT; - if (it->second.status.compare_exchange_strong(expected, status_t::DONE)) { - ldout(cct, 10) << __func__ << " decompress job id=" << decompress_id << " hasn't started, abort!"<< dendl; - if (compressor->decompress(it->second.data, data)) { - ldout(cct, 1) << __func__ << " decompress job id=" << decompress_id << " failed!"<< dendl; - it->second.status = status_t::ERROR; - return -EIO; - } - *finished = true; - } else { - job_lock.Unlock(); - usleep(1000); - job_lock.Lock(); - goto retry; - } - } else { - ldout(cct, 10) << __func__ << " decompress job id=" << decompress_id << " hasn't finished."<< dendl; - *finished = false; - } - return 0; -}