repo
stringlengths 1
152
⌀ | file
stringlengths 15
205
| code
stringlengths 0
41.6M
| file_length
int64 0
41.6M
| avg_line_length
float64 0
1.81M
| max_line_length
int64 0
12.7M
| extension_type
stringclasses 90
values |
---|---|---|---|---|---|---|
null |
ceph-main/src/tools/rbd_mirror/PoolMetaCache.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "common/debug.h"
#include "common/dout.h"
#include "tools/rbd_mirror/PoolMetaCache.h"
#include <shared_mutex>
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::PoolMetaCache: " \
<< this << " " << __func__ << ": "
namespace rbd {
namespace mirror {
int PoolMetaCache::get_local_pool_meta(
int64_t pool_id,
LocalPoolMeta* local_pool_meta) const {
dout(15) << "pool_id=" << pool_id << dendl;
std::shared_lock locker{m_lock};
auto it = m_local_pool_metas.find(pool_id);
if (it == m_local_pool_metas.end()) {
return -ENOENT;
}
*local_pool_meta = it->second;
return 0;
}
void PoolMetaCache::set_local_pool_meta(
int64_t pool_id,
const LocalPoolMeta& local_pool_meta) {
dout(15) << "pool_id=" << pool_id << ", "
<< "local_pool_meta=" << local_pool_meta << dendl;
std::unique_lock locker(m_lock);
m_local_pool_metas[pool_id] = local_pool_meta;
}
void PoolMetaCache::remove_local_pool_meta(int64_t pool_id) {
dout(15) << "pool_id=" << pool_id << dendl;
std::unique_lock locker(m_lock);
m_local_pool_metas.erase(pool_id);
}
int PoolMetaCache::get_remote_pool_meta(
int64_t pool_id,
RemotePoolMeta* remote_pool_meta) const {
dout(15) << "pool_id=" << pool_id << dendl;
std::shared_lock locker{m_lock};
auto it = m_remote_pool_metas.find(pool_id);
if (it == m_remote_pool_metas.end()) {
return -ENOENT;
}
*remote_pool_meta = it->second;
return 0;
}
void PoolMetaCache::set_remote_pool_meta(
int64_t pool_id,
const RemotePoolMeta& remote_pool_meta) {
dout(15) << "pool_id=" << pool_id << ", "
<< "remote_pool_meta=" << remote_pool_meta << dendl;
std::unique_lock locker(m_lock);
m_remote_pool_metas[pool_id] = remote_pool_meta;
}
void PoolMetaCache::remove_remote_pool_meta(int64_t pool_id) {
dout(15) << "pool_id=" << pool_id << dendl;
std::unique_lock locker(m_lock);
m_remote_pool_metas.erase(pool_id);
}
} // namespace mirror
} // namespace rbd
| 2,195 | 25.142857 | 70 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/PoolMetaCache.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef CEPH_RBD_MIRROR_POOL_META_CACHE_H
#define CEPH_RBD_MIRROR_POOL_META_CACHE_H
#include "include/int_types.h"
#include "common/ceph_mutex.h"
#include "tools/rbd_mirror/Types.h"
#include <map>
namespace rbd {
namespace mirror {
class PoolMetaCache {
public:
PoolMetaCache(CephContext* cct)
: m_cct(cct) {
}
PoolMetaCache(const PoolMetaCache&) = delete;
PoolMetaCache& operator=(const PoolMetaCache&) = delete;
int get_local_pool_meta(int64_t pool_id,
LocalPoolMeta* local_pool_meta) const;
void set_local_pool_meta(int64_t pool_id,
const LocalPoolMeta& local_pool_meta);
void remove_local_pool_meta(int64_t pool_id);
int get_remote_pool_meta(int64_t pool_id,
RemotePoolMeta* remote_pool_meta) const;
void set_remote_pool_meta(int64_t pool_id,
const RemotePoolMeta& remote_pool_meta);
void remove_remote_pool_meta(int64_t pool_id);
private:
CephContext* m_cct;
mutable ceph::shared_mutex m_lock =
ceph::make_shared_mutex("rbd::mirror::PoolMetaCache::m_lock");
std::map<int64_t, LocalPoolMeta> m_local_pool_metas;
std::map<int64_t, RemotePoolMeta> m_remote_pool_metas;
};
} // namespace mirror
} // namespace rbd
#endif // CEPH_RBD_MIRROR_POOL_META_CACHE_H
| 1,411 | 28.416667 | 70 |
h
|
null |
ceph-main/src/tools/rbd_mirror/PoolReplayer.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "PoolReplayer.h"
#include "common/Cond.h"
#include "common/Formatter.h"
#include "common/admin_socket.h"
#include "common/ceph_argparse.h"
#include "common/code_environment.h"
#include "common/common_init.h"
#include "common/debug.h"
#include "common/errno.h"
#include "cls/rbd/cls_rbd_client.h"
#include "global/global_context.h"
#include "librbd/api/Config.h"
#include "librbd/api/Namespace.h"
#include "PoolMetaCache.h"
#include "RemotePoolPoller.h"
#include "ServiceDaemon.h"
#include "Threads.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::PoolReplayer: " \
<< this << " " << __func__ << ": "
namespace rbd {
namespace mirror {
using ::operator<<;
namespace {
const std::string SERVICE_DAEMON_INSTANCE_ID_KEY("instance_id");
const std::string SERVICE_DAEMON_LEADER_KEY("leader");
const std::vector<std::string> UNIQUE_PEER_CONFIG_KEYS {
{"monmap", "mon_host", "mon_dns_srv_name", "key", "keyfile", "keyring"}};
template <typename I>
class PoolReplayerAdminSocketCommand {
public:
PoolReplayerAdminSocketCommand(PoolReplayer<I> *pool_replayer)
: pool_replayer(pool_replayer) {
}
virtual ~PoolReplayerAdminSocketCommand() {}
virtual int call(Formatter *f) = 0;
protected:
PoolReplayer<I> *pool_replayer;
};
template <typename I>
class StatusCommand : public PoolReplayerAdminSocketCommand<I> {
public:
explicit StatusCommand(PoolReplayer<I> *pool_replayer)
: PoolReplayerAdminSocketCommand<I>(pool_replayer) {
}
int call(Formatter *f) override {
this->pool_replayer->print_status(f);
return 0;
}
};
template <typename I>
class StartCommand : public PoolReplayerAdminSocketCommand<I> {
public:
explicit StartCommand(PoolReplayer<I> *pool_replayer)
: PoolReplayerAdminSocketCommand<I>(pool_replayer) {
}
int call(Formatter *f) override {
this->pool_replayer->start();
return 0;
}
};
template <typename I>
class StopCommand : public PoolReplayerAdminSocketCommand<I> {
public:
explicit StopCommand(PoolReplayer<I> *pool_replayer)
: PoolReplayerAdminSocketCommand<I>(pool_replayer) {
}
int call(Formatter *f) override {
this->pool_replayer->stop(true);
return 0;
}
};
template <typename I>
class RestartCommand : public PoolReplayerAdminSocketCommand<I> {
public:
explicit RestartCommand(PoolReplayer<I> *pool_replayer)
: PoolReplayerAdminSocketCommand<I>(pool_replayer) {
}
int call(Formatter *f) override {
this->pool_replayer->restart();
return 0;
}
};
template <typename I>
class FlushCommand : public PoolReplayerAdminSocketCommand<I> {
public:
explicit FlushCommand(PoolReplayer<I> *pool_replayer)
: PoolReplayerAdminSocketCommand<I>(pool_replayer) {
}
int call(Formatter *f) override {
this->pool_replayer->flush();
return 0;
}
};
template <typename I>
class LeaderReleaseCommand : public PoolReplayerAdminSocketCommand<I> {
public:
explicit LeaderReleaseCommand(PoolReplayer<I> *pool_replayer)
: PoolReplayerAdminSocketCommand<I>(pool_replayer) {
}
int call(Formatter *f) override {
this->pool_replayer->release_leader();
return 0;
}
};
template <typename I>
class PoolReplayerAdminSocketHook : public AdminSocketHook {
public:
PoolReplayerAdminSocketHook(CephContext *cct, const std::string &name,
PoolReplayer<I> *pool_replayer)
: admin_socket(cct->get_admin_socket()) {
std::string command;
int r;
command = "rbd mirror status " + name;
r = admin_socket->register_command(command, this,
"get status for rbd mirror " + name);
if (r == 0) {
commands[command] = new StatusCommand<I>(pool_replayer);
}
command = "rbd mirror start " + name;
r = admin_socket->register_command(command, this,
"start rbd mirror " + name);
if (r == 0) {
commands[command] = new StartCommand<I>(pool_replayer);
}
command = "rbd mirror stop " + name;
r = admin_socket->register_command(command, this,
"stop rbd mirror " + name);
if (r == 0) {
commands[command] = new StopCommand<I>(pool_replayer);
}
command = "rbd mirror restart " + name;
r = admin_socket->register_command(command, this,
"restart rbd mirror " + name);
if (r == 0) {
commands[command] = new RestartCommand<I>(pool_replayer);
}
command = "rbd mirror flush " + name;
r = admin_socket->register_command(command, this,
"flush rbd mirror " + name);
if (r == 0) {
commands[command] = new FlushCommand<I>(pool_replayer);
}
command = "rbd mirror leader release " + name;
r = admin_socket->register_command(command, this,
"release rbd mirror leader " + name);
if (r == 0) {
commands[command] = new LeaderReleaseCommand<I>(pool_replayer);
}
}
~PoolReplayerAdminSocketHook() override {
(void)admin_socket->unregister_commands(this);
for (auto i = commands.begin(); i != commands.end(); ++i) {
delete i->second;
}
}
int call(std::string_view command, const cmdmap_t& cmdmap,
const bufferlist&,
Formatter *f,
std::ostream& ss,
bufferlist& out) override {
auto i = commands.find(command);
ceph_assert(i != commands.end());
return i->second->call(f);
}
private:
typedef std::map<std::string, PoolReplayerAdminSocketCommand<I>*,
std::less<>> Commands;
AdminSocket *admin_socket;
Commands commands;
};
} // anonymous namespace
template <typename I>
struct PoolReplayer<I>::RemotePoolPollerListener
: public remote_pool_poller::Listener {
PoolReplayer<I>* m_pool_replayer;
RemotePoolPollerListener(PoolReplayer<I>* pool_replayer)
: m_pool_replayer(pool_replayer) {
}
void handle_updated(const RemotePoolMeta& remote_pool_meta) override {
m_pool_replayer->handle_remote_pool_meta_updated(remote_pool_meta);
}
};
template <typename I>
PoolReplayer<I>::PoolReplayer(
Threads<I> *threads, ServiceDaemon<I> *service_daemon,
journal::CacheManagerHandler *cache_manager_handler,
PoolMetaCache* pool_meta_cache, int64_t local_pool_id,
const PeerSpec &peer, const std::vector<const char*> &args) :
m_threads(threads),
m_service_daemon(service_daemon),
m_cache_manager_handler(cache_manager_handler),
m_pool_meta_cache(pool_meta_cache),
m_local_pool_id(local_pool_id),
m_peer(peer),
m_args(args),
m_lock(ceph::make_mutex("rbd::mirror::PoolReplayer " + stringify(peer))),
m_pool_replayer_thread(this),
m_leader_listener(this) {
}
template <typename I>
PoolReplayer<I>::~PoolReplayer()
{
shut_down();
ceph_assert(m_asok_hook == nullptr);
}
template <typename I>
bool PoolReplayer<I>::is_blocklisted() const {
std::lock_guard locker{m_lock};
return m_blocklisted;
}
template <typename I>
bool PoolReplayer<I>::is_leader() const {
std::lock_guard locker{m_lock};
return m_leader_watcher && m_leader_watcher->is_leader();
}
template <typename I>
bool PoolReplayer<I>::is_running() const {
return m_pool_replayer_thread.is_started() && !m_stopping;
}
template <typename I>
void PoolReplayer<I>::init(const std::string& site_name) {
std::lock_guard locker{m_lock};
ceph_assert(!m_pool_replayer_thread.is_started());
// reset state
m_stopping = false;
m_blocklisted = false;
m_site_name = site_name;
dout(10) << "replaying for " << m_peer << dendl;
int r = init_rados(g_ceph_context->_conf->cluster,
g_ceph_context->_conf->name.to_str(),
"", "", "local cluster", &m_local_rados, false);
if (r < 0) {
m_callout_id = m_service_daemon->add_or_update_callout(
m_local_pool_id, m_callout_id, service_daemon::CALLOUT_LEVEL_ERROR,
"unable to connect to local cluster");
return;
}
r = init_rados(m_peer.cluster_name, m_peer.client_name,
m_peer.mon_host, m_peer.key,
std::string("remote peer ") + stringify(m_peer),
&m_remote_rados, true);
if (r < 0) {
m_callout_id = m_service_daemon->add_or_update_callout(
m_local_pool_id, m_callout_id, service_daemon::CALLOUT_LEVEL_ERROR,
"unable to connect to remote cluster");
return;
}
r = m_local_rados->ioctx_create2(m_local_pool_id, m_local_io_ctx);
if (r < 0) {
derr << "error accessing local pool " << m_local_pool_id << ": "
<< cpp_strerror(r) << dendl;
return;
}
auto cct = reinterpret_cast<CephContext *>(m_local_io_ctx.cct());
librbd::api::Config<I>::apply_pool_overrides(m_local_io_ctx, &cct->_conf);
r = librbd::cls_client::mirror_uuid_get(&m_local_io_ctx,
&m_local_mirror_uuid);
if (r < 0) {
derr << "failed to retrieve local mirror uuid from pool "
<< m_local_io_ctx.get_pool_name() << ": " << cpp_strerror(r) << dendl;
m_callout_id = m_service_daemon->add_or_update_callout(
m_local_pool_id, m_callout_id, service_daemon::CALLOUT_LEVEL_ERROR,
"unable to query local mirror uuid");
return;
}
r = m_remote_rados->ioctx_create(m_local_io_ctx.get_pool_name().c_str(),
m_remote_io_ctx);
if (r < 0) {
derr << "error accessing remote pool " << m_local_io_ctx.get_pool_name()
<< ": " << cpp_strerror(r) << dendl;
m_callout_id = m_service_daemon->add_or_update_callout(
m_local_pool_id, m_callout_id, service_daemon::CALLOUT_LEVEL_WARNING,
"unable to access remote pool");
return;
}
dout(10) << "connected to " << m_peer << dendl;
m_image_sync_throttler.reset(
Throttler<I>::create(cct, "rbd_mirror_concurrent_image_syncs"));
m_image_deletion_throttler.reset(
Throttler<I>::create(cct, "rbd_mirror_concurrent_image_deletions"));
m_remote_pool_poller_listener.reset(new RemotePoolPollerListener(this));
m_remote_pool_poller.reset(RemotePoolPoller<I>::create(
m_threads, m_remote_io_ctx, m_site_name, m_local_mirror_uuid,
*m_remote_pool_poller_listener));
C_SaferCond on_pool_poller_init;
m_remote_pool_poller->init(&on_pool_poller_init);
r = on_pool_poller_init.wait();
if (r < 0) {
derr << "failed to initialize remote pool poller: " << cpp_strerror(r)
<< dendl;
m_callout_id = m_service_daemon->add_or_update_callout(
m_local_pool_id, m_callout_id, service_daemon::CALLOUT_LEVEL_ERROR,
"unable to initialize remote pool poller");
m_remote_pool_poller.reset();
return;
}
ceph_assert(!m_remote_pool_meta.mirror_uuid.empty());
m_pool_meta_cache->set_remote_pool_meta(
m_remote_io_ctx.get_id(), m_remote_pool_meta);
m_pool_meta_cache->set_local_pool_meta(
m_local_io_ctx.get_id(), {m_local_mirror_uuid});
m_default_namespace_replayer.reset(NamespaceReplayer<I>::create(
"", m_local_io_ctx, m_remote_io_ctx, m_local_mirror_uuid, m_peer.uuid,
m_remote_pool_meta, m_threads, m_image_sync_throttler.get(),
m_image_deletion_throttler.get(), m_service_daemon,
m_cache_manager_handler, m_pool_meta_cache));
C_SaferCond on_init;
m_default_namespace_replayer->init(&on_init);
r = on_init.wait();
if (r < 0) {
derr << "error initializing default namespace replayer: " << cpp_strerror(r)
<< dendl;
m_callout_id = m_service_daemon->add_or_update_callout(
m_local_pool_id, m_callout_id, service_daemon::CALLOUT_LEVEL_ERROR,
"unable to initialize default namespace replayer");
m_default_namespace_replayer.reset();
return;
}
m_leader_watcher.reset(LeaderWatcher<I>::create(m_threads, m_local_io_ctx,
&m_leader_listener));
r = m_leader_watcher->init();
if (r < 0) {
derr << "error initializing leader watcher: " << cpp_strerror(r) << dendl;
m_callout_id = m_service_daemon->add_or_update_callout(
m_local_pool_id, m_callout_id, service_daemon::CALLOUT_LEVEL_ERROR,
"unable to initialize leader messenger object");
m_leader_watcher.reset();
return;
}
if (m_callout_id != service_daemon::CALLOUT_ID_NONE) {
m_service_daemon->remove_callout(m_local_pool_id, m_callout_id);
m_callout_id = service_daemon::CALLOUT_ID_NONE;
}
m_service_daemon->add_or_update_attribute(
m_local_io_ctx.get_id(), SERVICE_DAEMON_INSTANCE_ID_KEY,
stringify(m_local_io_ctx.get_instance_id()));
m_pool_replayer_thread.create("pool replayer");
}
template <typename I>
void PoolReplayer<I>::shut_down() {
dout(20) << dendl;
{
std::lock_guard l{m_lock};
m_stopping = true;
m_cond.notify_all();
}
if (m_pool_replayer_thread.is_started()) {
m_pool_replayer_thread.join();
}
if (m_leader_watcher) {
m_leader_watcher->shut_down();
}
m_leader_watcher.reset();
if (m_default_namespace_replayer) {
C_SaferCond on_shut_down;
m_default_namespace_replayer->shut_down(&on_shut_down);
on_shut_down.wait();
}
m_default_namespace_replayer.reset();
if (m_remote_pool_poller) {
C_SaferCond ctx;
m_remote_pool_poller->shut_down(&ctx);
ctx.wait();
m_pool_meta_cache->remove_remote_pool_meta(m_remote_io_ctx.get_id());
m_pool_meta_cache->remove_local_pool_meta(m_local_io_ctx.get_id());
}
m_remote_pool_poller.reset();
m_remote_pool_poller_listener.reset();
m_image_sync_throttler.reset();
m_image_deletion_throttler.reset();
m_local_rados.reset();
m_remote_rados.reset();
}
template <typename I>
int PoolReplayer<I>::init_rados(const std::string &cluster_name,
const std::string &client_name,
const std::string &mon_host,
const std::string &key,
const std::string &description,
RadosRef *rados_ref,
bool strip_cluster_overrides) {
dout(10) << "cluster_name=" << cluster_name << ", client_name=" << client_name
<< ", mon_host=" << mon_host << ", strip_cluster_overrides="
<< strip_cluster_overrides << dendl;
// NOTE: manually bootstrap a CephContext here instead of via
// the librados API to avoid mixing global singletons between
// the librados shared library and the daemon
// TODO: eliminate intermingling of global singletons within Ceph APIs
CephInitParameters iparams(CEPH_ENTITY_TYPE_CLIENT);
if (client_name.empty() || !iparams.name.from_str(client_name)) {
derr << "error initializing cluster handle for " << description << dendl;
return -EINVAL;
}
CephContext *cct = common_preinit(iparams, CODE_ENVIRONMENT_LIBRARY,
CINIT_FLAG_UNPRIVILEGED_DAEMON_DEFAULTS);
cct->_conf->cluster = cluster_name;
// librados::Rados::conf_read_file
int r = cct->_conf.parse_config_files(nullptr, nullptr, 0);
if (r < 0 && r != -ENOENT) {
// do not treat this as fatal, it might still be able to connect
derr << "could not read ceph conf for " << description << ": "
<< cpp_strerror(r) << dendl;
}
// preserve cluster-specific config settings before applying environment/cli
// overrides
std::map<std::string, std::string> config_values;
if (strip_cluster_overrides) {
// remote peer connections shouldn't apply cluster-specific
// configuration settings
for (auto& key : UNIQUE_PEER_CONFIG_KEYS) {
config_values[key] = cct->_conf.get_val<std::string>(key);
}
}
cct->_conf.parse_env(cct->get_module_type());
// librados::Rados::conf_parse_env
std::vector<const char*> args;
r = cct->_conf.parse_argv(args);
if (r < 0) {
derr << "could not parse environment for " << description << ":"
<< cpp_strerror(r) << dendl;
cct->put();
return r;
}
cct->_conf.parse_env(cct->get_module_type());
if (!m_args.empty()) {
// librados::Rados::conf_parse_argv
args = m_args;
r = cct->_conf.parse_argv(args);
if (r < 0) {
derr << "could not parse command line args for " << description << ": "
<< cpp_strerror(r) << dendl;
cct->put();
return r;
}
}
if (strip_cluster_overrides) {
// remote peer connections shouldn't apply cluster-specific
// configuration settings
for (auto& pair : config_values) {
auto value = cct->_conf.get_val<std::string>(pair.first);
if (pair.second != value) {
dout(0) << "reverting global config option override: "
<< pair.first << ": " << value << " -> " << pair.second
<< dendl;
cct->_conf.set_val_or_die(pair.first, pair.second);
}
}
}
if (!g_ceph_context->_conf->admin_socket.empty()) {
cct->_conf.set_val_or_die("admin_socket",
"$run_dir/$name.$pid.$cluster.$cctid.asok");
}
if (!mon_host.empty()) {
r = cct->_conf.set_val("mon_host", mon_host);
if (r < 0) {
derr << "failed to set mon_host config for " << description << ": "
<< cpp_strerror(r) << dendl;
cct->put();
return r;
}
}
if (!key.empty()) {
r = cct->_conf.set_val("key", key);
if (r < 0) {
derr << "failed to set key config for " << description << ": "
<< cpp_strerror(r) << dendl;
cct->put();
return r;
}
}
// disable unnecessary librbd cache
cct->_conf.set_val_or_die("rbd_cache", "false");
cct->_conf.apply_changes(nullptr);
cct->_conf.complain_about_parse_error(cct);
rados_ref->reset(new librados::Rados());
r = (*rados_ref)->init_with_context(cct);
ceph_assert(r == 0);
cct->put();
r = (*rados_ref)->connect();
if (r < 0) {
derr << "error connecting to " << description << ": "
<< cpp_strerror(r) << dendl;
return r;
}
return 0;
}
template <typename I>
void PoolReplayer<I>::run() {
dout(20) << dendl;
while (true) {
std::string asok_hook_name = m_local_io_ctx.get_pool_name() + " " +
m_peer.cluster_name;
if (m_asok_hook_name != asok_hook_name || m_asok_hook == nullptr) {
m_asok_hook_name = asok_hook_name;
delete m_asok_hook;
m_asok_hook = new PoolReplayerAdminSocketHook<I>(g_ceph_context,
m_asok_hook_name, this);
}
with_namespace_replayers([this]() { update_namespace_replayers(); });
std::unique_lock locker{m_lock};
if (m_leader_watcher->is_blocklisted() ||
m_default_namespace_replayer->is_blocklisted()) {
m_blocklisted = true;
m_stopping = true;
}
for (auto &it : m_namespace_replayers) {
if (it.second->is_blocklisted()) {
m_blocklisted = true;
m_stopping = true;
break;
}
}
if (m_stopping) {
break;
}
auto seconds = g_ceph_context->_conf.get_val<uint64_t>(
"rbd_mirror_pool_replayers_refresh_interval");
m_cond.wait_for(locker, ceph::make_timespan(seconds));
}
// shut down namespace replayers
with_namespace_replayers([this]() { update_namespace_replayers(); });
delete m_asok_hook;
m_asok_hook = nullptr;
}
template <typename I>
void PoolReplayer<I>::update_namespace_replayers() {
dout(20) << dendl;
ceph_assert(ceph_mutex_is_locked(m_lock));
std::set<std::string> mirroring_namespaces;
if (!m_stopping) {
int r = list_mirroring_namespaces(&mirroring_namespaces);
if (r < 0) {
return;
}
}
auto cct = reinterpret_cast<CephContext *>(m_local_io_ctx.cct());
C_SaferCond cond;
auto gather_ctx = new C_Gather(cct, &cond);
for (auto it = m_namespace_replayers.begin();
it != m_namespace_replayers.end(); ) {
auto iter = mirroring_namespaces.find(it->first);
if (iter == mirroring_namespaces.end()) {
auto namespace_replayer = it->second;
auto on_shut_down = new LambdaContext(
[namespace_replayer, ctx=gather_ctx->new_sub()](int r) {
delete namespace_replayer;
ctx->complete(r);
});
m_service_daemon->remove_namespace(m_local_pool_id, it->first);
namespace_replayer->shut_down(on_shut_down);
it = m_namespace_replayers.erase(it);
} else {
mirroring_namespaces.erase(iter);
it++;
}
}
for (auto &name : mirroring_namespaces) {
auto namespace_replayer = NamespaceReplayer<I>::create(
name, m_local_io_ctx, m_remote_io_ctx, m_local_mirror_uuid, m_peer.uuid,
m_remote_pool_meta, m_threads, m_image_sync_throttler.get(),
m_image_deletion_throttler.get(), m_service_daemon,
m_cache_manager_handler, m_pool_meta_cache);
auto on_init = new LambdaContext(
[this, namespace_replayer, name, &mirroring_namespaces,
ctx=gather_ctx->new_sub()](int r) {
std::lock_guard locker{m_lock};
if (r < 0) {
derr << "failed to initialize namespace replayer for namespace "
<< name << ": " << cpp_strerror(r) << dendl;
delete namespace_replayer;
mirroring_namespaces.erase(name);
} else {
m_namespace_replayers[name] = namespace_replayer;
m_service_daemon->add_namespace(m_local_pool_id, name);
}
ctx->complete(r);
});
namespace_replayer->init(on_init);
}
gather_ctx->activate();
m_lock.unlock();
cond.wait();
m_lock.lock();
if (m_leader) {
C_SaferCond acquire_cond;
auto acquire_gather_ctx = new C_Gather(cct, &acquire_cond);
for (auto &name : mirroring_namespaces) {
namespace_replayer_acquire_leader(name, acquire_gather_ctx->new_sub());
}
acquire_gather_ctx->activate();
m_lock.unlock();
acquire_cond.wait();
m_lock.lock();
std::vector<std::string> instance_ids;
m_leader_watcher->list_instances(&instance_ids);
for (auto &name : mirroring_namespaces) {
auto it = m_namespace_replayers.find(name);
if (it == m_namespace_replayers.end()) {
// acquire leader for this namespace replayer failed
continue;
}
it->second->handle_instances_added(instance_ids);
}
} else {
std::string leader_instance_id;
if (m_leader_watcher->get_leader_instance_id(&leader_instance_id)) {
for (auto &name : mirroring_namespaces) {
m_namespace_replayers[name]->handle_update_leader(leader_instance_id);
}
}
}
}
template <typename I>
int PoolReplayer<I>::list_mirroring_namespaces(
std::set<std::string> *namespaces) {
dout(20) << dendl;
ceph_assert(ceph_mutex_is_locked(m_lock));
std::vector<std::string> names;
int r = librbd::api::Namespace<I>::list(m_local_io_ctx, &names);
if (r < 0) {
derr << "failed to list namespaces: " << cpp_strerror(r) << dendl;
return r;
}
for (auto &name : names) {
cls::rbd::MirrorMode mirror_mode = cls::rbd::MIRROR_MODE_DISABLED;
int r = librbd::cls_client::mirror_mode_get(&m_local_io_ctx, &mirror_mode);
if (r < 0 && r != -ENOENT) {
derr << "failed to get namespace mirror mode: " << cpp_strerror(r)
<< dendl;
if (m_namespace_replayers.count(name) == 0) {
continue;
}
} else if (mirror_mode == cls::rbd::MIRROR_MODE_DISABLED) {
dout(10) << "mirroring is disabled for namespace " << name << dendl;
continue;
}
namespaces->insert(name);
}
return 0;
}
template <typename I>
void PoolReplayer<I>::reopen_logs()
{
dout(20) << dendl;
std::lock_guard locker{m_lock};
if (m_local_rados) {
reinterpret_cast<CephContext *>(m_local_rados->cct())->reopen_logs();
}
if (m_remote_rados) {
reinterpret_cast<CephContext *>(m_remote_rados->cct())->reopen_logs();
}
}
template <typename I>
void PoolReplayer<I>::namespace_replayer_acquire_leader(const std::string &name,
Context *on_finish) {
dout(20) << dendl;
ceph_assert(ceph_mutex_is_locked(m_lock));
auto it = m_namespace_replayers.find(name);
ceph_assert(it != m_namespace_replayers.end());
on_finish = new LambdaContext(
[this, name, on_finish](int r) {
if (r < 0) {
derr << "failed to handle acquire leader for namespace: "
<< name << ": " << cpp_strerror(r) << dendl;
// remove the namespace replayer -- update_namespace_replayers will
// retry to create it and acquire leader.
std::lock_guard locker{m_lock};
auto namespace_replayer = m_namespace_replayers[name];
m_namespace_replayers.erase(name);
auto on_shut_down = new LambdaContext(
[namespace_replayer, on_finish](int r) {
delete namespace_replayer;
on_finish->complete(r);
});
m_service_daemon->remove_namespace(m_local_pool_id, name);
namespace_replayer->shut_down(on_shut_down);
return;
}
on_finish->complete(0);
});
it->second->handle_acquire_leader(on_finish);
}
template <typename I>
void PoolReplayer<I>::print_status(Formatter *f) {
dout(20) << dendl;
assert(f);
std::lock_guard l{m_lock};
f->open_object_section("pool_replayer_status");
f->dump_stream("peer") << m_peer;
if (m_local_io_ctx.is_valid()) {
f->dump_string("pool", m_local_io_ctx.get_pool_name());
f->dump_stream("instance_id") << m_local_io_ctx.get_instance_id();
}
std::string state("running");
if (m_manual_stop) {
state = "stopped (manual)";
} else if (m_stopping) {
state = "stopped";
} else if (!is_running()) {
state = "error";
}
f->dump_string("state", state);
if (m_leader_watcher) {
std::string leader_instance_id;
m_leader_watcher->get_leader_instance_id(&leader_instance_id);
f->dump_string("leader_instance_id", leader_instance_id);
bool leader = m_leader_watcher->is_leader();
f->dump_bool("leader", leader);
if (leader) {
std::vector<std::string> instance_ids;
m_leader_watcher->list_instances(&instance_ids);
f->open_array_section("instances");
for (auto instance_id : instance_ids) {
f->dump_string("instance_id", instance_id);
}
f->close_section(); // instances
}
}
if (m_local_rados) {
auto cct = reinterpret_cast<CephContext *>(m_local_rados->cct());
f->dump_string("local_cluster_admin_socket",
cct->_conf.get_val<std::string>("admin_socket"));
}
if (m_remote_rados) {
auto cct = reinterpret_cast<CephContext *>(m_remote_rados->cct());
f->dump_string("remote_cluster_admin_socket",
cct->_conf.get_val<std::string>("admin_socket"));
}
if (m_image_sync_throttler) {
f->open_object_section("sync_throttler");
m_image_sync_throttler->print_status(f);
f->close_section(); // sync_throttler
}
if (m_image_deletion_throttler) {
f->open_object_section("deletion_throttler");
m_image_deletion_throttler->print_status(f);
f->close_section(); // deletion_throttler
}
if (m_default_namespace_replayer) {
m_default_namespace_replayer->print_status(f);
}
f->open_array_section("namespaces");
for (auto &it : m_namespace_replayers) {
f->open_object_section("namespace");
f->dump_string("name", it.first);
it.second->print_status(f);
f->close_section(); // namespace
}
f->close_section(); // namespaces
f->close_section(); // pool_replayer_status
}
template <typename I>
void PoolReplayer<I>::start() {
dout(20) << dendl;
std::lock_guard l{m_lock};
if (m_stopping) {
return;
}
m_manual_stop = false;
if (m_default_namespace_replayer) {
m_default_namespace_replayer->start();
}
for (auto &it : m_namespace_replayers) {
it.second->start();
}
}
template <typename I>
void PoolReplayer<I>::stop(bool manual) {
dout(20) << "enter: manual=" << manual << dendl;
std::lock_guard l{m_lock};
if (!manual) {
m_stopping = true;
m_cond.notify_all();
return;
} else if (m_stopping) {
return;
}
m_manual_stop = true;
if (m_default_namespace_replayer) {
m_default_namespace_replayer->stop();
}
for (auto &it : m_namespace_replayers) {
it.second->stop();
}
}
template <typename I>
void PoolReplayer<I>::restart() {
dout(20) << dendl;
std::lock_guard l{m_lock};
if (m_stopping) {
return;
}
if (m_default_namespace_replayer) {
m_default_namespace_replayer->restart();
}
for (auto &it : m_namespace_replayers) {
it.second->restart();
}
}
template <typename I>
void PoolReplayer<I>::flush() {
dout(20) << dendl;
std::lock_guard l{m_lock};
if (m_stopping || m_manual_stop) {
return;
}
if (m_default_namespace_replayer) {
m_default_namespace_replayer->flush();
}
for (auto &it : m_namespace_replayers) {
it.second->flush();
}
}
template <typename I>
void PoolReplayer<I>::release_leader() {
dout(20) << dendl;
std::lock_guard l{m_lock};
if (m_stopping || !m_leader_watcher) {
return;
}
m_leader_watcher->release_leader();
}
template <typename I>
void PoolReplayer<I>::handle_post_acquire_leader(Context *on_finish) {
dout(20) << dendl;
with_namespace_replayers(
[this](Context *on_finish) {
dout(10) << "handle_post_acquire_leader" << dendl;
ceph_assert(ceph_mutex_is_locked(m_lock));
m_service_daemon->add_or_update_attribute(m_local_pool_id,
SERVICE_DAEMON_LEADER_KEY,
true);
auto ctx = new LambdaContext(
[this, on_finish](int r) {
if (r == 0) {
std::lock_guard locker{m_lock};
m_leader = true;
}
on_finish->complete(r);
});
auto cct = reinterpret_cast<CephContext *>(m_local_io_ctx.cct());
auto gather_ctx = new C_Gather(cct, ctx);
m_default_namespace_replayer->handle_acquire_leader(
gather_ctx->new_sub());
for (auto &it : m_namespace_replayers) {
namespace_replayer_acquire_leader(it.first, gather_ctx->new_sub());
}
gather_ctx->activate();
}, on_finish);
}
template <typename I>
void PoolReplayer<I>::handle_pre_release_leader(Context *on_finish) {
dout(20) << dendl;
with_namespace_replayers(
[this](Context *on_finish) {
dout(10) << "handle_pre_release_leader" << dendl;
ceph_assert(ceph_mutex_is_locked(m_lock));
m_leader = false;
m_service_daemon->remove_attribute(m_local_pool_id,
SERVICE_DAEMON_LEADER_KEY);
auto cct = reinterpret_cast<CephContext *>(m_local_io_ctx.cct());
auto gather_ctx = new C_Gather(cct, on_finish);
m_default_namespace_replayer->handle_release_leader(
gather_ctx->new_sub());
for (auto &it : m_namespace_replayers) {
it.second->handle_release_leader(gather_ctx->new_sub());
}
gather_ctx->activate();
}, on_finish);
}
template <typename I>
void PoolReplayer<I>::handle_update_leader(
const std::string &leader_instance_id) {
dout(10) << "leader_instance_id=" << leader_instance_id << dendl;
std::lock_guard locker{m_lock};
m_default_namespace_replayer->handle_update_leader(leader_instance_id);
for (auto &it : m_namespace_replayers) {
it.second->handle_update_leader(leader_instance_id);
}
}
template <typename I>
void PoolReplayer<I>::handle_instances_added(
const std::vector<std::string> &instance_ids) {
dout(5) << "instance_ids=" << instance_ids << dendl;
std::lock_guard locker{m_lock};
if (!m_leader_watcher->is_leader()) {
return;
}
m_default_namespace_replayer->handle_instances_added(instance_ids);
for (auto &it : m_namespace_replayers) {
it.second->handle_instances_added(instance_ids);
}
}
template <typename I>
void PoolReplayer<I>::handle_instances_removed(
const std::vector<std::string> &instance_ids) {
dout(5) << "instance_ids=" << instance_ids << dendl;
std::lock_guard locker{m_lock};
if (!m_leader_watcher->is_leader()) {
return;
}
m_default_namespace_replayer->handle_instances_removed(instance_ids);
for (auto &it : m_namespace_replayers) {
it.second->handle_instances_removed(instance_ids);
}
}
template <typename I>
void PoolReplayer<I>::handle_remote_pool_meta_updated(
const RemotePoolMeta& remote_pool_meta) {
dout(5) << "remote_pool_meta=" << remote_pool_meta << dendl;
if (!m_default_namespace_replayer) {
m_remote_pool_meta = remote_pool_meta;
return;
}
derr << "remote pool metadata updated unexpectedly" << dendl;
std::unique_lock locker{m_lock};
m_stopping = true;
m_cond.notify_all();
}
} // namespace mirror
} // namespace rbd
template class rbd::mirror::PoolReplayer<librbd::ImageCtx>;
| 32,868 | 28.373548 | 80 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/PoolReplayer.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef CEPH_RBD_MIRROR_POOL_REPLAYER_H
#define CEPH_RBD_MIRROR_POOL_REPLAYER_H
#include "common/Cond.h"
#include "common/ceph_mutex.h"
#include "include/rados/librados.hpp"
#include "librbd/Utils.h"
#include "librbd/asio/ContextWQ.h"
#include "tools/rbd_mirror/LeaderWatcher.h"
#include "tools/rbd_mirror/NamespaceReplayer.h"
#include "tools/rbd_mirror/Throttler.h"
#include "tools/rbd_mirror/Types.h"
#include "tools/rbd_mirror/leader_watcher/Types.h"
#include "tools/rbd_mirror/service_daemon/Types.h"
#include <map>
#include <memory>
#include <string>
#include <vector>
class AdminSocketHook;
namespace journal { struct CacheManagerHandler; }
namespace librbd { class ImageCtx; }
namespace rbd {
namespace mirror {
template <typename> class RemotePoolPoller;
namespace remote_pool_poller { struct Listener; }
struct PoolMetaCache;
template <typename> class ServiceDaemon;
template <typename> struct Threads;
/**
* Controls mirroring for a single remote cluster.
*/
template <typename ImageCtxT = librbd::ImageCtx>
class PoolReplayer {
public:
PoolReplayer(Threads<ImageCtxT> *threads,
ServiceDaemon<ImageCtxT> *service_daemon,
journal::CacheManagerHandler *cache_manager_handler,
PoolMetaCache* pool_meta_cache,
int64_t local_pool_id, const PeerSpec &peer,
const std::vector<const char*> &args);
~PoolReplayer();
PoolReplayer(const PoolReplayer&) = delete;
PoolReplayer& operator=(const PoolReplayer&) = delete;
bool is_blocklisted() const;
bool is_leader() const;
bool is_running() const;
void init(const std::string& site_name);
void shut_down();
void run();
void print_status(Formatter *f);
void start();
void stop(bool manual);
void restart();
void flush();
void release_leader();
void reopen_logs();
private:
/**
* @verbatim
*
* <start>
* |
* v
* INIT
* |
* v
* <follower> <---------------------\
* . |
* . (leader acquired) |
* v |
* NOTIFY_NAMESPACE_WATCHERS NOTIFY_NAMESPACE_WATCHERS
* | ^
* v .
* <leader> .
* . .
* . (leader lost / shut down) .
* . . . . . . . . . . . . . . . .
*
* @endverbatim
*/
struct RemotePoolPollerListener;
int init_rados(const std::string &cluster_name,
const std::string &client_name,
const std::string &mon_host,
const std::string &key,
const std::string &description, RadosRef *rados_ref,
bool strip_cluster_overrides);
void update_namespace_replayers();
int list_mirroring_namespaces(std::set<std::string> *namespaces);
void namespace_replayer_acquire_leader(const std::string &name,
Context *on_finish);
void handle_post_acquire_leader(Context *on_finish);
void handle_pre_release_leader(Context *on_finish);
void handle_update_leader(const std::string &leader_instance_id);
void handle_instances_added(const std::vector<std::string> &instance_ids);
void handle_instances_removed(const std::vector<std::string> &instance_ids);
// sync version, executed in the caller thread
template <typename L>
void with_namespace_replayers(L &&callback) {
std::lock_guard locker{m_lock};
if (m_namespace_replayers_locked) {
ceph_assert(m_on_namespace_replayers_unlocked == nullptr);
C_SaferCond cond;
m_on_namespace_replayers_unlocked = &cond;
m_lock.unlock();
cond.wait();
m_lock.lock();
} else {
m_namespace_replayers_locked = true;
}
ceph_assert(m_namespace_replayers_locked);
callback(); // may temporary release the lock
ceph_assert(m_namespace_replayers_locked);
if (m_on_namespace_replayers_unlocked == nullptr) {
m_namespace_replayers_locked = false;
return;
}
m_threads->work_queue->queue(m_on_namespace_replayers_unlocked);
m_on_namespace_replayers_unlocked = nullptr;
}
// async version
template <typename L>
void with_namespace_replayers(L &&callback, Context *on_finish) {
std::lock_guard locker{m_lock};
on_finish = librbd::util::create_async_context_callback(
m_threads->work_queue, new LambdaContext(
[this, on_finish](int r) {
{
std::lock_guard locker{m_lock};
ceph_assert(m_namespace_replayers_locked);
m_namespace_replayers_locked = false;
if (m_on_namespace_replayers_unlocked != nullptr) {
m_namespace_replayers_locked = true;
m_threads->work_queue->queue(m_on_namespace_replayers_unlocked);
m_on_namespace_replayers_unlocked = nullptr;
}
}
on_finish->complete(r);
}));
auto on_lock = new LambdaContext(
[this, callback, on_finish](int) {
std::lock_guard locker{m_lock};
ceph_assert(m_namespace_replayers_locked);
callback(on_finish);
});
if (m_namespace_replayers_locked) {
ceph_assert(m_on_namespace_replayers_unlocked == nullptr);
m_on_namespace_replayers_unlocked = on_lock;
return;
}
m_namespace_replayers_locked = true;
m_threads->work_queue->queue(on_lock);
}
void handle_remote_pool_meta_updated(const RemotePoolMeta& remote_pool_meta);
Threads<ImageCtxT> *m_threads;
ServiceDaemon<ImageCtxT> *m_service_daemon;
journal::CacheManagerHandler *m_cache_manager_handler;
PoolMetaCache* m_pool_meta_cache;
int64_t m_local_pool_id = -1;
PeerSpec m_peer;
std::vector<const char*> m_args;
mutable ceph::mutex m_lock;
ceph::condition_variable m_cond;
std::string m_site_name;
bool m_stopping = false;
bool m_manual_stop = false;
bool m_blocklisted = false;
RadosRef m_local_rados;
RadosRef m_remote_rados;
librados::IoCtx m_local_io_ctx;
librados::IoCtx m_remote_io_ctx;
std::string m_local_mirror_uuid;
RemotePoolMeta m_remote_pool_meta;
std::unique_ptr<remote_pool_poller::Listener> m_remote_pool_poller_listener;
std::unique_ptr<RemotePoolPoller<ImageCtxT>> m_remote_pool_poller;
std::unique_ptr<NamespaceReplayer<ImageCtxT>> m_default_namespace_replayer;
std::map<std::string, NamespaceReplayer<ImageCtxT> *> m_namespace_replayers;
std::string m_asok_hook_name;
AdminSocketHook *m_asok_hook = nullptr;
service_daemon::CalloutId m_callout_id = service_daemon::CALLOUT_ID_NONE;
bool m_leader = false;
bool m_namespace_replayers_locked = false;
Context *m_on_namespace_replayers_unlocked = nullptr;
class PoolReplayerThread : public Thread {
PoolReplayer *m_pool_replayer;
public:
PoolReplayerThread(PoolReplayer *pool_replayer)
: m_pool_replayer(pool_replayer) {
}
void *entry() override {
m_pool_replayer->run();
return 0;
}
} m_pool_replayer_thread;
class LeaderListener : public leader_watcher::Listener {
public:
LeaderListener(PoolReplayer *pool_replayer)
: m_pool_replayer(pool_replayer) {
}
protected:
void post_acquire_handler(Context *on_finish) override {
m_pool_replayer->handle_post_acquire_leader(on_finish);
}
void pre_release_handler(Context *on_finish) override {
m_pool_replayer->handle_pre_release_leader(on_finish);
}
void update_leader_handler(
const std::string &leader_instance_id) override {
m_pool_replayer->handle_update_leader(leader_instance_id);
}
void handle_instances_added(const InstanceIds& instance_ids) override {
m_pool_replayer->handle_instances_added(instance_ids);
}
void handle_instances_removed(const InstanceIds& instance_ids) override {
m_pool_replayer->handle_instances_removed(instance_ids);
}
private:
PoolReplayer *m_pool_replayer;
} m_leader_listener;
std::unique_ptr<LeaderWatcher<ImageCtxT>> m_leader_watcher;
std::unique_ptr<Throttler<ImageCtxT>> m_image_sync_throttler;
std::unique_ptr<Throttler<ImageCtxT>> m_image_deletion_throttler;
};
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::PoolReplayer<librbd::ImageCtx>;
#endif // CEPH_RBD_MIRROR_POOL_REPLAYER_H
| 8,480 | 28.346021 | 80 |
h
|
null |
ceph-main/src/tools/rbd_mirror/PoolWatcher.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "tools/rbd_mirror/PoolWatcher.h"
#include "include/rbd_types.h"
#include "cls/rbd/cls_rbd_client.h"
#include "common/debug.h"
#include "common/errno.h"
#include "common/Timer.h"
#include "librbd/ImageCtx.h"
#include "librbd/internal.h"
#include "librbd/MirroringWatcher.h"
#include "librbd/Utils.h"
#include "librbd/api/Image.h"
#include "librbd/api/Mirror.h"
#include "librbd/asio/ContextWQ.h"
#include "tools/rbd_mirror/Threads.h"
#include "tools/rbd_mirror/pool_watcher/RefreshImagesRequest.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::PoolWatcher: " << this << " " \
<< __func__ << ": "
using std::list;
using std::string;
using std::unique_ptr;
using std::vector;
using librbd::util::create_context_callback;
using librbd::util::create_rados_callback;
namespace rbd {
namespace mirror {
template <typename I>
class PoolWatcher<I>::MirroringWatcher : public librbd::MirroringWatcher<I> {
public:
using ContextWQ = typename std::decay<
typename std::remove_pointer<
decltype(Threads<I>::work_queue)>::type>::type;
MirroringWatcher(librados::IoCtx &io_ctx, ContextWQ *work_queue,
PoolWatcher *pool_watcher)
: librbd::MirroringWatcher<I>(io_ctx, work_queue),
m_pool_watcher(pool_watcher) {
}
void handle_rewatch_complete(int r) override {
m_pool_watcher->handle_rewatch_complete(r);
}
void handle_mode_updated(cls::rbd::MirrorMode mirror_mode) override {
// invalidate all image state and refresh the pool contents
m_pool_watcher->schedule_refresh_images(5);
}
void handle_image_updated(cls::rbd::MirrorImageState state,
const std::string &image_id,
const std::string &global_image_id) override {
bool enabled = (state == cls::rbd::MIRROR_IMAGE_STATE_ENABLED);
m_pool_watcher->handle_image_updated(image_id, global_image_id,
enabled);
}
private:
PoolWatcher *m_pool_watcher;
};
template <typename I>
PoolWatcher<I>::PoolWatcher(Threads<I> *threads,
librados::IoCtx &io_ctx,
const std::string& mirror_uuid,
pool_watcher::Listener &listener)
: m_threads(threads),
m_io_ctx(io_ctx),
m_mirror_uuid(mirror_uuid),
m_listener(listener),
m_lock(ceph::make_mutex(librbd::util::unique_lock_name(
"rbd::mirror::PoolWatcher", this))) {
m_mirroring_watcher = new MirroringWatcher(m_io_ctx,
m_threads->work_queue, this);
}
template <typename I>
PoolWatcher<I>::~PoolWatcher() {
delete m_mirroring_watcher;
}
template <typename I>
bool PoolWatcher<I>::is_blocklisted() const {
std::lock_guard locker{m_lock};
return m_blocklisted;
}
template <typename I>
void PoolWatcher<I>::init(Context *on_finish) {
dout(5) << dendl;
{
std::lock_guard locker{m_lock};
m_on_init_finish = on_finish;
ceph_assert(!m_refresh_in_progress);
m_refresh_in_progress = true;
}
// start async updates for mirror image directory
register_watcher();
}
template <typename I>
void PoolWatcher<I>::shut_down(Context *on_finish) {
dout(5) << dendl;
{
std::scoped_lock locker{m_threads->timer_lock, m_lock};
ceph_assert(!m_shutting_down);
m_shutting_down = true;
if (m_timer_ctx != nullptr) {
m_threads->timer->cancel_event(m_timer_ctx);
m_timer_ctx = nullptr;
}
}
// in-progress unregister tracked as async op
unregister_watcher();
m_async_op_tracker.wait_for_ops(on_finish);
}
template <typename I>
void PoolWatcher<I>::register_watcher() {
{
std::lock_guard locker{m_lock};
ceph_assert(m_image_ids_invalid);
ceph_assert(m_refresh_in_progress);
}
// if the watch registration is in-flight, let the watcher
// handle the transition -- only (re-)register if it's not registered
if (!m_mirroring_watcher->is_unregistered()) {
refresh_images();
return;
}
// first time registering or the watch failed
dout(5) << dendl;
m_async_op_tracker.start_op();
Context *ctx = create_context_callback<
PoolWatcher, &PoolWatcher<I>::handle_register_watcher>(this);
m_mirroring_watcher->register_watch(ctx);
}
template <typename I>
void PoolWatcher<I>::handle_register_watcher(int r) {
dout(5) << "r=" << r << dendl;
{
std::lock_guard locker{m_lock};
ceph_assert(m_image_ids_invalid);
ceph_assert(m_refresh_in_progress);
if (r < 0) {
m_refresh_in_progress = false;
}
}
Context *on_init_finish = nullptr;
if (r >= 0) {
refresh_images();
} else if (r == -EBLOCKLISTED) {
dout(0) << "detected client is blocklisted" << dendl;
std::lock_guard locker{m_lock};
m_blocklisted = true;
std::swap(on_init_finish, m_on_init_finish);
} else if (r == -ENOENT) {
dout(5) << "mirroring directory does not exist" << dendl;
{
std::lock_guard locker{m_lock};
std::swap(on_init_finish, m_on_init_finish);
}
schedule_refresh_images(30);
} else {
derr << "unexpected error registering mirroring directory watch: "
<< cpp_strerror(r) << dendl;
schedule_refresh_images(10);
}
m_async_op_tracker.finish_op();
if (on_init_finish != nullptr) {
on_init_finish->complete(r);
}
}
template <typename I>
void PoolWatcher<I>::unregister_watcher() {
dout(5) << dendl;
m_async_op_tracker.start_op();
Context *ctx = new LambdaContext([this](int r) {
dout(5) << "unregister_watcher: r=" << r << dendl;
if (r < 0) {
derr << "error unregistering watcher for "
<< m_mirroring_watcher->get_oid() << " object: " << cpp_strerror(r)
<< dendl;
}
m_async_op_tracker.finish_op();
});
m_mirroring_watcher->unregister_watch(ctx);
}
template <typename I>
void PoolWatcher<I>::refresh_images() {
dout(5) << dendl;
{
std::lock_guard locker{m_lock};
ceph_assert(m_image_ids_invalid);
ceph_assert(m_refresh_in_progress);
// clear all pending notification events since we need to perform
// a full image list refresh
m_pending_added_image_ids.clear();
m_pending_removed_image_ids.clear();
}
m_async_op_tracker.start_op();
m_refresh_image_ids.clear();
Context *ctx = create_context_callback<
PoolWatcher, &PoolWatcher<I>::handle_refresh_images>(this);
auto req = pool_watcher::RefreshImagesRequest<I>::create(m_io_ctx,
&m_refresh_image_ids,
ctx);
req->send();
}
template <typename I>
void PoolWatcher<I>::handle_refresh_images(int r) {
dout(5) << "r=" << r << dendl;
bool deferred_refresh = false;
bool retry_refresh = false;
Context *on_init_finish = nullptr;
{
std::lock_guard locker{m_lock};
ceph_assert(m_image_ids_invalid);
ceph_assert(m_refresh_in_progress);
m_refresh_in_progress = false;
if (r == -ENOENT) {
dout(5) << "mirroring directory not found" << dendl;
r = 0;
m_refresh_image_ids.clear();
}
if (m_deferred_refresh) {
// need to refresh -- skip the notification
deferred_refresh = true;
} else if (r >= 0) {
m_pending_image_ids = std::move(m_refresh_image_ids);
m_image_ids_invalid = false;
std::swap(on_init_finish, m_on_init_finish);
schedule_listener();
} else if (r == -EBLOCKLISTED) {
dout(0) << "detected client is blocklisted during image refresh" << dendl;
m_blocklisted = true;
std::swap(on_init_finish, m_on_init_finish);
} else {
retry_refresh = true;
}
}
if (deferred_refresh) {
dout(5) << "scheduling deferred refresh" << dendl;
schedule_refresh_images(0);
} else if (retry_refresh) {
derr << "failed to retrieve mirroring directory: " << cpp_strerror(r)
<< dendl;
schedule_refresh_images(10);
}
m_async_op_tracker.finish_op();
if (on_init_finish != nullptr) {
on_init_finish->complete(r);
}
}
template <typename I>
void PoolWatcher<I>::schedule_refresh_images(double interval) {
std::scoped_lock locker{m_threads->timer_lock, m_lock};
if (m_shutting_down || m_refresh_in_progress || m_timer_ctx != nullptr) {
if (m_refresh_in_progress && !m_deferred_refresh) {
dout(5) << "deferring refresh until in-flight refresh completes" << dendl;
m_deferred_refresh = true;
}
return;
}
m_image_ids_invalid = true;
m_timer_ctx = m_threads->timer->add_event_after(
interval,
new LambdaContext([this](int r) {
process_refresh_images();
}));
}
template <typename I>
void PoolWatcher<I>::handle_rewatch_complete(int r) {
dout(5) << "r=" << r << dendl;
if (r == -EBLOCKLISTED) {
dout(0) << "detected client is blocklisted" << dendl;
std::lock_guard locker{m_lock};
m_blocklisted = true;
return;
} else if (r == -ENOENT) {
dout(5) << "mirroring directory deleted" << dendl;
} else if (r < 0) {
derr << "unexpected error re-registering mirroring directory watch: "
<< cpp_strerror(r) << dendl;
}
schedule_refresh_images(5);
}
template <typename I>
void PoolWatcher<I>::handle_image_updated(const std::string &id,
const std::string &global_image_id,
bool enabled) {
dout(10) << "image_id=" << id << ", "
<< "global_image_id=" << global_image_id << ", "
<< "enabled=" << enabled << dendl;
std::lock_guard locker{m_lock};
ImageId image_id(global_image_id, id);
m_pending_added_image_ids.erase(image_id);
m_pending_removed_image_ids.erase(image_id);
if (enabled) {
m_pending_added_image_ids.insert(image_id);
schedule_listener();
} else {
m_pending_removed_image_ids.insert(image_id);
schedule_listener();
}
}
template <typename I>
void PoolWatcher<I>::process_refresh_images() {
ceph_assert(ceph_mutex_is_locked(m_threads->timer_lock));
ceph_assert(m_timer_ctx != nullptr);
m_timer_ctx = nullptr;
{
std::lock_guard locker{m_lock};
ceph_assert(!m_refresh_in_progress);
m_refresh_in_progress = true;
m_deferred_refresh = false;
}
// execute outside of the timer's lock
m_async_op_tracker.start_op();
Context *ctx = new LambdaContext([this](int r) {
register_watcher();
m_async_op_tracker.finish_op();
});
m_threads->work_queue->queue(ctx, 0);
}
template <typename I>
void PoolWatcher<I>::schedule_listener() {
ceph_assert(ceph_mutex_is_locked(m_lock));
m_pending_updates = true;
if (m_shutting_down || m_image_ids_invalid || m_notify_listener_in_progress) {
return;
}
dout(20) << dendl;
m_async_op_tracker.start_op();
Context *ctx = new LambdaContext([this](int r) {
notify_listener();
m_async_op_tracker.finish_op();
});
m_notify_listener_in_progress = true;
m_threads->work_queue->queue(ctx, 0);
}
template <typename I>
void PoolWatcher<I>::notify_listener() {
dout(10) << dendl;
std::string mirror_uuid;
ImageIds added_image_ids;
ImageIds removed_image_ids;
{
std::lock_guard locker{m_lock};
ceph_assert(m_notify_listener_in_progress);
}
if (!removed_image_ids.empty()) {
m_listener.handle_update(mirror_uuid, {}, std::move(removed_image_ids));
removed_image_ids.clear();
}
{
std::lock_guard locker{m_lock};
ceph_assert(m_notify_listener_in_progress);
// if the watch failed while we didn't own the lock, we are going
// to need to perform a full refresh
if (m_image_ids_invalid) {
m_notify_listener_in_progress = false;
return;
}
// merge add/remove notifications into pending set (a given image
// can only be in one set or another)
for (auto &image_id : m_pending_removed_image_ids) {
dout(20) << "image_id=" << image_id << dendl;
m_pending_image_ids.erase(image_id);
}
for (auto &image_id : m_pending_added_image_ids) {
dout(20) << "image_id=" << image_id << dendl;
m_pending_image_ids.erase(image_id);
m_pending_image_ids.insert(image_id);
}
m_pending_added_image_ids.clear();
// compute added/removed images
for (auto &image_id : m_image_ids) {
auto it = m_pending_image_ids.find(image_id);
if (it == m_pending_image_ids.end() || it->id != image_id.id) {
removed_image_ids.insert(image_id);
}
}
for (auto &image_id : m_pending_image_ids) {
auto it = m_image_ids.find(image_id);
if (it == m_image_ids.end() || it->id != image_id.id) {
added_image_ids.insert(image_id);
}
}
m_pending_updates = false;
m_image_ids = m_pending_image_ids;
}
m_listener.handle_update(m_mirror_uuid, std::move(added_image_ids),
std::move(removed_image_ids));
{
std::lock_guard locker{m_lock};
m_notify_listener_in_progress = false;
if (m_pending_updates) {
schedule_listener();
}
}
}
} // namespace mirror
} // namespace rbd
template class rbd::mirror::PoolWatcher<librbd::ImageCtx>;
| 13,360 | 27.187764 | 80 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/PoolWatcher.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef CEPH_RBD_MIRROR_POOL_WATCHER_H
#define CEPH_RBD_MIRROR_POOL_WATCHER_H
#include <map>
#include <memory>
#include <set>
#include <string>
#include "common/AsyncOpTracker.h"
#include "common/ceph_context.h"
#include "common/ceph_mutex.h"
#include "include/rados/librados.hpp"
#include "tools/rbd_mirror/Types.h"
#include <boost/functional/hash.hpp>
#include <boost/optional.hpp>
#include "include/ceph_assert.h"
#include "tools/rbd_mirror/pool_watcher/Types.h"
namespace librbd { struct ImageCtx; }
namespace rbd {
namespace mirror {
template <typename> struct Threads;
/**
* Keeps track of images that have mirroring enabled within all
* pools.
*/
template <typename ImageCtxT = librbd::ImageCtx>
class PoolWatcher {
public:
static PoolWatcher* create(Threads<ImageCtxT> *threads,
librados::IoCtx &io_ctx,
const std::string& mirror_uuid,
pool_watcher::Listener &listener) {
return new PoolWatcher(threads, io_ctx, mirror_uuid, listener);
}
PoolWatcher(Threads<ImageCtxT> *threads,
librados::IoCtx &io_ctx,
const std::string& mirror_uuid,
pool_watcher::Listener &listener);
~PoolWatcher();
PoolWatcher(const PoolWatcher&) = delete;
PoolWatcher& operator=(const PoolWatcher&) = delete;
bool is_blocklisted() const;
void init(Context *on_finish = nullptr);
void shut_down(Context *on_finish);
inline uint64_t get_image_count() const {
std::lock_guard locker{m_lock};
return m_image_ids.size();
}
private:
/**
* @verbatim
*
* <start>
* |
* v
* INIT
* |
* v
* REGISTER_WATCHER
* |
* |/--------------------------------\
* | |
* v |
* REFRESH_IMAGES |
* | |
* |/----------------------------\ |
* | | |
* v | |
* NOTIFY_LISTENER | |
* | | |
* v | |
* IDLE ---\ | |
* | | | |
* | |\---> IMAGE_UPDATED | |
* | | | | |
* | | v | |
* | | GET_IMAGE_NAME --/ |
* | | |
* | \----> WATCH_ERROR ---------/
* v
* SHUT_DOWN
* |
* v
* UNREGISTER_WATCHER
* |
* v
* <finish>
*
* @endverbatim
*/
class MirroringWatcher;
Threads<ImageCtxT> *m_threads;
librados::IoCtx m_io_ctx;
std::string m_mirror_uuid;
pool_watcher::Listener &m_listener;
ImageIds m_refresh_image_ids;
bufferlist m_out_bl;
mutable ceph::mutex m_lock;
Context *m_on_init_finish = nullptr;
ImageIds m_image_ids;
bool m_pending_updates = false;
bool m_notify_listener_in_progress = false;
ImageIds m_pending_image_ids;
ImageIds m_pending_added_image_ids;
ImageIds m_pending_removed_image_ids;
MirroringWatcher *m_mirroring_watcher;
Context *m_timer_ctx = nullptr;
AsyncOpTracker m_async_op_tracker;
bool m_blocklisted = false;
bool m_shutting_down = false;
bool m_image_ids_invalid = true;
bool m_refresh_in_progress = false;
bool m_deferred_refresh = false;
void register_watcher();
void handle_register_watcher(int r);
void unregister_watcher();
void refresh_images();
void handle_refresh_images(int r);
void schedule_refresh_images(double interval);
void process_refresh_images();
void handle_rewatch_complete(int r);
void handle_image_updated(const std::string &image_id,
const std::string &global_image_id,
bool enabled);
void schedule_listener();
void notify_listener();
};
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::PoolWatcher<librbd::ImageCtx>;
#endif // CEPH_RBD_MIRROR_POOL_WATCHER_H
| 4,213 | 25.012346 | 70 |
h
|
null |
ceph-main/src/tools/rbd_mirror/ProgressContext.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_PROGRESS_CONTEXT_H
#define RBD_MIRROR_PROGRESS_CONTEXT_H
namespace rbd {
namespace mirror {
class ProgressContext
{
public:
virtual ~ProgressContext() {}
virtual void update_progress(const std::string &description,
bool flush = true) = 0;
};
} // namespace mirror
} // namespace rbd
#endif // RBD_MIRROR_PROGRESS_CONTEXT_H
| 459 | 19.909091 | 70 |
h
|
null |
ceph-main/src/tools/rbd_mirror/RemotePoolPoller.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "RemotePoolPoller.h"
#include "include/ceph_assert.h"
#include "common/debug.h"
#include "common/errno.h"
#include "common/Timer.h"
#include "cls/rbd/cls_rbd_client.h"
#include "librbd/ImageCtx.h"
#include "librbd/Utils.h"
#include "librbd/asio/ContextWQ.h"
#include "tools/rbd_mirror/Threads.h"
#include "tools/rbd_mirror/Types.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::RemotePoolPoller: " << this << " " \
<< __func__ << ": "
namespace rbd {
namespace mirror {
static const double POLL_INTERVAL_SECONDS = 30;
using librbd::util::create_rados_callback;
template <typename I>
RemotePoolPoller<I>::~RemotePoolPoller() {
ceph_assert(m_timer_task == nullptr);
}
template <typename I>
void RemotePoolPoller<I>::init(Context* on_finish) {
dout(10) << dendl;
ceph_assert(m_state == STATE_INITIALIZING);
ceph_assert(m_on_finish == nullptr);
m_on_finish = on_finish;
get_mirror_uuid();
}
template <typename I>
void RemotePoolPoller<I>::shut_down(Context* on_finish) {
dout(10) << dendl;
std::unique_lock locker(m_threads->timer_lock);
ceph_assert(m_state == STATE_POLLING);
m_state = STATE_SHUTTING_DOWN;
if (m_timer_task == nullptr) {
// currently executing a poll
ceph_assert(m_on_finish == nullptr);
m_on_finish = on_finish;
return;
}
m_threads->timer->cancel_event(m_timer_task);
m_timer_task = nullptr;
m_threads->work_queue->queue(on_finish, 0);
}
template <typename I>
void RemotePoolPoller<I>::get_mirror_uuid() {
dout(10) << dendl;
librados::ObjectReadOperation op;
librbd::cls_client::mirror_uuid_get_start(&op);
auto aio_comp = create_rados_callback<
RemotePoolPoller<I>, &RemotePoolPoller<I>::handle_get_mirror_uuid>(this);
m_out_bl.clear();
int r = m_remote_io_ctx.aio_operate(RBD_MIRRORING, aio_comp, &op, &m_out_bl);
ceph_assert(r == 0);
aio_comp->release();
}
template <typename I>
void RemotePoolPoller<I>::handle_get_mirror_uuid(int r) {
dout(10) << "r=" << r << dendl;
std::string remote_mirror_uuid;
if (r >= 0) {
auto it = m_out_bl.cbegin();
r = librbd::cls_client::mirror_uuid_get_finish(&it, &remote_mirror_uuid);
if (r >= 0 && remote_mirror_uuid.empty()) {
r = -ENOENT;
}
}
if (r < 0) {
if (r == -ENOENT) {
dout(5) << "remote mirror uuid missing" << dendl;
} else {
derr << "failed to retrieve remote mirror uuid: " << cpp_strerror(r)
<< dendl;
}
m_remote_pool_meta.mirror_uuid = "";
}
// if we have the mirror uuid, we will poll until shut down
if (m_state == STATE_INITIALIZING) {
if (r < 0) {
schedule_task(r);
return;
}
m_state = STATE_POLLING;
}
dout(10) << "remote_mirror_uuid=" << remote_mirror_uuid << dendl;
if (m_remote_pool_meta.mirror_uuid != remote_mirror_uuid) {
m_remote_pool_meta.mirror_uuid = remote_mirror_uuid;
m_updated = true;
}
mirror_peer_ping();
}
template <typename I>
void RemotePoolPoller<I>::mirror_peer_ping() {
dout(10) << dendl;
librados::ObjectWriteOperation op;
librbd::cls_client::mirror_peer_ping(&op, m_site_name, m_local_mirror_uuid);
auto aio_comp = create_rados_callback<
RemotePoolPoller<I>, &RemotePoolPoller<I>::handle_mirror_peer_ping>(this);
int r = m_remote_io_ctx.aio_operate(RBD_MIRRORING, aio_comp, &op);
ceph_assert(r == 0);
aio_comp->release();
}
template <typename I>
void RemotePoolPoller<I>::handle_mirror_peer_ping(int r) {
dout(10) << "r=" << r << dendl;
if (r == -EOPNOTSUPP) {
// older OSD that doesn't support snapshot-based mirroring, so no need
// to query remote peers
dout(10) << "remote peer does not support snapshot-based mirroring"
<< dendl;
notify_listener();
return;
} else if (r < 0) {
// we can still see if we can perform a peer list and find ourselves
derr << "failed to ping remote mirror peer: " << cpp_strerror(r) << dendl;
}
mirror_peer_list();
}
template <typename I>
void RemotePoolPoller<I>::mirror_peer_list() {
dout(10) << dendl;
librados::ObjectReadOperation op;
librbd::cls_client::mirror_peer_list_start(&op);
auto aio_comp = create_rados_callback<
RemotePoolPoller<I>, &RemotePoolPoller<I>::handle_mirror_peer_list>(this);
m_out_bl.clear();
int r = m_remote_io_ctx.aio_operate(RBD_MIRRORING, aio_comp, &op, &m_out_bl);
ceph_assert(r == 0);
aio_comp->release();
}
template <typename I>
void RemotePoolPoller<I>::handle_mirror_peer_list(int r) {
dout(10) << "r=" << r << dendl;
std::vector<cls::rbd::MirrorPeer> peers;
if (r == 0) {
auto iter = m_out_bl.cbegin();
r = librbd::cls_client::mirror_peer_list_finish(&iter, &peers);
}
if (r < 0) {
derr << "failed to retrieve mirror peers: " << cpp_strerror(r) << dendl;
}
cls::rbd::MirrorPeer* matched_peer = nullptr;
for (auto& peer : peers) {
if (peer.mirror_peer_direction == cls::rbd::MIRROR_PEER_DIRECTION_RX) {
continue;
}
if (peer.mirror_uuid == m_local_mirror_uuid) {
matched_peer = &peer;
break;
} else if (peer.site_name == m_site_name) {
// keep searching in case we hit an exact match by fsid
matched_peer = &peer;
}
}
// older OSDs don't support peer ping so we might fail to find a match,
// which will prevent snapshot mirroring from functioning
std::string remote_mirror_peer_uuid;
if (matched_peer != nullptr) {
remote_mirror_peer_uuid = matched_peer->uuid;
}
dout(10) << "remote_mirror_peer_uuid=" << remote_mirror_peer_uuid << dendl;
if (m_remote_pool_meta.mirror_peer_uuid != remote_mirror_peer_uuid) {
m_remote_pool_meta.mirror_peer_uuid = remote_mirror_peer_uuid;
m_updated = true;
}
notify_listener();
}
template <typename I>
void RemotePoolPoller<I>::notify_listener() {
bool updated = false;
std::swap(updated, m_updated);
if (updated) {
dout(10) << dendl;
m_listener.handle_updated(m_remote_pool_meta);
}
schedule_task(0);
}
template <typename I>
void RemotePoolPoller<I>::schedule_task(int r) {
std::unique_lock locker{m_threads->timer_lock};
if (m_state == STATE_POLLING) {
dout(10) << dendl;
ceph_assert(m_timer_task == nullptr);
m_timer_task = new LambdaContext([this](int) {
handle_task();
});
m_threads->timer->add_event_after(POLL_INTERVAL_SECONDS, m_timer_task);
}
// finish init or shut down callback
if (m_on_finish != nullptr) {
locker.unlock();
Context* on_finish = nullptr;
std::swap(on_finish, m_on_finish);
on_finish->complete(m_state == STATE_SHUTTING_DOWN ? 0 : r);
}
}
template <typename I>
void RemotePoolPoller<I>::handle_task() {
dout(10) << dendl;
ceph_assert(ceph_mutex_is_locked_by_me(m_threads->timer_lock));
m_timer_task = nullptr;
auto ctx = new LambdaContext([this](int) {
get_mirror_uuid();
});
m_threads->work_queue->queue(ctx);
}
} // namespace mirror
} // namespace rbd
template class rbd::mirror::RemotePoolPoller<librbd::ImageCtx>;
| 7,198 | 25.86194 | 80 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/RemotePoolPoller.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef CEPH_RBD_MIRROR_REMOTE_POOL_POLLER_H
#define CEPH_RBD_MIRROR_REMOTE_POOL_POLLER_H
#include "include/rados/librados.hpp"
#include "tools/rbd_mirror/Types.h"
#include <string>
struct Context;
namespace librbd { struct ImageCtx; }
namespace rbd {
namespace mirror {
template <typename> struct Threads;
namespace remote_pool_poller {
struct Listener {
virtual ~Listener() {}
virtual void handle_updated(const RemotePoolMeta& remote_pool_meta) = 0;
};
}; // namespace remote_pool_poller
template <typename ImageCtxT>
class RemotePoolPoller {
public:
static RemotePoolPoller* create(
Threads<ImageCtxT>* threads,
librados::IoCtx& remote_io_ctx,
const std::string& site_name,
const std::string& local_mirror_uuid,
remote_pool_poller::Listener& listener) {
return new RemotePoolPoller(threads, remote_io_ctx, site_name,
local_mirror_uuid, listener);
}
RemotePoolPoller(
Threads<ImageCtxT>* threads,
librados::IoCtx& remote_io_ctx,
const std::string& site_name,
const std::string& local_mirror_uuid,
remote_pool_poller::Listener& listener)
: m_threads(threads),
m_remote_io_ctx(remote_io_ctx),
m_site_name(site_name),
m_local_mirror_uuid(local_mirror_uuid),
m_listener(listener) {
}
~RemotePoolPoller();
void init(Context* on_finish);
void shut_down(Context* on_finish);
private:
/**
* @verbatim
*
* <start>
* |
* |/----------------------------\
* | |
* v |
* MIRROR_UUID_GET |
* | |
* v |
* MIRROR_PEER_PING |
* | |
* v |
* MIRROR_PEER_LIST |
* | |
* v |
* MIRROR_UUID_GET |
* | |
* v (skip if no changes) |
* NOTIFY_LISTENER |
* | |
* | (repeat periodically) |
* |\----------------------------/
* |
* v
* <finish>
*
* @endverbatim
*/
enum State {
STATE_INITIALIZING,
STATE_POLLING,
STATE_SHUTTING_DOWN
};
Threads<ImageCtxT>* m_threads;
librados::IoCtx& m_remote_io_ctx;
std::string m_site_name;
std::string m_local_mirror_uuid;
remote_pool_poller::Listener& m_listener;
bufferlist m_out_bl;
RemotePoolMeta m_remote_pool_meta;
bool m_updated = false;
State m_state = STATE_INITIALIZING;
Context* m_timer_task = nullptr;
Context* m_on_finish = nullptr;
void get_mirror_uuid();
void handle_get_mirror_uuid(int r);
void mirror_peer_ping();
void handle_mirror_peer_ping(int r);
void mirror_peer_list();
void handle_mirror_peer_list(int r);
void notify_listener();
void schedule_task(int r);
void handle_task();
};
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::RemotePoolPoller<librbd::ImageCtx>;
#endif // CEPH_RBD_MIRROR_REMOTE_POOL_POLLER_H
| 3,308 | 23.69403 | 74 |
h
|
null |
ceph-main/src/tools/rbd_mirror/ServiceDaemon.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "tools/rbd_mirror/ServiceDaemon.h"
#include "include/Context.h"
#include "include/stringify.h"
#include "common/ceph_context.h"
#include "common/config.h"
#include "common/debug.h"
#include "common/errno.h"
#include "common/Formatter.h"
#include "common/Timer.h"
#include "tools/rbd_mirror/Threads.h"
#include <sstream>
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::ServiceDaemon: " << this << " " \
<< __func__ << ": "
namespace rbd {
namespace mirror {
namespace {
const std::string RBD_MIRROR_AUTH_ID_PREFIX("rbd-mirror.");
struct AttributeDumpVisitor : public boost::static_visitor<void> {
ceph::Formatter *f;
const std::string& name;
AttributeDumpVisitor(ceph::Formatter *f, const std::string& name)
: f(f), name(name) {
}
void operator()(bool val) const {
f->dump_bool(name.c_str(), val);
}
void operator()(uint64_t val) const {
f->dump_unsigned(name.c_str(), val);
}
void operator()(const std::string& val) const {
f->dump_string(name.c_str(), val);
}
};
} // anonymous namespace
using namespace service_daemon;
template <typename I>
ServiceDaemon<I>::ServiceDaemon(CephContext *cct, RadosRef rados,
Threads<I>* threads)
: m_cct(cct), m_rados(rados), m_threads(threads) {
dout(20) << dendl;
}
template <typename I>
ServiceDaemon<I>::~ServiceDaemon() {
dout(20) << dendl;
std::lock_guard timer_locker{m_threads->timer_lock};
if (m_timer_ctx != nullptr) {
m_threads->timer->cancel_event(m_timer_ctx);
update_status();
}
}
template <typename I>
int ServiceDaemon<I>::init() {
dout(20) << dendl;
std::string id = m_cct->_conf->name.get_id();
if (id.find(RBD_MIRROR_AUTH_ID_PREFIX) == 0) {
id = id.substr(RBD_MIRROR_AUTH_ID_PREFIX.size());
}
std::string instance_id = stringify(m_rados->get_instance_id());
std::map<std::string, std::string> service_metadata = {
{"id", id}, {"instance_id", instance_id}};
int r = m_rados->service_daemon_register("rbd-mirror", instance_id,
service_metadata);
if (r < 0) {
return r;
}
return 0;
}
template <typename I>
void ServiceDaemon<I>::add_pool(int64_t pool_id, const std::string& pool_name) {
dout(20) << "pool_id=" << pool_id << ", pool_name=" << pool_name << dendl;
{
std::lock_guard locker{m_lock};
m_pools.insert({pool_id, {pool_name}});
}
schedule_update_status();
}
template <typename I>
void ServiceDaemon<I>::remove_pool(int64_t pool_id) {
dout(20) << "pool_id=" << pool_id << dendl;
{
std::lock_guard locker{m_lock};
m_pools.erase(pool_id);
}
schedule_update_status();
}
template <typename I>
void ServiceDaemon<I>::add_namespace(int64_t pool_id,
const std::string& namespace_name) {
dout(20) << "pool_id=" << pool_id << ", namespace=" << namespace_name
<< dendl;
std::lock_guard locker{m_lock};
auto pool_it = m_pools.find(pool_id);
if (pool_it == m_pools.end()) {
return;
}
pool_it->second.ns_attributes[namespace_name];
// don't schedule update status as the namespace attributes are empty yet
}
template <typename I>
void ServiceDaemon<I>::remove_namespace(int64_t pool_id,
const std::string& namespace_name) {
dout(20) << "pool_id=" << pool_id << ", namespace=" << namespace_name
<< dendl;
{
std::lock_guard locker{m_lock};
auto pool_it = m_pools.find(pool_id);
if (pool_it == m_pools.end()) {
return;
}
pool_it->second.ns_attributes.erase(namespace_name);
}
schedule_update_status();
}
template <typename I>
uint64_t ServiceDaemon<I>::add_or_update_callout(int64_t pool_id,
uint64_t callout_id,
CalloutLevel callout_level,
const std::string& text) {
dout(20) << "pool_id=" << pool_id << ", "
<< "callout_id=" << callout_id << ", "
<< "callout_level=" << callout_level << ", "
<< "text=" << text << dendl;
{
std::lock_guard locker{m_lock};
auto pool_it = m_pools.find(pool_id);
if (pool_it == m_pools.end()) {
return CALLOUT_ID_NONE;
}
if (callout_id == CALLOUT_ID_NONE) {
callout_id = ++m_callout_id;
}
pool_it->second.callouts[callout_id] = {callout_level, text};
}
schedule_update_status();
return callout_id;
}
template <typename I>
void ServiceDaemon<I>::remove_callout(int64_t pool_id, uint64_t callout_id) {
dout(20) << "pool_id=" << pool_id << ", "
<< "callout_id=" << callout_id << dendl;
{
std::lock_guard locker{m_lock};
auto pool_it = m_pools.find(pool_id);
if (pool_it == m_pools.end()) {
return;
}
pool_it->second.callouts.erase(callout_id);
}
schedule_update_status();
}
template <typename I>
void ServiceDaemon<I>::add_or_update_attribute(int64_t pool_id,
const std::string& key,
const AttributeValue& value) {
dout(20) << "pool_id=" << pool_id << ", "
<< "key=" << key << ", "
<< "value=" << value << dendl;
{
std::lock_guard locker{m_lock};
auto pool_it = m_pools.find(pool_id);
if (pool_it == m_pools.end()) {
return;
}
pool_it->second.attributes[key] = value;
}
schedule_update_status();
}
template <typename I>
void ServiceDaemon<I>::add_or_update_namespace_attribute(
int64_t pool_id, const std::string& namespace_name, const std::string& key,
const AttributeValue& value) {
if (namespace_name.empty()) {
add_or_update_attribute(pool_id, key, value);
return;
}
dout(20) << "pool_id=" << pool_id << ", "
<< "namespace=" << namespace_name << ", "
<< "key=" << key << ", "
<< "value=" << value << dendl;
{
std::lock_guard locker{m_lock};
auto pool_it = m_pools.find(pool_id);
if (pool_it == m_pools.end()) {
return;
}
auto ns_it = pool_it->second.ns_attributes.find(namespace_name);
if (ns_it == pool_it->second.ns_attributes.end()) {
return;
}
ns_it->second[key] = value;
}
schedule_update_status();
}
template <typename I>
void ServiceDaemon<I>::remove_attribute(int64_t pool_id,
const std::string& key) {
dout(20) << "pool_id=" << pool_id << ", "
<< "key=" << key << dendl;
{
std::lock_guard locker{m_lock};
auto pool_it = m_pools.find(pool_id);
if (pool_it == m_pools.end()) {
return;
}
pool_it->second.attributes.erase(key);
}
schedule_update_status();
}
template <typename I>
void ServiceDaemon<I>::schedule_update_status() {
std::lock_guard timer_locker{m_threads->timer_lock};
if (m_timer_ctx != nullptr) {
return;
}
dout(20) << dendl;
m_timer_ctx = new LambdaContext([this](int) {
m_timer_ctx = nullptr;
update_status();
});
m_threads->timer->add_event_after(1, m_timer_ctx);
}
template <typename I>
void ServiceDaemon<I>::update_status() {
ceph_assert(ceph_mutex_is_locked(m_threads->timer_lock));
ceph::JSONFormatter f;
{
std::lock_guard locker{m_lock};
f.open_object_section("pools");
for (auto& pool_pair : m_pools) {
f.open_object_section(stringify(pool_pair.first).c_str());
f.dump_string("name", pool_pair.second.name);
f.open_object_section("callouts");
for (auto& callout : pool_pair.second.callouts) {
f.open_object_section(stringify(callout.first).c_str());
f.dump_string("level", stringify(callout.second.level).c_str());
f.dump_string("text", callout.second.text.c_str());
f.close_section();
}
f.close_section(); // callouts
for (auto& attribute : pool_pair.second.attributes) {
AttributeDumpVisitor attribute_dump_visitor(&f, attribute.first);
boost::apply_visitor(attribute_dump_visitor, attribute.second);
}
if (!pool_pair.second.ns_attributes.empty()) {
f.open_object_section("namespaces");
for (auto& [ns, attributes] : pool_pair.second.ns_attributes) {
f.open_object_section(ns.c_str());
for (auto& [key, value] : attributes) {
AttributeDumpVisitor attribute_dump_visitor(&f, key);
boost::apply_visitor(attribute_dump_visitor, value);
}
f.close_section(); // namespace
}
f.close_section(); // namespaces
}
f.close_section(); // pool
}
f.close_section(); // pools
}
std::stringstream ss;
f.flush(ss);
dout(20) << ss.str() << dendl;
int r = m_rados->service_daemon_update_status({{"json", ss.str()}});
if (r < 0) {
derr << "failed to update service daemon status: " << cpp_strerror(r)
<< dendl;
}
}
} // namespace mirror
} // namespace rbd
template class rbd::mirror::ServiceDaemon<librbd::ImageCtx>;
| 9,242 | 26.924471 | 80 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/ServiceDaemon.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef CEPH_RBD_MIRROR_SERVICE_DAEMON_H
#define CEPH_RBD_MIRROR_SERVICE_DAEMON_H
#include "common/ceph_mutex.h"
#include "include/common_fwd.h"
#include "tools/rbd_mirror/Types.h"
#include "tools/rbd_mirror/service_daemon/Types.h"
#include <map>
#include <string>
struct Context;
namespace librbd { struct ImageCtx; }
namespace rbd {
namespace mirror {
template <typename> struct Threads;
template <typename ImageCtxT = librbd::ImageCtx>
class ServiceDaemon {
public:
ServiceDaemon(CephContext *cct, RadosRef rados, Threads<ImageCtxT>* threads);
~ServiceDaemon();
int init();
void add_pool(int64_t pool_id, const std::string& pool_name);
void remove_pool(int64_t pool_id);
void add_namespace(int64_t pool_id, const std::string& namespace_name);
void remove_namespace(int64_t pool_id, const std::string& namespace_name);
uint64_t add_or_update_callout(int64_t pool_id, uint64_t callout_id,
service_daemon::CalloutLevel callout_level,
const std::string& text);
void remove_callout(int64_t pool_id, uint64_t callout_id);
void add_or_update_attribute(int64_t pool_id, const std::string& key,
const service_daemon::AttributeValue& value);
void add_or_update_namespace_attribute(
int64_t pool_id, const std::string& namespace_name,
const std::string& key, const service_daemon::AttributeValue& value);
void remove_attribute(int64_t pool_id, const std::string& key);
private:
struct Callout {
service_daemon::CalloutLevel level;
std::string text;
Callout() : level(service_daemon::CALLOUT_LEVEL_INFO) {
}
Callout(service_daemon::CalloutLevel level, const std::string& text)
: level(level), text(text) {
}
};
typedef std::map<uint64_t, Callout> Callouts;
typedef std::map<std::string, service_daemon::AttributeValue> Attributes;
typedef std::map<std::string, Attributes> NamespaceAttributes;
struct Pool {
std::string name;
Callouts callouts;
Attributes attributes;
NamespaceAttributes ns_attributes;
Pool(const std::string& name) : name(name) {
}
};
typedef std::map<int64_t, Pool> Pools;
CephContext *m_cct;
RadosRef m_rados;
Threads<ImageCtxT>* m_threads;
ceph::mutex m_lock = ceph::make_mutex("rbd::mirror::ServiceDaemon");
Pools m_pools;
uint64_t m_callout_id = service_daemon::CALLOUT_ID_NONE;
Context* m_timer_ctx = nullptr;
void schedule_update_status();
void update_status();
};
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::ServiceDaemon<librbd::ImageCtx>;
#endif // CEPH_RBD_MIRROR_SERVICE_DAEMON_H
| 2,775 | 28.221053 | 79 |
h
|
null |
ceph-main/src/tools/rbd_mirror/Threads.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "tools/rbd_mirror/Threads.h"
#include "common/Timer.h"
#include "librbd/AsioEngine.h"
#include "librbd/ImageCtx.h"
#include "librbd/asio/ContextWQ.h"
namespace rbd {
namespace mirror {
template <typename I>
Threads<I>::Threads(std::shared_ptr<librados::Rados>& rados) {
auto cct = static_cast<CephContext*>(rados->cct());
asio_engine = new librbd::AsioEngine(rados);
work_queue = asio_engine->get_work_queue();
timer = new SafeTimer(cct, timer_lock, true);
timer->init();
}
template <typename I>
Threads<I>::~Threads() {
{
std::lock_guard timer_locker{timer_lock};
timer->shutdown();
}
delete timer;
work_queue->drain();
delete asio_engine;
}
} // namespace mirror
} // namespace rbd
template class rbd::mirror::Threads<librbd::ImageCtx>;
| 882 | 21.641026 | 70 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/Threads.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef CEPH_RBD_MIRROR_THREADS_H
#define CEPH_RBD_MIRROR_THREADS_H
#include "include/common_fwd.h"
#include "include/rados/librados_fwd.hpp"
#include "common/ceph_mutex.h"
#include "common/Timer.h"
#include <memory>
class ThreadPool;
namespace librbd {
struct AsioEngine;
struct ImageCtx;
namespace asio { struct ContextWQ; }
} // namespace librbd
namespace rbd {
namespace mirror {
template <typename ImageCtxT = librbd::ImageCtx>
class Threads {
public:
librbd::AsioEngine* asio_engine = nullptr;
librbd::asio::ContextWQ* work_queue = nullptr;
SafeTimer *timer = nullptr;
ceph::mutex timer_lock = ceph::make_mutex("Threads::timer_lock");
explicit Threads(std::shared_ptr<librados::Rados>& rados);
Threads(const Threads&) = delete;
Threads& operator=(const Threads&) = delete;
~Threads();
};
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::Threads<librbd::ImageCtx>;
#endif // CEPH_RBD_MIRROR_THREADS_H
| 1,059 | 22.043478 | 70 |
h
|
null |
ceph-main/src/tools/rbd_mirror/Throttler.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
/*
* Ceph - scalable distributed file system
*
* Copyright (C) 2016 SUSE LINUX GmbH
*
* This is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License version 2.1, as published by the Free Software
* Foundation. See file COPYING.
*
*/
#include "Throttler.h"
#include "common/Formatter.h"
#include "common/debug.h"
#include "common/errno.h"
#include "librbd/Utils.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::Throttler:: " << this \
<< " " << __func__ << ": "
namespace rbd {
namespace mirror {
template <typename I>
Throttler<I>::Throttler(CephContext *cct, const std::string &config_key)
: m_cct(cct), m_config_key(config_key),
m_config_keys{m_config_key.c_str(), nullptr},
m_lock(ceph::make_mutex(
librbd::util::unique_lock_name("rbd::mirror::Throttler", this))),
m_max_concurrent_ops(cct->_conf.get_val<uint64_t>(m_config_key)) {
dout(20) << m_config_key << "=" << m_max_concurrent_ops << dendl;
m_cct->_conf.add_observer(this);
}
template <typename I>
Throttler<I>::~Throttler() {
m_cct->_conf.remove_observer(this);
std::lock_guard locker{m_lock};
ceph_assert(m_inflight_ops.empty());
ceph_assert(m_queue.empty());
}
template <typename I>
void Throttler<I>::start_op(const std::string &ns,
const std::string &id_,
Context *on_start) {
Id id{ns, id_};
dout(20) << "id=" << id << dendl;
int r = 0;
{
std::lock_guard locker{m_lock};
if (m_inflight_ops.count(id) > 0) {
dout(20) << "duplicate for already started op " << id << dendl;
} else if (m_queued_ops.count(id) > 0) {
dout(20) << "duplicate for already queued op " << id << dendl;
std::swap(m_queued_ops[id], on_start);
r = -ENOENT;
} else if (m_max_concurrent_ops == 0 ||
m_inflight_ops.size() < m_max_concurrent_ops) {
ceph_assert(m_queue.empty());
m_inflight_ops.insert(id);
dout(20) << "ready to start op for " << id << " ["
<< m_inflight_ops.size() << "/" << m_max_concurrent_ops << "]"
<< dendl;
} else {
m_queue.push_back(id);
std::swap(m_queued_ops[id], on_start);
dout(20) << "op for " << id << " has been queued" << dendl;
}
}
if (on_start != nullptr) {
on_start->complete(r);
}
}
template <typename I>
bool Throttler<I>::cancel_op(const std::string &ns,
const std::string &id_) {
Id id{ns, id_};
dout(20) << "id=" << id << dendl;
Context *on_start = nullptr;
{
std::lock_guard locker{m_lock};
auto it = m_queued_ops.find(id);
if (it != m_queued_ops.end()) {
dout(20) << "canceled queued op for " << id << dendl;
m_queue.remove(id);
on_start = it->second;
m_queued_ops.erase(it);
}
}
if (on_start == nullptr) {
return false;
}
on_start->complete(-ECANCELED);
return true;
}
template <typename I>
void Throttler<I>::finish_op(const std::string &ns,
const std::string &id_) {
Id id{ns, id_};
dout(20) << "id=" << id << dendl;
if (cancel_op(ns, id_)) {
return;
}
Context *on_start = nullptr;
{
std::lock_guard locker{m_lock};
m_inflight_ops.erase(id);
if (m_inflight_ops.size() < m_max_concurrent_ops && !m_queue.empty()) {
auto id = m_queue.front();
auto it = m_queued_ops.find(id);
ceph_assert(it != m_queued_ops.end());
m_inflight_ops.insert(id);
dout(20) << "ready to start op for " << id << " ["
<< m_inflight_ops.size() << "/" << m_max_concurrent_ops << "]"
<< dendl;
on_start = it->second;
m_queued_ops.erase(it);
m_queue.pop_front();
}
}
if (on_start != nullptr) {
on_start->complete(0);
}
}
template <typename I>
void Throttler<I>::drain(const std::string &ns, int r) {
dout(20) << "ns=" << ns << dendl;
std::map<Id, Context *> queued_ops;
{
std::lock_guard locker{m_lock};
for (auto it = m_queued_ops.begin(); it != m_queued_ops.end(); ) {
if (it->first.first == ns) {
queued_ops[it->first] = it->second;
m_queue.remove(it->first);
it = m_queued_ops.erase(it);
} else {
it++;
}
}
for (auto it = m_inflight_ops.begin(); it != m_inflight_ops.end(); ) {
if (it->first == ns) {
dout(20) << "inflight_op " << *it << dendl;
it = m_inflight_ops.erase(it);
} else {
it++;
}
}
}
for (auto &it : queued_ops) {
dout(20) << "queued_op " << it.first << dendl;
it.second->complete(r);
}
}
template <typename I>
void Throttler<I>::set_max_concurrent_ops(uint32_t max) {
dout(20) << "max=" << max << dendl;
std::list<Context *> ops;
{
std::lock_guard locker{m_lock};
m_max_concurrent_ops = max;
// Start waiting ops in the case of available free slots
while ((m_max_concurrent_ops == 0 ||
m_inflight_ops.size() < m_max_concurrent_ops) &&
!m_queue.empty()) {
auto id = m_queue.front();
m_inflight_ops.insert(id);
dout(20) << "ready to start op for " << id << " ["
<< m_inflight_ops.size() << "/" << m_max_concurrent_ops << "]"
<< dendl;
auto it = m_queued_ops.find(id);
ceph_assert(it != m_queued_ops.end());
ops.push_back(it->second);
m_queued_ops.erase(it);
m_queue.pop_front();
}
}
for (const auto& ctx : ops) {
ctx->complete(0);
}
}
template <typename I>
void Throttler<I>::print_status(ceph::Formatter *f) {
dout(20) << dendl;
std::lock_guard locker{m_lock};
f->dump_int("max_parallel_requests", m_max_concurrent_ops);
f->dump_int("running_requests", m_inflight_ops.size());
f->dump_int("waiting_requests", m_queue.size());
}
template <typename I>
const char** Throttler<I>::get_tracked_conf_keys() const {
return m_config_keys;
}
template <typename I>
void Throttler<I>::handle_conf_change(const ConfigProxy& conf,
const std::set<std::string> &changed) {
if (changed.count(m_config_key)) {
set_max_concurrent_ops(conf.get_val<uint64_t>(m_config_key));
}
}
} // namespace mirror
} // namespace rbd
template class rbd::mirror::Throttler<librbd::ImageCtx>;
| 6,593 | 26.360996 | 77 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/Throttler.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_THROTTLER_H
#define RBD_MIRROR_THROTTLER_H
#include <list>
#include <map>
#include <set>
#include <sstream>
#include <string>
#include <utility>
#include "common/ceph_mutex.h"
#include "common/config_obs.h"
#include "include/common_fwd.h"
class Context;
namespace ceph { class Formatter; }
namespace librbd { class ImageCtx; }
namespace rbd {
namespace mirror {
template <typename ImageCtxT = librbd::ImageCtx>
class Throttler : public md_config_obs_t {
public:
static Throttler *create(
CephContext *cct,
const std::string &config_key) {
return new Throttler(cct, config_key);
}
void destroy() {
delete this;
}
Throttler(CephContext *cct,
const std::string &config_key);
~Throttler() override;
void set_max_concurrent_ops(uint32_t max);
void start_op(const std::string &ns, const std::string &id,
Context *on_start);
bool cancel_op(const std::string &ns, const std::string &id);
void finish_op(const std::string &ns, const std::string &id);
void drain(const std::string &ns, int r);
void print_status(ceph::Formatter *f);
private:
typedef std::pair<std::string, std::string> Id;
CephContext *m_cct;
const std::string m_config_key;
mutable const char* m_config_keys[2];
ceph::mutex m_lock;
uint32_t m_max_concurrent_ops;
std::list<Id> m_queue;
std::map<Id, Context *> m_queued_ops;
std::set<Id> m_inflight_ops;
const char **get_tracked_conf_keys() const override;
void handle_conf_change(const ConfigProxy& conf,
const std::set<std::string> &changed) override;
};
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::Throttler<librbd::ImageCtx>;
#endif // RBD_MIRROR_THROTTLER_H
| 1,856 | 23.76 | 73 |
h
|
null |
ceph-main/src/tools/rbd_mirror/Types.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "tools/rbd_mirror/Types.h"
namespace rbd {
namespace mirror {
std::ostream &operator<<(std::ostream &os, const ImageId &image_id) {
return os << "global id=" << image_id.global_id << ", "
<< "id=" << image_id.id;
}
std::ostream& operator<<(std::ostream& lhs,
const LocalPoolMeta& rhs) {
return lhs << "mirror_uuid=" << rhs.mirror_uuid;
}
std::ostream& operator<<(std::ostream& lhs,
const RemotePoolMeta& rhs) {
return lhs << "mirror_uuid=" << rhs.mirror_uuid << ", "
"mirror_peer_uuid=" << rhs.mirror_peer_uuid;
}
std::ostream& operator<<(std::ostream& lhs, const PeerSpec &peer) {
return lhs << "uuid: " << peer.uuid
<< " cluster: " << peer.cluster_name
<< " client: " << peer.client_name;
}
} // namespace mirror
} // namespace rbd
| 946 | 27.69697 | 70 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/Types.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef CEPH_RBD_MIRROR_TYPES_H
#define CEPH_RBD_MIRROR_TYPES_H
#include <iostream>
#include <memory>
#include <set>
#include <string>
#include <vector>
#include "include/rados/librados.hpp"
#include "include/rbd/librbd.hpp"
namespace rbd {
namespace mirror {
template <typename> struct MirrorStatusUpdater;
// Performance counters
enum {
l_rbd_mirror_journal_first = 27000,
l_rbd_mirror_journal_entries,
l_rbd_mirror_journal_replay_bytes,
l_rbd_mirror_journal_replay_latency,
l_rbd_mirror_journal_last,
l_rbd_mirror_snapshot_first,
l_rbd_mirror_snapshot_snapshots,
l_rbd_mirror_snapshot_sync_time,
l_rbd_mirror_snapshot_sync_bytes,
// per-image only counters below
l_rbd_mirror_snapshot_remote_timestamp,
l_rbd_mirror_snapshot_local_timestamp,
l_rbd_mirror_snapshot_last_sync_time,
l_rbd_mirror_snapshot_last_sync_bytes,
l_rbd_mirror_snapshot_last,
};
typedef std::shared_ptr<librados::Rados> RadosRef;
typedef std::shared_ptr<librados::IoCtx> IoCtxRef;
typedef std::shared_ptr<librbd::Image> ImageRef;
struct ImageId {
std::string global_id;
std::string id;
explicit ImageId(const std::string &global_id) : global_id(global_id) {
}
ImageId(const std::string &global_id, const std::string &id)
: global_id(global_id), id(id) {
}
inline bool operator==(const ImageId &rhs) const {
return (global_id == rhs.global_id && id == rhs.id);
}
inline bool operator<(const ImageId &rhs) const {
return global_id < rhs.global_id;
}
};
std::ostream &operator<<(std::ostream &, const ImageId &image_id);
typedef std::set<ImageId> ImageIds;
struct LocalPoolMeta {
LocalPoolMeta() {}
LocalPoolMeta(const std::string& mirror_uuid)
: mirror_uuid(mirror_uuid) {
}
std::string mirror_uuid;
};
std::ostream& operator<<(std::ostream& lhs,
const LocalPoolMeta& local_pool_meta);
struct RemotePoolMeta {
RemotePoolMeta() {}
RemotePoolMeta(const std::string& mirror_uuid,
const std::string& mirror_peer_uuid)
: mirror_uuid(mirror_uuid),
mirror_peer_uuid(mirror_peer_uuid) {
}
std::string mirror_uuid;
std::string mirror_peer_uuid;
};
std::ostream& operator<<(std::ostream& lhs,
const RemotePoolMeta& remote_pool_meta);
template <typename I>
struct Peer {
std::string uuid;
mutable librados::IoCtx io_ctx;
RemotePoolMeta remote_pool_meta;
MirrorStatusUpdater<I>* mirror_status_updater = nullptr;
Peer() {
}
Peer(const std::string& uuid,
librados::IoCtx& io_ctx,
const RemotePoolMeta& remote_pool_meta,
MirrorStatusUpdater<I>* mirror_status_updater)
: io_ctx(io_ctx),
remote_pool_meta(remote_pool_meta),
mirror_status_updater(mirror_status_updater) {
}
inline bool operator<(const Peer &rhs) const {
return uuid < rhs.uuid;
}
};
template <typename I>
std::ostream& operator<<(std::ostream& lhs, const Peer<I>& peer) {
return lhs << peer.remote_pool_meta;
}
struct PeerSpec {
PeerSpec() = default;
PeerSpec(const std::string &uuid, const std::string &cluster_name,
const std::string &client_name)
: uuid(uuid), cluster_name(cluster_name), client_name(client_name)
{
}
PeerSpec(const librbd::mirror_peer_site_t &peer) :
uuid(peer.uuid),
cluster_name(peer.site_name),
client_name(peer.client_name)
{
}
std::string uuid;
std::string cluster_name;
std::string client_name;
/// optional config properties
std::string mon_host;
std::string key;
bool operator==(const PeerSpec& rhs) const {
return (uuid == rhs.uuid &&
cluster_name == rhs.cluster_name &&
client_name == rhs.client_name &&
mon_host == rhs.mon_host &&
key == rhs.key);
}
bool operator<(const PeerSpec& rhs) const {
if (uuid != rhs.uuid) {
return uuid < rhs.uuid;
} else if (cluster_name != rhs.cluster_name) {
return cluster_name < rhs.cluster_name;
} else if (client_name != rhs.client_name) {
return client_name < rhs.client_name;
} else if (mon_host < rhs.mon_host) {
return mon_host < rhs.mon_host;
} else {
return key < rhs.key;
}
}
};
std::ostream& operator<<(std::ostream& lhs, const PeerSpec &peer);
} // namespace mirror
} // namespace rbd
#endif // CEPH_RBD_MIRROR_TYPES_H
| 4,427 | 24.744186 | 73 |
h
|
null |
ceph-main/src/tools/rbd_mirror/main.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "common/ceph_argparse.h"
#include "common/config.h"
#include "common/debug.h"
#include "common/errno.h"
#include "common/perf_counters.h"
#include "global/global_init.h"
#include "global/signal_handler.h"
#include "Mirror.h"
#include "Types.h"
#include <vector>
rbd::mirror::Mirror *mirror = nullptr;
PerfCounters *g_journal_perf_counters = nullptr;
PerfCounters *g_snapshot_perf_counters = nullptr;
void usage() {
std::cout << "usage: rbd-mirror [options...]" << std::endl;
std::cout << "options:\n";
std::cout << " -m monaddress[:port] connect to specified monitor\n";
std::cout << " --keyring=<path> path to keyring for local cluster\n";
std::cout << " --log-file=<logfile> file to log debug output\n";
std::cout << " --debug-rbd-mirror=<log-level>/<memory-level> set rbd-mirror debug level\n";
generic_server_usage();
}
static void handle_signal(int signum)
{
if (mirror)
mirror->handle_signal(signum);
}
int main(int argc, const char **argv)
{
auto args = argv_to_vec(argc, argv);
if (args.empty()) {
std::cerr << argv[0] << ": -h or --help for usage" << std::endl;
exit(1);
}
if (ceph_argparse_need_usage(args)) {
usage();
exit(0);
}
auto cct = global_init(nullptr, args, CEPH_ENTITY_TYPE_CLIENT,
CODE_ENVIRONMENT_DAEMON,
CINIT_FLAG_UNPRIVILEGED_DAEMON_DEFAULTS);
if (g_conf()->daemonize) {
global_init_daemonize(g_ceph_context);
}
common_init_finish(g_ceph_context);
init_async_signal_handler();
register_async_signal_handler(SIGHUP, handle_signal);
register_async_signal_handler_oneshot(SIGINT, handle_signal);
register_async_signal_handler_oneshot(SIGTERM, handle_signal);
auto cmd_args = argv_to_vec(argc, argv);
// disable unnecessary librbd cache
g_ceph_context->_conf.set_val_or_die("rbd_cache", "false");
auto prio =
g_ceph_context->_conf.get_val<int64_t>("rbd_mirror_perf_stats_prio");
{
PerfCountersBuilder plb(g_ceph_context, "rbd_mirror_journal",
rbd::mirror::l_rbd_mirror_journal_first,
rbd::mirror::l_rbd_mirror_journal_last);
plb.add_u64_counter(rbd::mirror::l_rbd_mirror_journal_entries, "entries",
"Number of entries replayed", nullptr, prio);
plb.add_u64_counter(rbd::mirror::l_rbd_mirror_journal_replay_bytes,
"replay_bytes", "Total bytes replayed", nullptr, prio,
unit_t(UNIT_BYTES));
plb.add_time_avg(rbd::mirror::l_rbd_mirror_journal_replay_latency,
"replay_latency", "Replay latency", nullptr, prio);
g_journal_perf_counters = plb.create_perf_counters();
}
{
PerfCountersBuilder plb(
g_ceph_context, "rbd_mirror_snapshot",
rbd::mirror::l_rbd_mirror_snapshot_first,
rbd::mirror::l_rbd_mirror_snapshot_remote_timestamp);
plb.add_u64_counter(rbd::mirror::l_rbd_mirror_snapshot_snapshots,
"snapshots", "Number of snapshots synced", nullptr,
prio);
plb.add_time_avg(rbd::mirror::l_rbd_mirror_snapshot_sync_time, "sync_time",
"Average sync time", nullptr, prio);
plb.add_u64_counter(rbd::mirror::l_rbd_mirror_snapshot_sync_bytes,
"sync_bytes", "Total bytes synced", nullptr, prio,
unit_t(UNIT_BYTES));
g_snapshot_perf_counters = plb.create_perf_counters();
}
g_ceph_context->get_perfcounters_collection()->add(g_journal_perf_counters);
g_ceph_context->get_perfcounters_collection()->add(g_snapshot_perf_counters);
mirror = new rbd::mirror::Mirror(g_ceph_context, cmd_args);
int r = mirror->init();
if (r < 0) {
std::cerr << "failed to initialize: " << cpp_strerror(r) << std::endl;
goto cleanup;
}
mirror->run();
cleanup:
unregister_async_signal_handler(SIGHUP, handle_signal);
unregister_async_signal_handler(SIGINT, handle_signal);
unregister_async_signal_handler(SIGTERM, handle_signal);
shutdown_async_signal_handler();
g_ceph_context->get_perfcounters_collection()->remove(g_journal_perf_counters);
g_ceph_context->get_perfcounters_collection()->remove(g_snapshot_perf_counters);
delete mirror;
delete g_journal_perf_counters;
delete g_snapshot_perf_counters;
return r < 0 ? EXIT_FAILURE : EXIT_SUCCESS;
}
| 4,454 | 34.64 | 95 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_deleter/SnapshotPurgeRequest.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "tools/rbd_mirror/image_deleter/SnapshotPurgeRequest.h"
#include "common/debug.h"
#include "common/errno.h"
#include "librbd/ExclusiveLock.h"
#include "librbd/ImageCtx.h"
#include "librbd/ImageState.h"
#include "librbd/Operations.h"
#include "librbd/Utils.h"
#include "librbd/journal/Policy.h"
#include "tools/rbd_mirror/image_deleter/Types.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_deleter::SnapshotPurgeRequest: " \
<< this << " " << __func__ << ": "
namespace rbd {
namespace mirror {
namespace image_deleter {
using librbd::util::create_context_callback;
template <typename I>
void SnapshotPurgeRequest<I>::send() {
open_image();
}
template <typename I>
void SnapshotPurgeRequest<I>::open_image() {
dout(10) << dendl;
m_image_ctx = I::create("", m_image_id, nullptr, m_io_ctx, false);
// ensure non-primary images can be modified
m_image_ctx->read_only_mask &= ~librbd::IMAGE_READ_ONLY_FLAG_NON_PRIMARY;
{
std::unique_lock image_locker{m_image_ctx->image_lock};
m_image_ctx->set_journal_policy(new JournalPolicy());
}
Context *ctx = create_context_callback<
SnapshotPurgeRequest<I>, &SnapshotPurgeRequest<I>::handle_open_image>(
this);
m_image_ctx->state->open(librbd::OPEN_FLAG_SKIP_OPEN_PARENT, ctx);
}
template <typename I>
void SnapshotPurgeRequest<I>::handle_open_image(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to open image '" << m_image_id << "': " << cpp_strerror(r)
<< dendl;
m_image_ctx = nullptr;
finish(r);
return;
}
acquire_lock();
}
template <typename I>
void SnapshotPurgeRequest<I>::acquire_lock() {
dout(10) << dendl;
m_image_ctx->owner_lock.lock_shared();
if (m_image_ctx->exclusive_lock == nullptr) {
m_image_ctx->owner_lock.unlock_shared();
start_snap_unprotect();
return;
}
m_image_ctx->exclusive_lock->acquire_lock(create_context_callback<
SnapshotPurgeRequest<I>, &SnapshotPurgeRequest<I>::handle_acquire_lock>(
this));
m_image_ctx->owner_lock.unlock_shared();
}
template <typename I>
void SnapshotPurgeRequest<I>::handle_acquire_lock(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to acquire exclusive lock: " << cpp_strerror(r) << dendl;
m_ret_val = r;
close_image();
return;
}
start_snap_unprotect();
}
template <typename I>
void SnapshotPurgeRequest<I>::start_snap_unprotect() {
dout(10) << dendl;
{
std::shared_lock image_locker{m_image_ctx->image_lock};
m_snaps = m_image_ctx->snaps;
}
snap_unprotect();
}
template <typename I>
void SnapshotPurgeRequest<I>::snap_unprotect() {
if (m_snaps.empty()) {
close_image();
return;
}
librados::snap_t snap_id = m_snaps.back();
m_image_ctx->image_lock.lock_shared();
int r = m_image_ctx->get_snap_namespace(snap_id, &m_snap_namespace);
if (r < 0) {
m_image_ctx->image_lock.unlock_shared();
derr << "failed to get snap namespace: " << cpp_strerror(r) << dendl;
m_ret_val = r;
close_image();
return;
}
r = m_image_ctx->get_snap_name(snap_id, &m_snap_name);
if (r < 0) {
m_image_ctx->image_lock.unlock_shared();
derr << "failed to get snap name: " << cpp_strerror(r) << dendl;
m_ret_val = r;
close_image();
return;
}
bool is_protected;
r = m_image_ctx->is_snap_protected(snap_id, &is_protected);
if (r < 0) {
m_image_ctx->image_lock.unlock_shared();
derr << "failed to get snap protection status: " << cpp_strerror(r)
<< dendl;
m_ret_val = r;
close_image();
return;
}
m_image_ctx->image_lock.unlock_shared();
if (!is_protected) {
snap_remove();
return;
}
dout(10) << "snap_id=" << snap_id << ", "
<< "snap_namespace=" << m_snap_namespace << ", "
<< "snap_name=" << m_snap_name << dendl;
auto finish_op_ctx = start_lock_op(&r);
if (finish_op_ctx == nullptr) {
derr << "lost exclusive lock" << dendl;
m_ret_val = r;
close_image();
return;
}
auto ctx = new LambdaContext([this, finish_op_ctx](int r) {
handle_snap_unprotect(r);
finish_op_ctx->complete(0);
});
std::shared_lock owner_locker{m_image_ctx->owner_lock};
m_image_ctx->operations->execute_snap_unprotect(
m_snap_namespace, m_snap_name.c_str(), ctx);
}
template <typename I>
void SnapshotPurgeRequest<I>::handle_snap_unprotect(int r) {
dout(10) << "r=" << r << dendl;
if (r == -EBUSY) {
dout(10) << "snapshot in-use" << dendl;
m_ret_val = r;
close_image();
return;
} else if (r < 0) {
derr << "failed to unprotect snapshot: " << cpp_strerror(r) << dendl;
m_ret_val = r;
close_image();
return;
}
{
// avoid the need to refresh to delete the newly unprotected snapshot
std::shared_lock image_locker{m_image_ctx->image_lock};
librados::snap_t snap_id = m_snaps.back();
auto snap_info_it = m_image_ctx->snap_info.find(snap_id);
if (snap_info_it != m_image_ctx->snap_info.end()) {
snap_info_it->second.protection_status =
RBD_PROTECTION_STATUS_UNPROTECTED;
}
}
snap_remove();
}
template <typename I>
void SnapshotPurgeRequest<I>::snap_remove() {
librados::snap_t snap_id = m_snaps.back();
dout(10) << "snap_id=" << snap_id << ", "
<< "snap_namespace=" << m_snap_namespace << ", "
<< "snap_name=" << m_snap_name << dendl;
int r;
auto finish_op_ctx = start_lock_op(&r);
if (finish_op_ctx == nullptr) {
derr << "lost exclusive lock" << dendl;
m_ret_val = r;
close_image();
return;
}
auto ctx = new LambdaContext([this, finish_op_ctx](int r) {
handle_snap_remove(r);
finish_op_ctx->complete(0);
});
std::shared_lock owner_locker{m_image_ctx->owner_lock};
m_image_ctx->operations->execute_snap_remove(
m_snap_namespace, m_snap_name.c_str(), ctx);
}
template <typename I>
void SnapshotPurgeRequest<I>::handle_snap_remove(int r) {
dout(10) << "r=" << r << dendl;
if (r == -EBUSY) {
dout(10) << "snapshot in-use" << dendl;
m_ret_val = r;
close_image();
return;
} else if (r < 0) {
derr << "failed to remove snapshot: " << cpp_strerror(r) << dendl;
m_ret_val = r;
close_image();
return;
}
m_snaps.pop_back();
snap_unprotect();
}
template <typename I>
void SnapshotPurgeRequest<I>::close_image() {
dout(10) << dendl;
m_image_ctx->state->close(create_context_callback<
SnapshotPurgeRequest<I>,
&SnapshotPurgeRequest<I>::handle_close_image>(this));
}
template <typename I>
void SnapshotPurgeRequest<I>::handle_close_image(int r) {
dout(10) << "r=" << r << dendl;
m_image_ctx = nullptr;
if (r < 0) {
derr << "failed to close: " << cpp_strerror(r) << dendl;
finish(r);
return;
}
finish(0);
}
template <typename I>
void SnapshotPurgeRequest<I>::finish(int r) {
if (m_ret_val < 0) {
r = m_ret_val;
}
m_on_finish->complete(r);
delete this;
}
template <typename I>
Context *SnapshotPurgeRequest<I>::start_lock_op(int* r) {
std::shared_lock owner_locker{m_image_ctx->owner_lock};
if (m_image_ctx->exclusive_lock == nullptr) {
return new LambdaContext([](int r) {});
}
return m_image_ctx->exclusive_lock->start_op(r);
}
} // namespace image_deleter
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_deleter::SnapshotPurgeRequest<librbd::ImageCtx>;
| 7,618 | 24.396667 | 84 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_deleter/SnapshotPurgeRequest.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef CEPH_RBD_MIRROR_IMAGE_DELETER_SNAPSHOT_PURGE_REQUEST_H
#define CEPH_RBD_MIRROR_IMAGE_DELETER_SNAPSHOT_PURGE_REQUEST_H
#include "include/rados/librados.hpp"
#include "cls/rbd/cls_rbd_types.h"
#include <string>
#include <vector>
class Context;
namespace librbd { struct ImageCtx; }
namespace rbd {
namespace mirror {
namespace image_deleter {
template <typename ImageCtxT = librbd::ImageCtx>
class SnapshotPurgeRequest {
public:
static SnapshotPurgeRequest* create(librados::IoCtx &io_ctx,
const std::string &image_id,
Context *on_finish) {
return new SnapshotPurgeRequest(io_ctx, image_id, on_finish);
}
SnapshotPurgeRequest(librados::IoCtx &io_ctx, const std::string &image_id,
Context *on_finish)
: m_io_ctx(io_ctx), m_image_id(image_id), m_on_finish(on_finish) {
}
void send();
private:
/*
* @verbatim
*
* <start>
* |
* v
* OPEN_IMAGE
* |
* v
* ACQUIRE_LOCK
* |
* | (repeat for each snapshot)
* |/------------------------\
* | |
* v (skip if not needed) |
* SNAP_UNPROTECT |
* | |
* v (skip if not needed) |
* SNAP_REMOVE -----------------/
* |
* v
* CLOSE_IMAGE
* |
* v
* <finish>
*
* @endverbatim
*/
librados::IoCtx &m_io_ctx;
std::string m_image_id;
Context *m_on_finish;
ImageCtxT *m_image_ctx = nullptr;
int m_ret_val = 0;
std::vector<librados::snap_t> m_snaps;
cls::rbd::SnapshotNamespace m_snap_namespace;
std::string m_snap_name;
void open_image();
void handle_open_image(int r);
void acquire_lock();
void handle_acquire_lock(int r);
void start_snap_unprotect();
void snap_unprotect();
void handle_snap_unprotect(int r);
void snap_remove();
void handle_snap_remove(int r);
void close_image();
void handle_close_image(int r);
void finish(int r);
Context *start_lock_op(int* r);
};
} // namespace image_deleter
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_deleter::SnapshotPurgeRequest<librbd::ImageCtx>;
#endif // CEPH_RBD_MIRROR_IMAGE_DELETER_SNAPSHOT_PURGE_REQUEST_H
| 2,399 | 21.641509 | 89 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_deleter/TrashMoveRequest.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "tools/rbd_mirror/image_deleter/TrashMoveRequest.h"
#include "include/rbd_types.h"
#include "cls/rbd/cls_rbd_client.h"
#include "common/debug.h"
#include "common/errno.h"
#include "common/WorkQueue.h"
#include "librbd/ExclusiveLock.h"
#include "librbd/ImageCtx.h"
#include "librbd/ImageState.h"
#include "librbd/Journal.h"
#include "librbd/TrashWatcher.h"
#include "librbd/Utils.h"
#include "librbd/asio/ContextWQ.h"
#include "librbd/journal/ResetRequest.h"
#include "librbd/mirror/ImageRemoveRequest.h"
#include "librbd/mirror/GetInfoRequest.h"
#include "librbd/trash/MoveRequest.h"
#include "tools/rbd_mirror/image_deleter/Types.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_deleter::TrashMoveRequest: " \
<< this << " " << __func__ << ": "
namespace rbd {
namespace mirror {
namespace image_deleter {
using librbd::util::create_context_callback;
using librbd::util::create_rados_callback;
template <typename I>
void TrashMoveRequest<I>::send() {
get_mirror_image_id();
}
template <typename I>
void TrashMoveRequest<I>::get_mirror_image_id() {
dout(10) << dendl;
librados::ObjectReadOperation op;
librbd::cls_client::mirror_image_get_image_id_start(&op, m_global_image_id);
auto aio_comp = create_rados_callback<
TrashMoveRequest<I>,
&TrashMoveRequest<I>::handle_get_mirror_image_id>(this);
m_out_bl.clear();
int r = m_io_ctx.aio_operate(RBD_MIRRORING, aio_comp, &op, &m_out_bl);
ceph_assert(r == 0);
aio_comp->release();
}
template <typename I>
void TrashMoveRequest<I>::handle_get_mirror_image_id(int r) {
dout(10) << "r=" << r << dendl;
if (r == 0) {
auto bl_it = m_out_bl.cbegin();
r = librbd::cls_client::mirror_image_get_image_id_finish(&bl_it,
&m_image_id);
}
if (r == -ENOENT) {
dout(10) << "image " << m_global_image_id << " is not mirrored" << dendl;
finish(r);
return;
} else if (r < 0) {
derr << "error retrieving local id for image " << m_global_image_id << ": "
<< cpp_strerror(r) << dendl;
finish(r);
return;
}
get_mirror_info();
}
template <typename I>
void TrashMoveRequest<I>::get_mirror_info() {
dout(10) << dendl;
auto ctx = create_context_callback<
TrashMoveRequest<I>, &TrashMoveRequest<I>::handle_get_mirror_info>(this);
auto req = librbd::mirror::GetInfoRequest<I>::create(
m_io_ctx, m_op_work_queue, m_image_id, &m_mirror_image, &m_promotion_state,
&m_primary_mirror_uuid, ctx);
req->send();
}
template <typename I>
void TrashMoveRequest<I>::handle_get_mirror_info(int r) {
dout(10) << "r=" << r << dendl;
if (r == -ENOENT) {
dout(5) << "image " << m_global_image_id << " is not mirrored" << dendl;
finish(r);
return;
} else if (r < 0) {
derr << "error retrieving image primary info for image "
<< m_global_image_id << ": " << cpp_strerror(r) << dendl;
finish(r);
return;
}
if (m_promotion_state == librbd::mirror::PROMOTION_STATE_PRIMARY) {
dout(10) << "image " << m_global_image_id << " is local primary" << dendl;
finish(-EPERM);
return;
} else if (m_promotion_state == librbd::mirror::PROMOTION_STATE_ORPHAN &&
!m_resync) {
dout(10) << "image " << m_global_image_id << " is orphaned" << dendl;
finish(-EPERM);
return;
}
disable_mirror_image();
}
template <typename I>
void TrashMoveRequest<I>::disable_mirror_image() {
dout(10) << dendl;
m_mirror_image.state = cls::rbd::MIRROR_IMAGE_STATE_DISABLING;
librados::ObjectWriteOperation op;
librbd::cls_client::mirror_image_set(&op, m_image_id, m_mirror_image);
auto aio_comp = create_rados_callback<
TrashMoveRequest<I>,
&TrashMoveRequest<I>::handle_disable_mirror_image>(this);
int r = m_io_ctx.aio_operate(RBD_MIRRORING, aio_comp, &op);
ceph_assert(r == 0);
aio_comp->release();
}
template <typename I>
void TrashMoveRequest<I>::handle_disable_mirror_image(int r) {
dout(10) << "r=" << r << dendl;
if (r == -ENOENT) {
dout(10) << "local image is not mirrored, aborting deletion." << dendl;
finish(r);
return;
} else if (r == -EEXIST || r == -EINVAL) {
derr << "cannot disable mirroring for image " << m_global_image_id
<< ": global_image_id has changed/reused: "
<< cpp_strerror(r) << dendl;
finish(r);
return;
} else if (r < 0) {
derr << "cannot disable mirroring for image " << m_global_image_id
<< ": " << cpp_strerror(r) << dendl;
finish(r);
return;
}
open_image();
}
template <typename I>
void TrashMoveRequest<I>::open_image() {
dout(10) << dendl;
m_image_ctx = I::create("", m_image_id, nullptr, m_io_ctx, false);
// ensure non-primary images can be modified
m_image_ctx->read_only_mask &= ~librbd::IMAGE_READ_ONLY_FLAG_NON_PRIMARY;
{
// don't attempt to open the journal
std::unique_lock image_locker{m_image_ctx->image_lock};
m_image_ctx->set_journal_policy(new JournalPolicy());
}
Context *ctx = create_context_callback<
TrashMoveRequest<I>, &TrashMoveRequest<I>::handle_open_image>(this);
m_image_ctx->state->open(librbd::OPEN_FLAG_SKIP_OPEN_PARENT, ctx);
}
template <typename I>
void TrashMoveRequest<I>::handle_open_image(int r) {
dout(10) << "r=" << r << dendl;
if (r == -ENOENT) {
dout(5) << "mirror image does not exist, removing orphaned metadata" << dendl;
m_image_ctx = nullptr;
remove_mirror_image();
return;
}
if (r < 0) {
derr << "failed to open image: " << cpp_strerror(r) << dendl;
m_image_ctx = nullptr;
finish(r);
return;
}
if (m_image_ctx->old_format) {
derr << "cannot move v1 image to trash" << dendl;
m_ret_val = -EINVAL;
close_image();
return;
}
reset_journal();
}
template <typename I>
void TrashMoveRequest<I>::reset_journal() {
if (m_mirror_image.mode == cls::rbd::MIRROR_IMAGE_MODE_SNAPSHOT) {
// snapshot-based mirroring doesn't require journal feature
acquire_lock();
return;
}
dout(10) << dendl;
// TODO use Journal thread pool for journal ops until converted to ASIO
ContextWQ* context_wq;
librbd::Journal<>::get_work_queue(
reinterpret_cast<CephContext*>(m_io_ctx.cct()), &context_wq);
// ensure that if the image is recovered any peers will split-brain
auto ctx = create_context_callback<
TrashMoveRequest<I>, &TrashMoveRequest<I>::handle_reset_journal>(this);
auto req = librbd::journal::ResetRequest<I>::create(
m_io_ctx, m_image_id, librbd::Journal<>::IMAGE_CLIENT_ID,
librbd::Journal<>::LOCAL_MIRROR_UUID, context_wq, ctx);
req->send();
}
template <typename I>
void TrashMoveRequest<I>::handle_reset_journal(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0 && r != -ENOENT) {
derr << "failed to reset journal: " << cpp_strerror(r) << dendl;
m_ret_val = r;
close_image();
return;
}
acquire_lock();
}
template <typename I>
void TrashMoveRequest<I>::acquire_lock() {
m_image_ctx->owner_lock.lock_shared();
if (m_image_ctx->exclusive_lock == nullptr) {
m_image_ctx->owner_lock.unlock_shared();
if (m_mirror_image.mode == cls::rbd::MIRROR_IMAGE_MODE_SNAPSHOT) {
// snapshot-based mirroring doesn't require exclusive-lock
trash_move();
} else {
derr << "exclusive lock feature not enabled" << dendl;
m_ret_val = -EINVAL;
close_image();
}
return;
}
dout(10) << dendl;
Context *ctx = create_context_callback<
TrashMoveRequest<I>, &TrashMoveRequest<I>::handle_acquire_lock>(this);
m_image_ctx->exclusive_lock->block_requests(0);
m_image_ctx->exclusive_lock->acquire_lock(ctx);
m_image_ctx->owner_lock.unlock_shared();
}
template <typename I>
void TrashMoveRequest<I>::handle_acquire_lock(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to acquire exclusive lock: " << cpp_strerror(r) << dendl;
m_ret_val = r;
close_image();
return;
}
trash_move();
}
template <typename I>
void TrashMoveRequest<I>::trash_move() {
dout(10) << dendl;
utime_t delete_time{ceph_clock_now()};
utime_t deferment_end_time{delete_time};
deferment_end_time +=
m_image_ctx->config.template get_val<uint64_t>("rbd_mirroring_delete_delay");
m_trash_image_spec = {
cls::rbd::TRASH_IMAGE_SOURCE_MIRRORING, m_image_ctx->name, delete_time,
deferment_end_time};
Context *ctx = create_context_callback<
TrashMoveRequest<I>, &TrashMoveRequest<I>::handle_trash_move>(this);
auto req = librbd::trash::MoveRequest<I>::create(
m_io_ctx, m_image_id, m_trash_image_spec, ctx);
req->send();
}
template <typename I>
void TrashMoveRequest<I>::handle_trash_move(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to move image to trash: " << cpp_strerror(r) << dendl;
m_ret_val = r;
close_image();
return;
}
m_moved_to_trash = true;
remove_mirror_image();
}
template <typename I>
void TrashMoveRequest<I>::remove_mirror_image() {
dout(10) << dendl;
auto ctx = create_context_callback<
TrashMoveRequest<I>,
&TrashMoveRequest<I>::handle_remove_mirror_image>(this);
auto req = librbd::mirror::ImageRemoveRequest<I>::create(
m_io_ctx, m_global_image_id, m_image_id, ctx);
req->send();
}
template <typename I>
void TrashMoveRequest<I>::handle_remove_mirror_image(int r) {
dout(10) << "r=" << r << dendl;
if (r == -ENOENT) {
dout(10) << "local image is not mirrored" << dendl;
} else if (r < 0) {
derr << "failed to remove mirror image state for " << m_global_image_id
<< ": " << cpp_strerror(r) << dendl;
m_ret_val = r;
}
close_image();
}
template <typename I>
void TrashMoveRequest<I>::close_image() {
dout(10) << dendl;
if (m_image_ctx == nullptr) {
handle_close_image(0);
return;
}
Context *ctx = create_context_callback<
TrashMoveRequest<I>, &TrashMoveRequest<I>::handle_close_image>(this);
m_image_ctx->state->close(ctx);
}
template <typename I>
void TrashMoveRequest<I>::handle_close_image(int r) {
dout(10) << "r=" << r << dendl;
m_image_ctx = nullptr;
if (r < 0) {
derr << "failed to close image: " << cpp_strerror(r) << dendl;
}
// don't send notification if we failed
if (!m_moved_to_trash) {
finish(0);
return;
}
notify_trash_add();
}
template <typename I>
void TrashMoveRequest<I>::notify_trash_add() {
dout(10) << dendl;
Context *ctx = create_context_callback<
TrashMoveRequest<I>, &TrashMoveRequest<I>::handle_notify_trash_add>(this);
librbd::TrashWatcher<I>::notify_image_added(m_io_ctx, m_image_id,
m_trash_image_spec, ctx);
}
template <typename I>
void TrashMoveRequest<I>::handle_notify_trash_add(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to notify trash watchers: " << cpp_strerror(r) << dendl;
}
finish(0);
}
template <typename I>
void TrashMoveRequest<I>::finish(int r) {
if (m_ret_val < 0) {
r = m_ret_val;
}
dout(10) << "r=" << r << dendl;
m_on_finish->complete(r);
delete this;
}
} // namespace image_deleter
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_deleter::TrashMoveRequest<librbd::ImageCtx>;
| 11,458 | 26.283333 | 82 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_deleter/TrashMoveRequest.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef CEPH_RBD_MIRROR_IMAGE_DELETE_TRASH_MOVE_REQUEST_H
#define CEPH_RBD_MIRROR_IMAGE_DELETE_TRASH_MOVE_REQUEST_H
#include "include/buffer.h"
#include "include/rados/librados.hpp"
#include "cls/rbd/cls_rbd_types.h"
#include "librbd/mirror/Types.h"
#include <string>
struct Context;
namespace librbd {
struct ImageCtx;
namespace asio { struct ContextWQ; }
} // namespace librbd
namespace rbd {
namespace mirror {
namespace image_deleter {
template <typename ImageCtxT = librbd::ImageCtx>
class TrashMoveRequest {
public:
static TrashMoveRequest* create(librados::IoCtx& io_ctx,
const std::string& global_image_id,
bool resync,
librbd::asio::ContextWQ* op_work_queue,
Context* on_finish) {
return new TrashMoveRequest(io_ctx, global_image_id, resync, op_work_queue,
on_finish);
}
TrashMoveRequest(librados::IoCtx& io_ctx, const std::string& global_image_id,
bool resync, librbd::asio::ContextWQ* op_work_queue,
Context* on_finish)
: m_io_ctx(io_ctx), m_global_image_id(global_image_id), m_resync(resync),
m_op_work_queue(op_work_queue), m_on_finish(on_finish) {
}
void send();
private:
/*
* @verbatim
*
* <start>
* |
* v
* GET_MIRROR_IMAGE_ID
* |
* v
* GET_MIRROR_INFO
* |
* v
* DISABLE_MIRROR_IMAGE
* |
* v
* OPEN_IMAGE
* |
* v (skip if not needed)
* RESET_JOURNAL
* |
* v (skip if not needed)
* ACQUIRE_LOCK
* |
* v
* TRASH_MOVE
* |
* v
* REMOVE_MIRROR_IMAGE
* |
* v
* CLOSE_IMAGE
* |
* v
* NOTIFY_TRASH_ADD
* |
* v
* <finish>
*
* @endverbatim
*/
librados::IoCtx &m_io_ctx;
std::string m_global_image_id;
bool m_resync;
librbd::asio::ContextWQ *m_op_work_queue;
Context *m_on_finish;
ceph::bufferlist m_out_bl;
std::string m_image_id;
cls::rbd::MirrorImage m_mirror_image;
librbd::mirror::PromotionState m_promotion_state;
std::string m_primary_mirror_uuid;
cls::rbd::TrashImageSpec m_trash_image_spec;
ImageCtxT *m_image_ctx = nullptr;;
int m_ret_val = 0;
bool m_moved_to_trash = false;
void get_mirror_image_id();
void handle_get_mirror_image_id(int r);
void get_mirror_info();
void handle_get_mirror_info(int r);
void disable_mirror_image();
void handle_disable_mirror_image(int r);
void open_image();
void handle_open_image(int r);
void reset_journal();
void handle_reset_journal(int r);
void acquire_lock();
void handle_acquire_lock(int r);
void trash_move();
void handle_trash_move(int r);
void remove_mirror_image();
void handle_remove_mirror_image(int r);
void close_image();
void handle_close_image(int r);
void notify_trash_add();
void handle_notify_trash_add(int r);
void finish(int r);
};
} // namespace image_deleter
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_deleter::TrashMoveRequest<librbd::ImageCtx>;
#endif // CEPH_RBD_MIRROR_IMAGE_DELETE_TRASH_WATCHER_H
| 3,322 | 22.237762 | 85 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_deleter/TrashRemoveRequest.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "tools/rbd_mirror/image_deleter/TrashRemoveRequest.h"
#include "include/ceph_assert.h"
#include "common/debug.h"
#include "common/errno.h"
#include "common/WorkQueue.h"
#include "cls/rbd/cls_rbd_client.h"
#include "librbd/ImageCtx.h"
#include "librbd/TrashWatcher.h"
#include "librbd/Utils.h"
#include "librbd/asio/ContextWQ.h"
#include "librbd/trash/RemoveRequest.h"
#include "tools/rbd_mirror/image_deleter/SnapshotPurgeRequest.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_deleter::TrashRemoveRequest: " \
<< this << " " << __func__ << ": "
namespace rbd {
namespace mirror {
namespace image_deleter {
using librbd::util::create_context_callback;
using librbd::util::create_rados_callback;
template <typename I>
void TrashRemoveRequest<I>::send() {
*m_error_result = ERROR_RESULT_RETRY;
get_trash_image_spec();
}
template <typename I>
void TrashRemoveRequest<I>::get_trash_image_spec() {
dout(10) << dendl;
librados::ObjectReadOperation op;
librbd::cls_client::trash_get_start(&op, m_image_id);
auto aio_comp = create_rados_callback<
TrashRemoveRequest<I>,
&TrashRemoveRequest<I>::handle_get_trash_image_spec>(this);
m_out_bl.clear();
int r = m_io_ctx.aio_operate(RBD_TRASH, aio_comp, &op, &m_out_bl);
ceph_assert(r == 0);
aio_comp->release();
}
template <typename I>
void TrashRemoveRequest<I>::handle_get_trash_image_spec(int r) {
dout(10) << "r=" << r << dendl;
if (r == 0) {
auto bl_it = m_out_bl.cbegin();
r = librbd::cls_client::trash_get_finish(&bl_it, &m_trash_image_spec);
}
if (r == -ENOENT || (r >= 0 && m_trash_image_spec.source !=
cls::rbd::TRASH_IMAGE_SOURCE_MIRRORING)) {
dout(10) << "image id " << m_image_id << " not in mirroring trash" << dendl;
finish(0);
return;
} else if (r < 0) {
derr << "error getting image id " << m_image_id << " info from trash: "
<< cpp_strerror(r) << dendl;
finish(r);
return;
}
if (m_trash_image_spec.state != cls::rbd::TRASH_IMAGE_STATE_NORMAL &&
m_trash_image_spec.state != cls::rbd::TRASH_IMAGE_STATE_REMOVING) {
dout(10) << "image " << m_image_id << " is not in an expected trash state: "
<< m_trash_image_spec.state << dendl;
*m_error_result = ERROR_RESULT_RETRY_IMMEDIATELY;
finish(-EBUSY);
return;
}
set_trash_state();
}
template <typename I>
void TrashRemoveRequest<I>::set_trash_state() {
if (m_trash_image_spec.state == cls::rbd::TRASH_IMAGE_STATE_REMOVING) {
get_snap_context();
return;
}
dout(10) << dendl;
librados::ObjectWriteOperation op;
librbd::cls_client::trash_state_set(&op, m_image_id,
cls::rbd::TRASH_IMAGE_STATE_REMOVING,
cls::rbd::TRASH_IMAGE_STATE_NORMAL);
auto aio_comp = create_rados_callback<
TrashRemoveRequest<I>,
&TrashRemoveRequest<I>::handle_set_trash_state>(this);
int r = m_io_ctx.aio_operate(RBD_TRASH, aio_comp, &op);
ceph_assert(r == 0);
aio_comp->release();
}
template <typename I>
void TrashRemoveRequest<I>::handle_set_trash_state(int r) {
dout(10) << "r=" << r << dendl;
if (r == -ENOENT) {
dout(10) << "image id " << m_image_id << " not in mirroring trash" << dendl;
finish(0);
return;
} else if (r < 0 && r != -EOPNOTSUPP) {
derr << "error setting trash image state for image id " << m_image_id
<< ": " << cpp_strerror(r) << dendl;
finish(r);
return;
}
get_snap_context();
}
template <typename I>
void TrashRemoveRequest<I>::get_snap_context() {
dout(10) << dendl;
librados::ObjectReadOperation op;
librbd::cls_client::get_snapcontext_start(&op);
std::string header_oid = librbd::util::header_name(m_image_id);
auto aio_comp = create_rados_callback<
TrashRemoveRequest<I>,
&TrashRemoveRequest<I>::handle_get_snap_context>(this);
m_out_bl.clear();
int r = m_io_ctx.aio_operate(header_oid, aio_comp, &op, &m_out_bl);
ceph_assert(r == 0);
aio_comp->release();
}
template <typename I>
void TrashRemoveRequest<I>::handle_get_snap_context(int r) {
dout(10) << "r=" << r << dendl;
::SnapContext snapc;
if (r == 0) {
auto bl_it = m_out_bl.cbegin();
r = librbd::cls_client::get_snapcontext_finish(&bl_it, &snapc);
}
if (r < 0 && r != -ENOENT) {
derr << "error retrieving snapshot context for image "
<< m_image_id << ": " << cpp_strerror(r) << dendl;
finish(r);
return;
}
m_has_snapshots = (!snapc.empty());
purge_snapshots();
}
template <typename I>
void TrashRemoveRequest<I>::purge_snapshots() {
if (!m_has_snapshots) {
remove_image();
return;
}
dout(10) << dendl;
auto ctx = create_context_callback<
TrashRemoveRequest<I>,
&TrashRemoveRequest<I>::handle_purge_snapshots>(this);
auto req = SnapshotPurgeRequest<I>::create(m_io_ctx, m_image_id, ctx);
req->send();
}
template <typename I>
void TrashRemoveRequest<I>::handle_purge_snapshots(int r) {
dout(10) << "r=" << r << dendl;
if (r == -EBUSY) {
dout(10) << "snapshots still in-use" << dendl;
*m_error_result = ERROR_RESULT_RETRY_IMMEDIATELY;
finish(r);
return;
} else if (r < 0) {
derr << "failed to purge image snapshots: " << cpp_strerror(r) << dendl;
finish(r);
return;
}
remove_image();
}
template <typename I>
void TrashRemoveRequest<I>::remove_image() {
dout(10) << dendl;
auto ctx = create_context_callback<
TrashRemoveRequest<I>,
&TrashRemoveRequest<I>::handle_remove_image>(this);
auto req = librbd::trash::RemoveRequest<I>::create(
m_io_ctx, m_image_id, m_op_work_queue, true, m_progress_ctx,
ctx);
req->send();
}
template <typename I>
void TrashRemoveRequest<I>::handle_remove_image(int r) {
dout(10) << "r=" << r << dendl;
if (r == -ENOTEMPTY) {
// image must have clone v2 snapshot still associated to child
dout(10) << "snapshots still in-use" << dendl;
*m_error_result = ERROR_RESULT_RETRY_IMMEDIATELY;
finish(-EBUSY);
return;
}
if (r < 0 && r != -ENOENT) {
derr << "error removing image " << m_image_id << " "
<< "(" << m_image_id << ") from local pool: "
<< cpp_strerror(r) << dendl;
finish(r);
return;
}
notify_trash_removed();
}
template <typename I>
void TrashRemoveRequest<I>::notify_trash_removed() {
dout(10) << dendl;
Context *ctx = create_context_callback<
TrashRemoveRequest<I>,
&TrashRemoveRequest<I>::handle_notify_trash_removed>(this);
librbd::TrashWatcher<I>::notify_image_removed(m_io_ctx, m_image_id, ctx);
}
template <typename I>
void TrashRemoveRequest<I>::handle_notify_trash_removed(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to notify trash watchers: " << cpp_strerror(r) << dendl;
}
finish(0);
}
template <typename I>
void TrashRemoveRequest<I>::finish(int r) {
dout(10) << "r=" << r << dendl;
m_on_finish->complete(r);
delete this;
}
} // namespace image_deleter
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_deleter::TrashRemoveRequest<librbd::ImageCtx>;
| 7,343 | 26.609023 | 82 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_deleter/TrashRemoveRequest.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef CEPH_RBD_MIRROR_IMAGE_DELETER_TRASH_REMOVE_REQUEST_H
#define CEPH_RBD_MIRROR_IMAGE_DELETER_TRASH_REMOVE_REQUEST_H
#include "include/rados/librados.hpp"
#include "include/buffer.h"
#include "cls/rbd/cls_rbd_types.h"
#include "librbd/internal.h"
#include "tools/rbd_mirror/image_deleter/Types.h"
#include <string>
#include <vector>
class Context;
class ContextWQ;
namespace librbd {
struct ImageCtx;
namespace asio { struct ContextWQ; }
} // namespace librbd
namespace rbd {
namespace mirror {
namespace image_deleter {
template <typename ImageCtxT = librbd::ImageCtx>
class TrashRemoveRequest {
public:
static TrashRemoveRequest* create(librados::IoCtx &io_ctx,
const std::string &image_id,
ErrorResult *error_result,
librbd::asio::ContextWQ *op_work_queue,
Context *on_finish) {
return new TrashRemoveRequest(io_ctx, image_id, error_result, op_work_queue,
on_finish);
}
TrashRemoveRequest(librados::IoCtx &io_ctx, const std::string &image_id,
ErrorResult *error_result,
librbd::asio::ContextWQ *op_work_queue,
Context *on_finish)
: m_io_ctx(io_ctx), m_image_id(image_id), m_error_result(error_result),
m_op_work_queue(op_work_queue), m_on_finish(on_finish) {
}
void send();
private:
/*
* @verbatim
*
* <start>
* |
* v
* GET_TRASH_IMAGE_SPEC
* |
* v
* SET_TRASH_STATE
* |
* v
* GET_SNAP_CONTEXT
* |
* v
* PURGE_SNAPSHOTS
* |
* v
* TRASH_REMOVE
* |
* v
* NOTIFY_TRASH_REMOVE
* |
* v
* <finish>
*
* @endverbatim
*/
librados::IoCtx &m_io_ctx;
std::string m_image_id;
ErrorResult *m_error_result;
librbd::asio::ContextWQ *m_op_work_queue;
Context *m_on_finish;
ceph::bufferlist m_out_bl;
cls::rbd::TrashImageSpec m_trash_image_spec;
bool m_has_snapshots = false;
librbd::NoOpProgressContext m_progress_ctx;
void get_trash_image_spec();
void handle_get_trash_image_spec(int r);
void set_trash_state();
void handle_set_trash_state(int r);
void get_snap_context();
void handle_get_snap_context(int r);
void purge_snapshots();
void handle_purge_snapshots(int r);
void remove_image();
void handle_remove_image(int r);
void notify_trash_removed();
void handle_notify_trash_removed(int r);
void finish(int r);
};
} // namespace image_deleter
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_deleter::TrashRemoveRequest<librbd::ImageCtx>;
#endif // CEPH_RBD_MIRROR_IMAGE_DELETER_TRASH_REMOVE_REQUEST_H
| 2,883 | 23.440678 | 87 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_deleter/TrashWatcher.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "tools/rbd_mirror/image_deleter/TrashWatcher.h"
#include "include/rbd_types.h"
#include "cls/rbd/cls_rbd_client.h"
#include "common/debug.h"
#include "common/errno.h"
#include "common/Timer.h"
#include "librbd/ImageCtx.h"
#include "librbd/Utils.h"
#include "librbd/asio/ContextWQ.h"
#include "tools/rbd_mirror/Threads.h"
#include "tools/rbd_mirror/image_deleter/Types.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_deleter::TrashWatcher: " \
<< this << " " << __func__ << ": "
using librbd::util::create_context_callback;
using librbd::util::create_rados_callback;
namespace rbd {
namespace mirror {
namespace image_deleter {
namespace {
const size_t MAX_RETURN = 1024;
} // anonymous namespace
template <typename I>
TrashWatcher<I>::TrashWatcher(librados::IoCtx &io_ctx, Threads<I> *threads,
TrashListener& trash_listener)
: librbd::TrashWatcher<I>(io_ctx, threads->work_queue),
m_io_ctx(io_ctx), m_threads(threads), m_trash_listener(trash_listener),
m_lock(ceph::make_mutex(librbd::util::unique_lock_name(
"rbd::mirror::image_deleter::TrashWatcher", this))) {
}
template <typename I>
void TrashWatcher<I>::init(Context *on_finish) {
dout(5) << dendl;
{
std::lock_guard locker{m_lock};
m_on_init_finish = on_finish;
ceph_assert(!m_trash_list_in_progress);
m_trash_list_in_progress = true;
}
create_trash();
}
template <typename I>
void TrashWatcher<I>::shut_down(Context *on_finish) {
dout(5) << dendl;
{
std::scoped_lock locker{m_threads->timer_lock, m_lock};
ceph_assert(!m_shutting_down);
m_shutting_down = true;
if (m_timer_ctx != nullptr) {
m_threads->timer->cancel_event(m_timer_ctx);
m_timer_ctx = nullptr;
}
}
auto ctx = new LambdaContext([this, on_finish](int r) {
unregister_watcher(on_finish);
});
m_async_op_tracker.wait_for_ops(ctx);
}
template <typename I>
void TrashWatcher<I>::handle_image_added(const std::string &image_id,
const cls::rbd::TrashImageSpec& spec) {
dout(10) << "image_id=" << image_id << dendl;
std::lock_guard locker{m_lock};
add_image(image_id, spec);
}
template <typename I>
void TrashWatcher<I>::handle_image_removed(const std::string &image_id) {
// ignore removals -- the image deleter will ignore -ENOENTs
}
template <typename I>
void TrashWatcher<I>::handle_rewatch_complete(int r) {
dout(5) << "r=" << r << dendl;
if (r == -EBLOCKLISTED) {
dout(0) << "detected client is blocklisted" << dendl;
return;
} else if (r == -ENOENT) {
dout(5) << "trash directory deleted" << dendl;
} else if (r < 0) {
derr << "unexpected error re-registering trash directory watch: "
<< cpp_strerror(r) << dendl;
}
schedule_trash_list(30);
}
template <typename I>
void TrashWatcher<I>::create_trash() {
dout(20) << dendl;
{
std::lock_guard locker{m_lock};
ceph_assert(m_trash_list_in_progress);
}
librados::ObjectWriteOperation op;
op.create(false);
m_async_op_tracker.start_op();
auto aio_comp = create_rados_callback<
TrashWatcher<I>, &TrashWatcher<I>::handle_create_trash>(this);
int r = m_io_ctx.aio_operate(RBD_TRASH, aio_comp, &op);
ceph_assert(r == 0);
aio_comp->release();
}
template <typename I>
void TrashWatcher<I>::handle_create_trash(int r) {
dout(20) << "r=" << r << dendl;
{
std::lock_guard locker{m_lock};
ceph_assert(m_trash_list_in_progress);
}
Context* on_init_finish = nullptr;
if (r == -EBLOCKLISTED || r == -ENOENT) {
if (r == -EBLOCKLISTED) {
dout(0) << "detected client is blocklisted" << dendl;
} else {
dout(0) << "detected pool no longer exists" << dendl;
}
std::lock_guard locker{m_lock};
std::swap(on_init_finish, m_on_init_finish);
m_trash_list_in_progress = false;
} else if (r < 0 && r != -EEXIST) {
derr << "failed to create trash object: " << cpp_strerror(r) << dendl;
{
std::lock_guard locker{m_lock};
m_trash_list_in_progress = false;
}
schedule_trash_list(30);
} else {
register_watcher();
}
m_async_op_tracker.finish_op();
if (on_init_finish != nullptr) {
on_init_finish->complete(r);
}
}
template <typename I>
void TrashWatcher<I>::register_watcher() {
{
std::lock_guard locker{m_lock};
ceph_assert(m_trash_list_in_progress);
}
// if the watch registration is in-flight, let the watcher
// handle the transition -- only (re-)register if it's not registered
if (!this->is_unregistered()) {
trash_list(true);
return;
}
// first time registering or the watch failed
dout(5) << dendl;
m_async_op_tracker.start_op();
Context *ctx = create_context_callback<
TrashWatcher, &TrashWatcher<I>::handle_register_watcher>(this);
this->register_watch(ctx);
}
template <typename I>
void TrashWatcher<I>::handle_register_watcher(int r) {
dout(5) << "r=" << r << dendl;
{
std::lock_guard locker{m_lock};
ceph_assert(m_trash_list_in_progress);
if (r < 0) {
m_trash_list_in_progress = false;
}
}
Context *on_init_finish = nullptr;
if (r >= 0) {
trash_list(true);
} else if (r == -EBLOCKLISTED) {
dout(0) << "detected client is blocklisted" << dendl;
std::lock_guard locker{m_lock};
std::swap(on_init_finish, m_on_init_finish);
} else {
derr << "unexpected error registering trash directory watch: "
<< cpp_strerror(r) << dendl;
schedule_trash_list(10);
}
m_async_op_tracker.finish_op();
if (on_init_finish != nullptr) {
on_init_finish->complete(r);
}
}
template <typename I>
void TrashWatcher<I>::unregister_watcher(Context* on_finish) {
dout(5) << dendl;
m_async_op_tracker.start_op();
Context *ctx = new LambdaContext([this, on_finish](int r) {
handle_unregister_watcher(r, on_finish);
});
this->unregister_watch(ctx);
}
template <typename I>
void TrashWatcher<I>::handle_unregister_watcher(int r, Context* on_finish) {
dout(5) << "unregister_watcher: r=" << r << dendl;
if (r < 0) {
derr << "error unregistering watcher for trash directory: "
<< cpp_strerror(r) << dendl;
}
m_async_op_tracker.finish_op();
on_finish->complete(0);
}
template <typename I>
void TrashWatcher<I>::trash_list(bool initial_request) {
if (initial_request) {
m_async_op_tracker.start_op();
m_last_image_id = "";
}
dout(5) << "last_image_id=" << m_last_image_id << dendl;
{
std::lock_guard locker{m_lock};
ceph_assert(m_trash_list_in_progress);
}
librados::ObjectReadOperation op;
librbd::cls_client::trash_list_start(&op, m_last_image_id, MAX_RETURN);
librados::AioCompletion *aio_comp = create_rados_callback<
TrashWatcher<I>, &TrashWatcher<I>::handle_trash_list>(this);
m_out_bl.clear();
int r = m_io_ctx.aio_operate(RBD_TRASH, aio_comp, &op, &m_out_bl);
ceph_assert(r == 0);
aio_comp->release();
}
template <typename I>
void TrashWatcher<I>::handle_trash_list(int r) {
dout(5) << "r=" << r << dendl;
std::map<std::string, cls::rbd::TrashImageSpec> images;
if (r >= 0) {
auto bl_it = m_out_bl.cbegin();
r = librbd::cls_client::trash_list_finish(&bl_it, &images);
}
Context *on_init_finish = nullptr;
{
std::lock_guard locker{m_lock};
ceph_assert(m_trash_list_in_progress);
if (r >= 0) {
for (auto& image : images) {
add_image(image.first, image.second);
}
} else if (r == -ENOENT) {
r = 0;
}
if (r == -EBLOCKLISTED) {
dout(0) << "detected client is blocklisted during trash refresh" << dendl;
m_trash_list_in_progress = false;
std::swap(on_init_finish, m_on_init_finish);
} else if (r >= 0 && images.size() < MAX_RETURN) {
m_trash_list_in_progress = false;
std::swap(on_init_finish, m_on_init_finish);
} else if (r < 0) {
m_trash_list_in_progress = false;
}
}
if (r >= 0 && images.size() == MAX_RETURN) {
m_last_image_id = images.rbegin()->first;
trash_list(false);
return;
} else if (r < 0 && r != -EBLOCKLISTED) {
derr << "failed to retrieve trash directory: " << cpp_strerror(r) << dendl;
schedule_trash_list(10);
}
m_async_op_tracker.finish_op();
if (on_init_finish != nullptr) {
on_init_finish->complete(r);
}
}
template <typename I>
void TrashWatcher<I>::schedule_trash_list(double interval) {
std::scoped_lock locker{m_threads->timer_lock, m_lock};
if (m_shutting_down || m_trash_list_in_progress || m_timer_ctx != nullptr) {
if (m_trash_list_in_progress && !m_deferred_trash_list) {
dout(5) << "deferring refresh until in-flight refresh completes" << dendl;
m_deferred_trash_list = true;
}
return;
}
dout(5) << dendl;
m_timer_ctx = m_threads->timer->add_event_after(
interval,
new LambdaContext([this](int r) {
process_trash_list();
}));
}
template <typename I>
void TrashWatcher<I>::process_trash_list() {
dout(5) << dendl;
ceph_assert(ceph_mutex_is_locked(m_threads->timer_lock));
ceph_assert(m_timer_ctx != nullptr);
m_timer_ctx = nullptr;
{
std::lock_guard locker{m_lock};
ceph_assert(!m_trash_list_in_progress);
m_trash_list_in_progress = true;
}
// execute outside of the timer's lock
m_async_op_tracker.start_op();
Context *ctx = new LambdaContext([this](int r) {
create_trash();
m_async_op_tracker.finish_op();
});
m_threads->work_queue->queue(ctx, 0);
}
template <typename I>
void TrashWatcher<I>::add_image(const std::string& image_id,
const cls::rbd::TrashImageSpec& spec) {
if (spec.source != cls::rbd::TRASH_IMAGE_SOURCE_MIRRORING) {
return;
}
ceph_assert(ceph_mutex_is_locked(m_lock));
auto& deferment_end_time = spec.deferment_end_time;
dout(10) << "image_id=" << image_id << ", "
<< "deferment_end_time=" << deferment_end_time << dendl;
m_async_op_tracker.start_op();
auto ctx = new LambdaContext([this, image_id, deferment_end_time](int r) {
m_trash_listener.handle_trash_image(image_id,
deferment_end_time.to_real_time());
m_async_op_tracker.finish_op();
});
m_threads->work_queue->queue(ctx, 0);
}
} // namespace image_deleter;
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_deleter::TrashWatcher<librbd::ImageCtx>;
| 10,581 | 26.485714 | 80 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_deleter/TrashWatcher.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef CEPH_RBD_MIRROR_IMAGE_DELETE_TRASH_WATCHER_H
#define CEPH_RBD_MIRROR_IMAGE_DELETE_TRASH_WATCHER_H
#include "include/rados/librados.hpp"
#include "common/AsyncOpTracker.h"
#include "common/ceph_mutex.h"
#include "librbd/TrashWatcher.h"
#include <set>
#include <string>
struct Context;
namespace librbd { struct ImageCtx; }
namespace rbd {
namespace mirror {
template <typename> struct Threads;
namespace image_deleter {
struct TrashListener;
template <typename ImageCtxT = librbd::ImageCtx>
class TrashWatcher : public librbd::TrashWatcher<ImageCtxT> {
public:
static TrashWatcher* create(librados::IoCtx &io_ctx,
Threads<ImageCtxT> *threads,
TrashListener& trash_listener) {
return new TrashWatcher(io_ctx, threads, trash_listener);
}
TrashWatcher(librados::IoCtx &io_ctx, Threads<ImageCtxT> *threads,
TrashListener& trash_listener);
TrashWatcher(const TrashWatcher&) = delete;
TrashWatcher& operator=(const TrashWatcher&) = delete;
void init(Context *on_finish);
void shut_down(Context *on_finish);
protected:
void handle_image_added(const std::string &image_id,
const cls::rbd::TrashImageSpec& spec) override;
void handle_image_removed(const std::string &image_id) override;
void handle_rewatch_complete(int r) override;
private:
/**
* @verbatim
*
* <start>
* |
* v
* INIT
* |
* v
* CREATE_TRASH
* |
* v
* REGISTER_WATCHER
* |
* |/--------------------------------\
* | |
* |/---------\ |
* | | |
* v | (more images) |
* TRASH_LIST ---/ |
* | |
* |/----------------------------\ |
* | | |
* v | |
* <idle> --\ | |
* | | | |
* | |\---> IMAGE_ADDED -----/ |
* | | |
* | \----> WATCH_ERROR ---------/
* v
* SHUT_DOWN
* |
* v
* UNREGISTER_WATCHER
* |
* v
* <finish>
*
* @endverbatim
*/
librados::IoCtx m_io_ctx;
Threads<ImageCtxT> *m_threads;
TrashListener& m_trash_listener;
std::string m_last_image_id;
bufferlist m_out_bl;
mutable ceph::mutex m_lock;
Context *m_on_init_finish = nullptr;
Context *m_timer_ctx = nullptr;
AsyncOpTracker m_async_op_tracker;
bool m_trash_list_in_progress = false;
bool m_deferred_trash_list = false;
bool m_shutting_down = false;
void register_watcher();
void handle_register_watcher(int r);
void create_trash();
void handle_create_trash(int r);
void unregister_watcher(Context* on_finish);
void handle_unregister_watcher(int r, Context* on_finish);
void trash_list(bool initial_request);
void handle_trash_list(int r);
void schedule_trash_list(double interval);
void process_trash_list();
void get_mirror_uuid();
void handle_get_mirror_uuid(int r);
void add_image(const std::string& image_id,
const cls::rbd::TrashImageSpec& spec);
};
} // namespace image_deleter
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_deleter::TrashWatcher<librbd::ImageCtx>;
#endif // CEPH_RBD_MIRROR_IMAGE_DELETE_TRASH_WATCHER_H
| 3,602 | 24.735714 | 81 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_deleter/Types.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef CEPH_RBD_MIRROR_IMAGE_DELETER_TYPES_H
#define CEPH_RBD_MIRROR_IMAGE_DELETER_TYPES_H
#include "include/Context.h"
#include "librbd/journal/Policy.h"
#include <string>
struct utime_t;
namespace rbd {
namespace mirror {
namespace image_deleter {
enum ErrorResult {
ERROR_RESULT_COMPLETE,
ERROR_RESULT_RETRY,
ERROR_RESULT_RETRY_IMMEDIATELY
};
struct TrashListener {
TrashListener() {
}
TrashListener(const TrashListener&) = delete;
TrashListener& operator=(const TrashListener&) = delete;
virtual ~TrashListener() {
}
virtual void handle_trash_image(const std::string& image_id,
const ceph::real_clock::time_point& deferment_end_time) = 0;
};
struct JournalPolicy : public librbd::journal::Policy {
bool append_disabled() const override {
return true;
}
bool journal_disabled() const override {
return true;
}
void allocate_tag_on_lock(Context *on_finish) override {
on_finish->complete(0);
}
};
} // namespace image_deleter
} // namespace mirror
} // namespace rbd
#endif // CEPH_RBD_MIRROR_IMAGE_DELETER_TYPES_H
| 1,177 | 20.418182 | 70 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_map/LoadRequest.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "common/debug.h"
#include "common/errno.h"
#include "librbd/Utils.h"
#include "include/rbd_types.h"
#include "cls/rbd/cls_rbd_client.h"
#include "UpdateRequest.h"
#include "LoadRequest.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_map::LoadRequest: " \
<< this << " " << __func__
namespace rbd {
namespace mirror {
namespace image_map {
static const uint32_t MAX_RETURN = 1024;
using librbd::util::create_rados_callback;
using librbd::util::create_context_callback;
template<typename I>
LoadRequest<I>::LoadRequest(librados::IoCtx &ioctx,
std::map<std::string, cls::rbd::MirrorImageMap> *image_mapping,
Context *on_finish)
: m_ioctx(ioctx),
m_image_mapping(image_mapping),
m_on_finish(on_finish) {
}
template<typename I>
void LoadRequest<I>::send() {
dout(20) << dendl;
image_map_list();
}
template<typename I>
void LoadRequest<I>::image_map_list() {
dout(20) << dendl;
librados::ObjectReadOperation op;
librbd::cls_client::mirror_image_map_list_start(&op, m_start_after, MAX_RETURN);
librados::AioCompletion *aio_comp = create_rados_callback<
LoadRequest, &LoadRequest::handle_image_map_list>(this);
m_out_bl.clear();
int r = m_ioctx.aio_operate(RBD_MIRROR_LEADER, aio_comp, &op, &m_out_bl);
ceph_assert(r == 0);
aio_comp->release();
}
template<typename I>
void LoadRequest<I>::handle_image_map_list(int r) {
dout(20) << ": r=" << r << dendl;
std::map<std::string, cls::rbd::MirrorImageMap> image_mapping;
if (r == 0) {
auto it = m_out_bl.cbegin();
r = librbd::cls_client::mirror_image_map_list_finish(&it, &image_mapping);
}
if (r < 0) {
derr << ": failed to get image map: " << cpp_strerror(r) << dendl;
finish(r);
return;
}
m_image_mapping->insert(image_mapping.begin(), image_mapping.end());
if (image_mapping.size() == MAX_RETURN) {
m_start_after = image_mapping.rbegin()->first;
image_map_list();
return;
}
mirror_image_list();
}
template<typename I>
void LoadRequest<I>::mirror_image_list() {
dout(20) << dendl;
librados::ObjectReadOperation op;
librbd::cls_client::mirror_image_list_start(&op, m_start_after, MAX_RETURN);
m_out_bl.clear();
librados::AioCompletion *aio_comp = create_rados_callback<
LoadRequest<I>,
&LoadRequest<I>::handle_mirror_image_list>(this);
int r = m_ioctx.aio_operate(RBD_MIRRORING, aio_comp, &op, &m_out_bl);
ceph_assert(r == 0);
aio_comp->release();
}
template<typename I>
void LoadRequest<I>::handle_mirror_image_list(int r) {
dout(20) << ": r=" << r << dendl;
std::map<std::string, std::string> ids;
if (r == 0) {
auto it = m_out_bl.cbegin();
r = librbd::cls_client::mirror_image_list_finish(&it, &ids);
}
if (r < 0 && r != -ENOENT) {
derr << "failed to list mirrored images: " << cpp_strerror(r) << dendl;
finish(r);
return;
}
for (auto &id : ids) {
m_global_image_ids.emplace(id.second);
}
if (ids.size() == MAX_RETURN) {
m_start_after = ids.rbegin()->first;
mirror_image_list();
return;
}
cleanup_image_map();
}
template<typename I>
void LoadRequest<I>::cleanup_image_map() {
dout(20) << dendl;
std::set<std::string> map_removals;
auto it = m_image_mapping->begin();
while (it != m_image_mapping->end()) {
if (m_global_image_ids.count(it->first) > 0) {
++it;
continue;
}
map_removals.emplace(it->first);
it = m_image_mapping->erase(it);
}
if (map_removals.size() == 0) {
finish(0);
return;
}
auto ctx = create_context_callback<
LoadRequest<I>,
&LoadRequest<I>::finish>(this);
image_map::UpdateRequest<I> *req = image_map::UpdateRequest<I>::create(
m_ioctx, {}, std::move(map_removals), ctx);
req->send();
}
template<typename I>
void LoadRequest<I>::finish(int r) {
dout(20) << ": r=" << r << dendl;
m_on_finish->complete(r);
delete this;
}
} // namespace image_map
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_map::LoadRequest<librbd::ImageCtx>;
| 4,296 | 23.554286 | 91 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_map/LoadRequest.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef CEPH_RBD_MIRROR_IMAGE_MAP_LOAD_REQUEST_H
#define CEPH_RBD_MIRROR_IMAGE_MAP_LOAD_REQUEST_H
#include "cls/rbd/cls_rbd_types.h"
#include "include/rados/librados.hpp"
class Context;
namespace librbd { class ImageCtx; }
namespace rbd {
namespace mirror {
namespace image_map {
template<typename ImageCtxT = librbd::ImageCtx>
class LoadRequest {
public:
static LoadRequest *create(librados::IoCtx &ioctx,
std::map<std::string, cls::rbd::MirrorImageMap> *image_mapping,
Context *on_finish) {
return new LoadRequest(ioctx, image_mapping, on_finish);
}
void send();
private:
/**
* @verbatim
*
* <start>
* | . . . . . . . .
* v v . MAX_RETURN
* IMAGE_MAP_LIST. . . . . . .
* |
* v
* MIRROR_IMAGE_LIST
* |
* v
* CLEANUP_IMAGE_MAP
* |
* v
* <finish>
*
* @endverbatim
*/
LoadRequest(librados::IoCtx &ioctx,
std::map<std::string, cls::rbd::MirrorImageMap> *image_mapping,
Context *on_finish);
librados::IoCtx &m_ioctx;
std::map<std::string, cls::rbd::MirrorImageMap> *m_image_mapping;
Context *m_on_finish;
std::set<std::string> m_global_image_ids;
bufferlist m_out_bl;
std::string m_start_after;
void image_map_list();
void handle_image_map_list(int r);
void mirror_image_list();
void handle_mirror_image_list(int r);
void cleanup_image_map();
void finish(int r);
};
} // namespace image_map
} // namespace mirror
} // namespace rbd
#endif // CEPH_RBD_MIRROR_IMAGE_MAP_LOAD_REQUEST_H
| 1,758 | 21.551282 | 92 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_map/Policy.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "common/debug.h"
#include "common/errno.h"
#include "librbd/Utils.h"
#include "Policy.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_map::Policy: " << this \
<< " " << __func__ << ": "
namespace rbd {
namespace mirror {
namespace image_map {
namespace {
bool is_instance_action(ActionType action_type) {
switch (action_type) {
case ACTION_TYPE_ACQUIRE:
case ACTION_TYPE_RELEASE:
return true;
case ACTION_TYPE_NONE:
case ACTION_TYPE_MAP_UPDATE:
case ACTION_TYPE_MAP_REMOVE:
break;
}
return false;
}
} // anonymous namespace
using ::operator<<;
using librbd::util::unique_lock_name;
Policy::Policy(librados::IoCtx &ioctx)
: m_ioctx(ioctx),
m_map_lock(ceph::make_shared_mutex(
unique_lock_name("rbd::mirror::image_map::Policy::m_map_lock", this))) {
// map should at least have once instance
std::string instance_id = stringify(ioctx.get_instance_id());
m_map.emplace(instance_id, std::set<std::string>{});
}
void Policy::init(
const std::map<std::string, cls::rbd::MirrorImageMap> &image_mapping) {
dout(20) << dendl;
std::unique_lock map_lock{m_map_lock};
for (auto& it : image_mapping) {
ceph_assert(!it.second.instance_id.empty());
auto map_result = m_map[it.second.instance_id].emplace(it.first);
ceph_assert(map_result.second);
auto image_state_result = m_image_states.emplace(
it.first, ImageState{it.second.instance_id, it.second.mapped_time});
ceph_assert(image_state_result.second);
// ensure we (re)send image acquire actions to the instance
auto& image_state = image_state_result.first->second;
auto start_action = set_state(&image_state,
StateTransition::STATE_INITIALIZING, false);
ceph_assert(start_action);
}
}
LookupInfo Policy::lookup(const std::string &global_image_id) {
dout(20) << "global_image_id=" << global_image_id << dendl;
std::shared_lock map_lock{m_map_lock};
LookupInfo info;
auto it = m_image_states.find(global_image_id);
if (it != m_image_states.end()) {
info.instance_id = it->second.instance_id;
info.mapped_time = it->second.mapped_time;
}
return info;
}
bool Policy::add_image(const std::string &global_image_id) {
dout(5) << "global_image_id=" << global_image_id << dendl;
std::unique_lock map_lock{m_map_lock};
auto image_state_result = m_image_states.emplace(global_image_id,
ImageState{});
auto& image_state = image_state_result.first->second;
if (image_state.state == StateTransition::STATE_INITIALIZING) {
// avoid duplicate acquire notifications upon leader startup
return false;
}
return set_state(&image_state, StateTransition::STATE_ASSOCIATING, false);
}
bool Policy::remove_image(const std::string &global_image_id) {
dout(5) << "global_image_id=" << global_image_id << dendl;
std::unique_lock map_lock{m_map_lock};
auto it = m_image_states.find(global_image_id);
if (it == m_image_states.end()) {
return false;
}
auto& image_state = it->second;
return set_state(&image_state, StateTransition::STATE_DISSOCIATING, false);
}
void Policy::add_instances(const InstanceIds &instance_ids,
GlobalImageIds* global_image_ids) {
dout(5) << "instance_ids=" << instance_ids << dendl;
std::unique_lock map_lock{m_map_lock};
for (auto& instance : instance_ids) {
ceph_assert(!instance.empty());
m_map.emplace(instance, std::set<std::string>{});
}
// post-failover, remove any dead instances and re-shuffle their images
if (m_initial_update) {
dout(5) << "initial instance update" << dendl;
m_initial_update = false;
std::set<std::string> alive_instances(instance_ids.begin(),
instance_ids.end());
InstanceIds dead_instances;
for (auto& map_pair : m_map) {
if (alive_instances.find(map_pair.first) == alive_instances.end()) {
dead_instances.push_back(map_pair.first);
}
}
if (!dead_instances.empty()) {
remove_instances(m_map_lock, dead_instances, global_image_ids);
}
}
GlobalImageIds shuffle_global_image_ids;
do_shuffle_add_instances(m_map, m_image_states.size(), &shuffle_global_image_ids);
dout(5) << "shuffling global_image_ids=[" << shuffle_global_image_ids
<< "]" << dendl;
for (auto& global_image_id : shuffle_global_image_ids) {
auto it = m_image_states.find(global_image_id);
ceph_assert(it != m_image_states.end());
auto& image_state = it->second;
if (set_state(&image_state, StateTransition::STATE_SHUFFLING, false)) {
global_image_ids->emplace(global_image_id);
}
}
}
void Policy::remove_instances(const InstanceIds &instance_ids,
GlobalImageIds* global_image_ids) {
std::unique_lock map_lock{m_map_lock};
remove_instances(m_map_lock, instance_ids, global_image_ids);
}
void Policy::remove_instances(const ceph::shared_mutex& lock,
const InstanceIds &instance_ids,
GlobalImageIds* global_image_ids) {
ceph_assert(ceph_mutex_is_wlocked(m_map_lock));
dout(5) << "instance_ids=" << instance_ids << dendl;
for (auto& instance_id : instance_ids) {
auto map_it = m_map.find(instance_id);
if (map_it == m_map.end()) {
continue;
}
auto& instance_global_image_ids = map_it->second;
if (instance_global_image_ids.empty()) {
m_map.erase(map_it);
continue;
}
m_dead_instances.insert(instance_id);
dout(5) << "force shuffling: instance_id=" << instance_id << ", "
<< "global_image_ids=[" << instance_global_image_ids << "]"<< dendl;
for (auto& global_image_id : instance_global_image_ids) {
auto it = m_image_states.find(global_image_id);
ceph_assert(it != m_image_states.end());
auto& image_state = it->second;
if (is_state_scheduled(image_state,
StateTransition::STATE_DISSOCIATING)) {
// don't shuffle images that no longer exist
continue;
}
if (set_state(&image_state, StateTransition::STATE_SHUFFLING, true)) {
global_image_ids->emplace(global_image_id);
}
}
}
}
ActionType Policy::start_action(const std::string &global_image_id) {
std::unique_lock map_lock{m_map_lock};
auto it = m_image_states.find(global_image_id);
ceph_assert(it != m_image_states.end());
auto& image_state = it->second;
auto& transition = image_state.transition;
ceph_assert(transition.action_type != ACTION_TYPE_NONE);
dout(5) << "global_image_id=" << global_image_id << ", "
<< "state=" << image_state.state << ", "
<< "action_type=" << transition.action_type << dendl;
if (transition.start_policy_action) {
execute_policy_action(global_image_id, &image_state,
*transition.start_policy_action);
transition.start_policy_action = boost::none;
}
return transition.action_type;
}
bool Policy::finish_action(const std::string &global_image_id, int r) {
std::unique_lock map_lock{m_map_lock};
auto it = m_image_states.find(global_image_id);
ceph_assert(it != m_image_states.end());
auto& image_state = it->second;
auto& transition = image_state.transition;
dout(5) << "global_image_id=" << global_image_id << ", "
<< "state=" << image_state.state << ", "
<< "action_type=" << transition.action_type << ", "
<< "r=" << r << dendl;
// retry on failure unless it's an RPC message to an instance that is dead
if (r < 0 &&
(!is_instance_action(image_state.transition.action_type) ||
image_state.instance_id == UNMAPPED_INSTANCE_ID ||
m_dead_instances.find(image_state.instance_id) ==
m_dead_instances.end())) {
return true;
}
auto finish_policy_action = transition.finish_policy_action;
StateTransition::transit(image_state.state, &image_state.transition);
if (transition.finish_state) {
// in-progress state machine complete
ceph_assert(StateTransition::is_idle(*transition.finish_state));
image_state.state = *transition.finish_state;
image_state.transition = {};
}
if (StateTransition::is_idle(image_state.state) && image_state.next_state) {
// advance to pending state machine
bool start_action = set_state(&image_state, *image_state.next_state, false);
ceph_assert(start_action);
}
// image state may get purged in execute_policy_action()
bool pending_action = image_state.transition.action_type != ACTION_TYPE_NONE;
if (finish_policy_action) {
execute_policy_action(global_image_id, &image_state, *finish_policy_action);
}
return pending_action;
}
void Policy::execute_policy_action(
const std::string& global_image_id, ImageState* image_state,
StateTransition::PolicyAction policy_action) {
dout(5) << "global_image_id=" << global_image_id << ", "
<< "policy_action=" << policy_action << dendl;
switch (policy_action) {
case StateTransition::POLICY_ACTION_MAP:
map(global_image_id, image_state);
break;
case StateTransition::POLICY_ACTION_UNMAP:
unmap(global_image_id, image_state);
break;
case StateTransition::POLICY_ACTION_REMOVE:
if (image_state->state == StateTransition::STATE_UNASSOCIATED) {
ceph_assert(image_state->instance_id == UNMAPPED_INSTANCE_ID);
ceph_assert(!image_state->next_state);
m_image_states.erase(global_image_id);
}
break;
}
}
void Policy::map(const std::string& global_image_id, ImageState* image_state) {
ceph_assert(ceph_mutex_is_wlocked(m_map_lock));
std::string instance_id = image_state->instance_id;
if (instance_id != UNMAPPED_INSTANCE_ID && !is_dead_instance(instance_id)) {
return;
}
if (is_dead_instance(instance_id)) {
unmap(global_image_id, image_state);
}
instance_id = do_map(m_map, global_image_id);
ceph_assert(!instance_id.empty());
dout(5) << "global_image_id=" << global_image_id << ", "
<< "instance_id=" << instance_id << dendl;
image_state->instance_id = instance_id;
image_state->mapped_time = ceph_clock_now();
auto ins = m_map[instance_id].emplace(global_image_id);
ceph_assert(ins.second);
}
void Policy::unmap(const std::string &global_image_id,
ImageState* image_state) {
ceph_assert(ceph_mutex_is_wlocked(m_map_lock));
std::string instance_id = image_state->instance_id;
if (instance_id == UNMAPPED_INSTANCE_ID) {
return;
}
dout(5) << "global_image_id=" << global_image_id << ", "
<< "instance_id=" << instance_id << dendl;
ceph_assert(!instance_id.empty());
m_map[instance_id].erase(global_image_id);
image_state->instance_id = UNMAPPED_INSTANCE_ID;
image_state->mapped_time = {};
if (is_dead_instance(instance_id) && m_map[instance_id].empty()) {
dout(5) << "removing dead instance_id=" << instance_id << dendl;
m_map.erase(instance_id);
m_dead_instances.erase(instance_id);
}
}
bool Policy::is_image_shuffling(const std::string &global_image_id) {
ceph_assert(ceph_mutex_is_locked(m_map_lock));
auto it = m_image_states.find(global_image_id);
ceph_assert(it != m_image_states.end());
auto& image_state = it->second;
// avoid attempting to re-shuffle a pending shuffle
auto result = is_state_scheduled(image_state,
StateTransition::STATE_SHUFFLING);
dout(20) << "global_image_id=" << global_image_id << ", "
<< "result=" << result << dendl;
return result;
}
bool Policy::can_shuffle_image(const std::string &global_image_id) {
ceph_assert(ceph_mutex_is_locked(m_map_lock));
CephContext *cct = reinterpret_cast<CephContext *>(m_ioctx.cct());
int migration_throttle = cct->_conf.get_val<uint64_t>(
"rbd_mirror_image_policy_migration_throttle");
auto it = m_image_states.find(global_image_id);
ceph_assert(it != m_image_states.end());
auto& image_state = it->second;
utime_t last_shuffled_time = image_state.mapped_time;
// idle images that haven't been recently remapped can shuffle
utime_t now = ceph_clock_now();
auto result = (StateTransition::is_idle(image_state.state) &&
((migration_throttle <= 0) ||
(now - last_shuffled_time >= migration_throttle)));
dout(10) << "global_image_id=" << global_image_id << ", "
<< "migration_throttle=" << migration_throttle << ", "
<< "last_shuffled_time=" << last_shuffled_time << ", "
<< "result=" << result << dendl;
return result;
}
bool Policy::set_state(ImageState* image_state, StateTransition::State state,
bool ignore_current_state) {
if (!ignore_current_state && image_state->state == state) {
image_state->next_state = boost::none;
return false;
} else if (StateTransition::is_idle(image_state->state)) {
image_state->state = state;
image_state->next_state = boost::none;
StateTransition::transit(image_state->state, &image_state->transition);
ceph_assert(image_state->transition.action_type != ACTION_TYPE_NONE);
ceph_assert(!image_state->transition.finish_state);
return true;
}
image_state->next_state = state;
return false;
}
bool Policy::is_state_scheduled(const ImageState& image_state,
StateTransition::State state) const {
return (image_state.state == state ||
(image_state.next_state && *image_state.next_state == state));
}
} // namespace image_map
} // namespace mirror
} // namespace rbd
| 13,796 | 32.816176 | 84 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_map/Policy.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef CEPH_RBD_MIRROR_IMAGE_MAP_POLICY_H
#define CEPH_RBD_MIRROR_IMAGE_MAP_POLICY_H
#include <map>
#include <tuple>
#include <boost/optional.hpp>
#include "cls/rbd/cls_rbd_types.h"
#include "include/rados/librados.hpp"
#include "tools/rbd_mirror/image_map/StateTransition.h"
#include "tools/rbd_mirror/image_map/Types.h"
class Context;
namespace rbd {
namespace mirror {
namespace image_map {
class Policy {
public:
Policy(librados::IoCtx &ioctx);
virtual ~Policy() {
}
// init -- called during initialization
void init(
const std::map<std::string, cls::rbd::MirrorImageMap> &image_mapping);
// lookup an image from the map
LookupInfo lookup(const std::string &global_image_id);
// add, remove
bool add_image(const std::string &global_image_id);
bool remove_image(const std::string &global_image_id);
// shuffle images when instances are added/removed
void add_instances(const InstanceIds &instance_ids,
GlobalImageIds* global_image_ids);
void remove_instances(const InstanceIds &instance_ids,
GlobalImageIds* global_image_ids);
ActionType start_action(const std::string &global_image_id);
bool finish_action(const std::string &global_image_id, int r);
protected:
typedef std::map<std::string, std::set<std::string> > InstanceToImageMap;
bool is_dead_instance(const std::string instance_id) {
ceph_assert(ceph_mutex_is_locked(m_map_lock));
return m_dead_instances.find(instance_id) != m_dead_instances.end();
}
bool is_image_shuffling(const std::string &global_image_id);
bool can_shuffle_image(const std::string &global_image_id);
// map an image (global image id) to an instance
virtual std::string do_map(const InstanceToImageMap& map,
const std::string &global_image_id) = 0;
// shuffle images when instances are added/removed
virtual void do_shuffle_add_instances(
const InstanceToImageMap& map, size_t image_count,
std::set<std::string> *remap_global_image_ids) = 0;
private:
struct ImageState {
std::string instance_id = UNMAPPED_INSTANCE_ID;
utime_t mapped_time;
ImageState() {}
ImageState(const std::string& instance_id, const utime_t& mapped_time)
: instance_id(instance_id), mapped_time(mapped_time) {
}
// active state and action
StateTransition::State state = StateTransition::STATE_UNASSOCIATED;
StateTransition::Transition transition;
// next scheduled state
boost::optional<StateTransition::State> next_state = boost::none;
};
typedef std::map<std::string, ImageState> ImageStates;
librados::IoCtx &m_ioctx;
ceph::shared_mutex m_map_lock; // protects m_map
InstanceToImageMap m_map; // instance_id -> global_id map
ImageStates m_image_states;
std::set<std::string> m_dead_instances;
bool m_initial_update = true;
void remove_instances(const ceph::shared_mutex& lock,
const InstanceIds &instance_ids,
GlobalImageIds* global_image_ids);
bool set_state(ImageState* image_state, StateTransition::State state,
bool ignore_current_state);
void execute_policy_action(const std::string& global_image_id,
ImageState* image_state,
StateTransition::PolicyAction policy_action);
void map(const std::string& global_image_id, ImageState* image_state);
void unmap(const std::string &global_image_id, ImageState* image_state);
bool is_state_scheduled(const ImageState& image_state,
StateTransition::State state) const;
};
} // namespace image_map
} // namespace mirror
} // namespace rbd
#endif // CEPH_RBD_MIRROR_IMAGE_MAP_POLICY_H
| 3,826 | 30.113821 | 76 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_map/SimplePolicy.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "common/debug.h"
#include "common/errno.h"
#include "SimplePolicy.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_map::SimplePolicy: " << this \
<< " " << __func__ << ": "
namespace rbd {
namespace mirror {
namespace image_map {
SimplePolicy::SimplePolicy(librados::IoCtx &ioctx)
: Policy(ioctx) {
}
size_t SimplePolicy::calc_images_per_instance(const InstanceToImageMap& map,
size_t image_count) {
size_t nr_instances = 0;
for (auto const &it : map) {
if (!Policy::is_dead_instance(it.first)) {
++nr_instances;
}
}
ceph_assert(nr_instances > 0);
size_t images_per_instance = image_count / nr_instances;
if (images_per_instance == 0) {
++images_per_instance;
}
return images_per_instance;
}
void SimplePolicy::do_shuffle_add_instances(
const InstanceToImageMap& map, size_t image_count,
std::set<std::string> *remap_global_image_ids) {
uint64_t images_per_instance = calc_images_per_instance(map, image_count);
dout(5) << "images per instance=" << images_per_instance << dendl;
for (auto const &instance : map) {
if (instance.second.size() <= images_per_instance) {
continue;
}
auto it = instance.second.begin();
uint64_t cut_off = instance.second.size() - images_per_instance;
while (it != instance.second.end() && cut_off > 0) {
if (Policy::is_image_shuffling(*it)) {
--cut_off;
} else if (Policy::can_shuffle_image(*it)) {
--cut_off;
remap_global_image_ids->emplace(*it);
}
++it;
}
}
}
std::string SimplePolicy::do_map(const InstanceToImageMap& map,
const std::string &global_image_id) {
auto min_it = map.end();
for (auto it = map.begin(); it != map.end(); ++it) {
ceph_assert(it->second.find(global_image_id) == it->second.end());
if (Policy::is_dead_instance(it->first)) {
continue;
} else if (min_it == map.end()) {
min_it = it;
} else if (it->second.size() < min_it->second.size()) {
min_it = it;
}
}
ceph_assert(min_it != map.end());
dout(20) << "global_image_id=" << global_image_id << " maps to instance_id="
<< min_it->first << dendl;
return min_it->first;
}
} // namespace image_map
} // namespace mirror
} // namespace rbd
| 2,559 | 27.444444 | 80 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_map/SimplePolicy.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef CEPH_RBD_MIRROR_IMAGE_MAP_SIMPLE_POLICY_H
#define CEPH_RBD_MIRROR_IMAGE_MAP_SIMPLE_POLICY_H
#include "Policy.h"
namespace rbd {
namespace mirror {
namespace image_map {
class SimplePolicy : public Policy {
public:
static SimplePolicy *create(librados::IoCtx &ioctx) {
return new SimplePolicy(ioctx);
}
protected:
SimplePolicy(librados::IoCtx &ioctx);
std::string do_map(const InstanceToImageMap& map,
const std::string &global_image_id) override;
void do_shuffle_add_instances(
const InstanceToImageMap& map, size_t image_count,
std::set<std::string> *remap_global_image_ids) override;
private:
size_t calc_images_per_instance(const InstanceToImageMap& map,
size_t image_count);
};
} // namespace image_map
} // namespace mirror
} // namespace rbd
#endif // CEPH_RBD_MIRROR_IMAGE_MAP_SIMPLE_POLICY_H
| 1,000 | 24.025 | 70 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_map/StateTransition.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include <ostream>
#include "include/ceph_assert.h"
#include "StateTransition.h"
namespace rbd {
namespace mirror {
namespace image_map {
std::ostream &operator<<(std::ostream &os,
const StateTransition::State &state) {
switch(state) {
case StateTransition::STATE_INITIALIZING:
os << "INITIALIZING";
break;
case StateTransition::STATE_ASSOCIATING:
os << "ASSOCIATING";
break;
case StateTransition::STATE_ASSOCIATED:
os << "ASSOCIATED";
break;
case StateTransition::STATE_SHUFFLING:
os << "SHUFFLING";
break;
case StateTransition::STATE_DISSOCIATING:
os << "DISSOCIATING";
break;
case StateTransition::STATE_UNASSOCIATED:
os << "UNASSOCIATED";
break;
}
return os;
}
std::ostream &operator<<(std::ostream &os,
const StateTransition::PolicyAction &policy_action) {
switch(policy_action) {
case StateTransition::POLICY_ACTION_MAP:
os << "MAP";
break;
case StateTransition::POLICY_ACTION_UNMAP:
os << "UNMAP";
break;
case StateTransition::POLICY_ACTION_REMOVE:
os << "REMOVE";
break;
}
return os;
}
const StateTransition::TransitionTable StateTransition::s_transition_table {
// state current_action Transition
// ---------------------------------------------------------------------------
{{STATE_INITIALIZING, ACTION_TYPE_NONE}, {ACTION_TYPE_ACQUIRE, {}, {},
{}}},
{{STATE_INITIALIZING, ACTION_TYPE_ACQUIRE}, {ACTION_TYPE_NONE, {}, {},
{STATE_ASSOCIATED}}},
{{STATE_ASSOCIATING, ACTION_TYPE_NONE}, {ACTION_TYPE_MAP_UPDATE,
{POLICY_ACTION_MAP}, {}, {}}},
{{STATE_ASSOCIATING, ACTION_TYPE_MAP_UPDATE}, {ACTION_TYPE_ACQUIRE, {}, {},
{}}},
{{STATE_ASSOCIATING, ACTION_TYPE_ACQUIRE}, {ACTION_TYPE_NONE, {}, {},
{STATE_ASSOCIATED}}},
{{STATE_DISSOCIATING, ACTION_TYPE_NONE}, {ACTION_TYPE_RELEASE, {},
{POLICY_ACTION_UNMAP}, {}}},
{{STATE_DISSOCIATING, ACTION_TYPE_RELEASE}, {ACTION_TYPE_MAP_REMOVE, {},
{POLICY_ACTION_REMOVE}, {}}},
{{STATE_DISSOCIATING, ACTION_TYPE_MAP_REMOVE}, {ACTION_TYPE_NONE, {},
{}, {STATE_UNASSOCIATED}}},
{{STATE_SHUFFLING, ACTION_TYPE_NONE}, {ACTION_TYPE_RELEASE, {},
{POLICY_ACTION_UNMAP}, {}}},
{{STATE_SHUFFLING, ACTION_TYPE_RELEASE}, {ACTION_TYPE_MAP_UPDATE,
{POLICY_ACTION_MAP}, {}, {}}},
{{STATE_SHUFFLING, ACTION_TYPE_MAP_UPDATE}, {ACTION_TYPE_ACQUIRE, {}, {},
{}}},
{{STATE_SHUFFLING, ACTION_TYPE_ACQUIRE}, {ACTION_TYPE_NONE, {}, {},
{STATE_ASSOCIATED}}}
};
void StateTransition::transit(State state, Transition* transition) {
auto it = s_transition_table.find({state, transition->action_type});
ceph_assert(it != s_transition_table.end());
*transition = it->second;
}
} // namespace image_map
} // namespace mirror
} // namespace rbd
| 3,531 | 36.178947 | 80 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_map/StateTransition.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef CEPH_RBD_MIRROR_IMAGE_MAP_STATE_TRANSITION_H
#define CEPH_RBD_MIRROR_IMAGE_MAP_STATE_TRANSITION_H
#include "tools/rbd_mirror/image_map/Types.h"
#include <boost/optional.hpp>
#include <map>
namespace rbd {
namespace mirror {
namespace image_map {
class StateTransition {
public:
enum State {
STATE_UNASSOCIATED,
STATE_INITIALIZING,
STATE_ASSOCIATING,
STATE_ASSOCIATED,
STATE_SHUFFLING,
STATE_DISSOCIATING
};
enum PolicyAction {
POLICY_ACTION_MAP,
POLICY_ACTION_UNMAP,
POLICY_ACTION_REMOVE
};
struct Transition {
// image map action
ActionType action_type = ACTION_TYPE_NONE;
// policy internal action
boost::optional<PolicyAction> start_policy_action;
boost::optional<PolicyAction> finish_policy_action;
// state machine complete
boost::optional<State> finish_state;
Transition() {
}
Transition(ActionType action_type,
const boost::optional<PolicyAction>& start_policy_action,
const boost::optional<PolicyAction>& finish_policy_action,
const boost::optional<State>& finish_state)
: action_type(action_type), start_policy_action(start_policy_action),
finish_policy_action(finish_policy_action), finish_state(finish_state) {
}
};
static bool is_idle(State state) {
return (state == STATE_UNASSOCIATED || state == STATE_ASSOCIATED);
}
static void transit(State state, Transition* transition);
private:
typedef std::pair<State, ActionType> TransitionKey;
typedef std::map<TransitionKey, Transition> TransitionTable;
// image transition table
static const TransitionTable s_transition_table;
};
std::ostream &operator<<(std::ostream &os, const StateTransition::State &state);
std::ostream &operator<<(std::ostream &os,
const StateTransition::PolicyAction &policy_action);
} // namespace image_map
} // namespace mirror
} // namespace rbd
#endif // CEPH_RBD_MIRROR_IMAGE_MAP_STATE_TRANSITION_H
| 2,103 | 26.324675 | 80 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_map/Types.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "Types.h"
#include "include/ceph_assert.h"
#include "include/stringify.h"
#include "common/Formatter.h"
#include <iostream>
namespace rbd {
namespace mirror {
namespace image_map {
const std::string UNMAPPED_INSTANCE_ID("");
namespace {
template <typename E>
class GetTypeVisitor : public boost::static_visitor<E> {
public:
template <typename T>
inline E operator()(const T&) const {
return T::TYPE;
}
};
class EncodeVisitor : public boost::static_visitor<void> {
public:
explicit EncodeVisitor(bufferlist &bl) : m_bl(bl) {
}
template <typename T>
inline void operator()(const T& t) const {
using ceph::encode;
encode(static_cast<uint32_t>(T::TYPE), m_bl);
t.encode(m_bl);
}
private:
bufferlist &m_bl;
};
class DecodeVisitor : public boost::static_visitor<void> {
public:
DecodeVisitor(__u8 version, bufferlist::const_iterator &iter)
: m_version(version), m_iter(iter) {
}
template <typename T>
inline void operator()(T& t) const {
t.decode(m_version, m_iter);
}
private:
__u8 m_version;
bufferlist::const_iterator &m_iter;
};
class DumpVisitor : public boost::static_visitor<void> {
public:
explicit DumpVisitor(Formatter *formatter, const std::string &key)
: m_formatter(formatter), m_key(key) {}
template <typename T>
inline void operator()(const T& t) const {
auto type = T::TYPE;
m_formatter->dump_string(m_key.c_str(), stringify(type));
t.dump(m_formatter);
}
private:
ceph::Formatter *m_formatter;
std::string m_key;
};
} // anonymous namespace
PolicyMetaType PolicyData::get_policy_meta_type() const {
return boost::apply_visitor(GetTypeVisitor<PolicyMetaType>(), policy_meta);
}
void PolicyData::encode(bufferlist& bl) const {
ENCODE_START(1, 1, bl);
boost::apply_visitor(EncodeVisitor(bl), policy_meta);
ENCODE_FINISH(bl);
}
void PolicyData::decode(bufferlist::const_iterator& it) {
DECODE_START(1, it);
uint32_t policy_meta_type;
decode(policy_meta_type, it);
switch (policy_meta_type) {
case POLICY_META_TYPE_NONE:
policy_meta = PolicyMetaNone();
break;
default:
policy_meta = PolicyMetaUnknown();
break;
}
boost::apply_visitor(DecodeVisitor(struct_v, it), policy_meta);
DECODE_FINISH(it);
}
void PolicyData::dump(Formatter *f) const {
boost::apply_visitor(DumpVisitor(f, "policy_meta_type"), policy_meta);
}
void PolicyData::generate_test_instances(std::list<PolicyData *> &o) {
o.push_back(new PolicyData(PolicyMetaNone()));
}
std::ostream &operator<<(std::ostream &os, const ActionType& action_type) {
switch (action_type) {
case ACTION_TYPE_NONE:
os << "NONE";
break;
case ACTION_TYPE_MAP_UPDATE:
os << "MAP_UPDATE";
break;
case ACTION_TYPE_MAP_REMOVE:
os << "MAP_REMOVE";
break;
case ACTION_TYPE_ACQUIRE:
os << "ACQUIRE";
break;
case ACTION_TYPE_RELEASE:
os << "RELEASE";
break;
default:
os << "UNKNOWN (" << static_cast<uint32_t>(action_type) << ")";
break;
}
return os;
}
} // namespace image_map
} // namespace mirror
} // namespace rbd
| 3,182 | 21.899281 | 77 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_map/Types.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef CEPH_RBD_MIRROR_IMAGE_MAP_TYPES_H
#define CEPH_RBD_MIRROR_IMAGE_MAP_TYPES_H
#include <iosfwd>
#include <map>
#include <set>
#include <string>
#include <boost/variant.hpp>
#include "include/buffer.h"
#include "include/encoding.h"
#include "include/utime.h"
#include "tools/rbd_mirror/Types.h"
struct Context;
namespace ceph {
class Formatter;
}
namespace rbd {
namespace mirror {
namespace image_map {
extern const std::string UNMAPPED_INSTANCE_ID;
struct Listener {
virtual ~Listener() {
}
virtual void acquire_image(const std::string &global_image_id,
const std::string &instance_id,
Context* on_finish) = 0;
virtual void release_image(const std::string &global_image_id,
const std::string &instance_id,
Context* on_finish) = 0;
virtual void remove_image(const std::string &mirror_uuid,
const std::string &global_image_id,
const std::string &instance_id,
Context* on_finish) = 0;
};
struct LookupInfo {
std::string instance_id = UNMAPPED_INSTANCE_ID;
utime_t mapped_time;
};
enum ActionType {
ACTION_TYPE_NONE,
ACTION_TYPE_MAP_UPDATE,
ACTION_TYPE_MAP_REMOVE,
ACTION_TYPE_ACQUIRE,
ACTION_TYPE_RELEASE
};
typedef std::vector<std::string> InstanceIds;
typedef std::set<std::string> GlobalImageIds;
typedef std::map<std::string, ActionType> ImageActionTypes;
enum PolicyMetaType {
POLICY_META_TYPE_NONE = 0,
};
struct PolicyMetaNone {
static const PolicyMetaType TYPE = POLICY_META_TYPE_NONE;
PolicyMetaNone() {
}
void encode(bufferlist& bl) const {
}
void decode(__u8 version, bufferlist::const_iterator& it) {
}
void dump(Formatter *f) const {
}
};
struct PolicyMetaUnknown {
static const PolicyMetaType TYPE = static_cast<PolicyMetaType>(-1);
PolicyMetaUnknown() {
}
void encode(bufferlist& bl) const {
ceph_abort();
}
void decode(__u8 version, bufferlist::const_iterator& it) {
}
void dump(Formatter *f) const {
}
};
typedef boost::variant<PolicyMetaNone,
PolicyMetaUnknown> PolicyMeta;
struct PolicyData {
PolicyData()
: policy_meta(PolicyMetaUnknown()) {
}
PolicyData(const PolicyMeta &policy_meta)
: policy_meta(policy_meta) {
}
PolicyMeta policy_meta;
PolicyMetaType get_policy_meta_type() const;
void encode(bufferlist& bl) const;
void decode(bufferlist::const_iterator& it);
void dump(Formatter *f) const;
static void generate_test_instances(std::list<PolicyData *> &o);
};
WRITE_CLASS_ENCODER(PolicyData);
std::ostream &operator<<(std::ostream &os, const ActionType &action_type);
} // namespace image_map
} // namespace mirror
} // namespace rbd
#endif // CEPH_RBD_MIRROR_IMAGE_MAP_TYPES_H
| 2,955 | 21.564885 | 74 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_map/UpdateRequest.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "common/debug.h"
#include "common/errno.h"
#include "librbd/Utils.h"
#include "include/rbd_types.h"
#include "cls/rbd/cls_rbd_client.h"
#include "UpdateRequest.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_map::UpdateRequest: " \
<< this << " " << __func__
namespace rbd {
namespace mirror {
namespace image_map {
using librbd::util::create_rados_callback;
static const uint32_t MAX_UPDATE = 256;
template <typename I>
UpdateRequest<I>::UpdateRequest(librados::IoCtx &ioctx,
std::map<std::string, cls::rbd::MirrorImageMap> &&update_mapping,
std::set<std::string> &&remove_global_image_ids, Context *on_finish)
: m_ioctx(ioctx),
m_update_mapping(update_mapping),
m_remove_global_image_ids(remove_global_image_ids),
m_on_finish(on_finish) {
}
template <typename I>
void UpdateRequest<I>::send() {
dout(20) << dendl;
update_image_map();
}
template <typename I>
void UpdateRequest<I>::update_image_map() {
dout(20) << dendl;
if (m_update_mapping.empty() && m_remove_global_image_ids.empty()) {
finish(0);
return;
}
uint32_t nr_updates = 0;
librados::ObjectWriteOperation op;
auto it1 = m_update_mapping.begin();
while (it1 != m_update_mapping.end() && nr_updates++ < MAX_UPDATE) {
librbd::cls_client::mirror_image_map_update(&op, it1->first, it1->second);
it1 = m_update_mapping.erase(it1);
}
auto it2 = m_remove_global_image_ids.begin();
while (it2 != m_remove_global_image_ids.end() && nr_updates++ < MAX_UPDATE) {
librbd::cls_client::mirror_image_map_remove(&op, *it2);
it2 = m_remove_global_image_ids.erase(it2);
}
librados::AioCompletion *aio_comp = create_rados_callback<
UpdateRequest, &UpdateRequest::handle_update_image_map>(this);
int r = m_ioctx.aio_operate(RBD_MIRROR_LEADER, aio_comp, &op);
ceph_assert(r == 0);
aio_comp->release();
}
template <typename I>
void UpdateRequest<I>::handle_update_image_map(int r) {
dout(20) << ": r=" << r << dendl;
if (r < 0) {
derr << ": failed to update image map: " << cpp_strerror(r) << dendl;
finish(r);
return;
}
update_image_map();
}
template <typename I>
void UpdateRequest<I>::finish(int r) {
dout(20) << ": r=" << r << dendl;
m_on_finish->complete(r);
delete this;
}
} // namespace image_map
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_map::UpdateRequest<librbd::ImageCtx>;
| 2,686 | 25.60396 | 100 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_map/UpdateRequest.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef CEPH_RBD_MIRROR_IMAGE_MAP_UPDATE_REQUEST_H
#define CEPH_RBD_MIRROR_IMAGE_MAP_UPDATE_REQUEST_H
#include "cls/rbd/cls_rbd_types.h"
#include "include/rados/librados.hpp"
class Context;
namespace librbd { class ImageCtx; }
namespace rbd {
namespace mirror {
namespace image_map {
template<typename ImageCtxT = librbd::ImageCtx>
class UpdateRequest {
public:
// accepts an image map for updation and a collection of
// global image ids to purge.
static UpdateRequest *create(librados::IoCtx &ioctx,
std::map<std::string, cls::rbd::MirrorImageMap> &&update_mapping,
std::set<std::string> &&remove_global_image_ids, Context *on_finish) {
return new UpdateRequest(ioctx, std::move(update_mapping), std::move(remove_global_image_ids),
on_finish);
}
void send();
private:
/**
* @verbatim
*
* <start>
* | . . . . . . . .
* v v . MAX_UPDATE
* UPDATE_IMAGE_MAP. . . . . . .
* |
* v
* <finish>
*
* @endverbatim
*/
UpdateRequest(librados::IoCtx &ioctx,
std::map<std::string, cls::rbd::MirrorImageMap> &&update_mapping,
std::set<std::string> &&remove_global_image_ids, Context *on_finish);
librados::IoCtx &m_ioctx;
std::map<std::string, cls::rbd::MirrorImageMap> m_update_mapping;
std::set<std::string> m_remove_global_image_ids;
Context *m_on_finish;
void update_image_map();
void handle_update_image_map(int r);
void finish(int r);
};
} // namespace image_map
} // namespace mirror
} // namespace rbd
#endif // CEPH_RBD_MIRROR_IMAGE_MAP_UPDATE_REQUEST_H
| 1,811 | 26.454545 | 101 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/BootstrapRequest.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "include/compat.h"
#include "BootstrapRequest.h"
#include "CreateImageRequest.h"
#include "OpenImageRequest.h"
#include "OpenLocalImageRequest.h"
#include "common/debug.h"
#include "common/dout.h"
#include "common/errno.h"
#include "cls/rbd/cls_rbd_client.h"
#include "journal/Journaler.h"
#include "journal/Settings.h"
#include "librbd/ImageCtx.h"
#include "librbd/ImageState.h"
#include "librbd/internal.h"
#include "librbd/Journal.h"
#include "librbd/Utils.h"
#include "librbd/asio/ContextWQ.h"
#include "librbd/journal/Types.h"
#include "tools/rbd_mirror/BaseRequest.h"
#include "tools/rbd_mirror/ImageSync.h"
#include "tools/rbd_mirror/ProgressContext.h"
#include "tools/rbd_mirror/Threads.h"
#include "tools/rbd_mirror/image_replayer/PrepareLocalImageRequest.h"
#include "tools/rbd_mirror/image_replayer/PrepareRemoteImageRequest.h"
#include "tools/rbd_mirror/image_replayer/journal/StateBuilder.h"
#include "tools/rbd_mirror/image_replayer/journal/SyncPointHandler.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_replayer::" \
<< "BootstrapRequest: " << this << " " \
<< __func__ << ": "
namespace rbd {
namespace mirror {
namespace image_replayer {
using librbd::util::create_context_callback;
using librbd::util::unique_lock_name;
template <typename I>
BootstrapRequest<I>::BootstrapRequest(
Threads<I>* threads,
librados::IoCtx& local_io_ctx,
librados::IoCtx& remote_io_ctx,
InstanceWatcher<I>* instance_watcher,
const std::string& global_image_id,
const std::string& local_mirror_uuid,
const RemotePoolMeta& remote_pool_meta,
::journal::CacheManagerHandler* cache_manager_handler,
PoolMetaCache* pool_meta_cache,
ProgressContext* progress_ctx,
StateBuilder<I>** state_builder,
bool* do_resync,
Context* on_finish)
: CancelableRequest("rbd::mirror::image_replayer::BootstrapRequest",
reinterpret_cast<CephContext*>(local_io_ctx.cct()),
on_finish),
m_threads(threads),
m_local_io_ctx(local_io_ctx),
m_remote_io_ctx(remote_io_ctx),
m_instance_watcher(instance_watcher),
m_global_image_id(global_image_id),
m_local_mirror_uuid(local_mirror_uuid),
m_remote_pool_meta(remote_pool_meta),
m_cache_manager_handler(cache_manager_handler),
m_pool_meta_cache(pool_meta_cache),
m_progress_ctx(progress_ctx),
m_state_builder(state_builder),
m_do_resync(do_resync),
m_lock(ceph::make_mutex(unique_lock_name("BootstrapRequest::m_lock",
this))) {
dout(10) << dendl;
}
template <typename I>
bool BootstrapRequest<I>::is_syncing() const {
std::lock_guard locker{m_lock};
return (m_image_sync != nullptr);
}
template <typename I>
void BootstrapRequest<I>::send() {
*m_do_resync = false;
prepare_local_image();
}
template <typename I>
void BootstrapRequest<I>::cancel() {
dout(10) << dendl;
std::lock_guard locker{m_lock};
m_canceled = true;
if (m_image_sync != nullptr) {
m_image_sync->cancel();
}
}
template <typename I>
std::string BootstrapRequest<I>::get_local_image_name() const {
std::unique_lock locker{m_lock};
return m_local_image_name;
}
template <typename I>
void BootstrapRequest<I>::prepare_local_image() {
dout(10) << dendl;
update_progress("PREPARE_LOCAL_IMAGE");
{
std::unique_lock locker{m_lock};
m_local_image_name = m_global_image_id;
}
ceph_assert(*m_state_builder == nullptr);
auto ctx = create_context_callback<
BootstrapRequest, &BootstrapRequest<I>::handle_prepare_local_image>(this);
auto req = image_replayer::PrepareLocalImageRequest<I>::create(
m_local_io_ctx, m_global_image_id, &m_prepare_local_image_name,
m_state_builder, m_threads->work_queue, ctx);
req->send();
}
template <typename I>
void BootstrapRequest<I>::handle_prepare_local_image(int r) {
dout(10) << "r=" << r << dendl;
ceph_assert(r < 0 || *m_state_builder != nullptr);
if (r == -ENOENT) {
dout(10) << "local image does not exist" << dendl;
} else if (r < 0) {
derr << "error preparing local image for replay: " << cpp_strerror(r)
<< dendl;
finish(r);
return;
}
// image replayer will detect the name change (if any) at next
// status update
if (r >= 0 && !m_prepare_local_image_name.empty()) {
std::unique_lock locker{m_lock};
m_local_image_name = m_prepare_local_image_name;
}
prepare_remote_image();
}
template <typename I>
void BootstrapRequest<I>::prepare_remote_image() {
dout(10) << dendl;
update_progress("PREPARE_REMOTE_IMAGE");
Context *ctx = create_context_callback<
BootstrapRequest, &BootstrapRequest<I>::handle_prepare_remote_image>(this);
auto req = image_replayer::PrepareRemoteImageRequest<I>::create(
m_threads, m_local_io_ctx, m_remote_io_ctx, m_global_image_id,
m_local_mirror_uuid, m_remote_pool_meta, m_cache_manager_handler,
m_state_builder, ctx);
req->send();
}
template <typename I>
void BootstrapRequest<I>::handle_prepare_remote_image(int r) {
dout(10) << "r=" << r << dendl;
auto state_builder = *m_state_builder;
ceph_assert(state_builder == nullptr ||
!state_builder->remote_mirror_uuid.empty());
if (state_builder != nullptr && state_builder->is_local_primary()) {
dout(5) << "local image is primary" << dendl;
finish(-ENOMSG);
return;
} else if (r == -ENOENT || state_builder == nullptr) {
dout(10) << "remote image does not exist";
if (state_builder != nullptr) {
*_dout << ": "
<< "local_image_id=" << state_builder->local_image_id << ", "
<< "remote_image_id=" << state_builder->remote_image_id << ", "
<< "is_linked=" << state_builder->is_linked();
}
*_dout << dendl;
// TODO need to support multiple remote images
if (state_builder != nullptr &&
state_builder->remote_image_id.empty() &&
(state_builder->local_image_id.empty() ||
state_builder->is_linked())) {
// both images doesn't exist or local image exists and is non-primary
// and linked to the missing remote image
finish(-ENOLINK);
} else {
finish(-ENOENT);
}
return;
} else if (r < 0) {
derr << "error preparing remote image for replay: " << cpp_strerror(r)
<< dendl;
finish(r);
return;
}
if (!state_builder->is_remote_primary()) {
ceph_assert(!state_builder->remote_image_id.empty());
if (state_builder->local_image_id.empty()) {
dout(10) << "local image does not exist and remote image is not primary"
<< dendl;
finish(-EREMOTEIO);
return;
} else if (!state_builder->is_linked()) {
dout(10) << "local image is unlinked and remote image is not primary"
<< dendl;
finish(-EREMOTEIO);
return;
}
// if the local image is linked to the remote image, we ignore that
// the remote image is not primary so that we can replay demotion
}
open_remote_image();
}
template <typename I>
void BootstrapRequest<I>::open_remote_image() {
ceph_assert(*m_state_builder != nullptr);
auto remote_image_id = (*m_state_builder)->remote_image_id;
dout(15) << "remote_image_id=" << remote_image_id << dendl;
update_progress("OPEN_REMOTE_IMAGE");
auto ctx = create_context_callback<
BootstrapRequest<I>,
&BootstrapRequest<I>::handle_open_remote_image>(this);
ceph_assert(*m_state_builder != nullptr);
OpenImageRequest<I> *request = OpenImageRequest<I>::create(
m_remote_io_ctx, &(*m_state_builder)->remote_image_ctx, remote_image_id,
false, ctx);
request->send();
}
template <typename I>
void BootstrapRequest<I>::handle_open_remote_image(int r) {
dout(15) << "r=" << r << dendl;
ceph_assert(*m_state_builder != nullptr);
if (r < 0) {
derr << "failed to open remote image: " << cpp_strerror(r) << dendl;
ceph_assert((*m_state_builder)->remote_image_ctx == nullptr);
finish(r);
return;
}
if ((*m_state_builder)->local_image_id.empty()) {
create_local_image();
return;
}
open_local_image();
}
template <typename I>
void BootstrapRequest<I>::open_local_image() {
ceph_assert(*m_state_builder != nullptr);
auto local_image_id = (*m_state_builder)->local_image_id;
dout(15) << "local_image_id=" << local_image_id << dendl;
update_progress("OPEN_LOCAL_IMAGE");
Context *ctx = create_context_callback<
BootstrapRequest<I>, &BootstrapRequest<I>::handle_open_local_image>(
this);
OpenLocalImageRequest<I> *request = OpenLocalImageRequest<I>::create(
m_local_io_ctx, &(*m_state_builder)->local_image_ctx, local_image_id,
m_threads->work_queue, ctx);
request->send();
}
template <typename I>
void BootstrapRequest<I>::handle_open_local_image(int r) {
dout(15) << "r=" << r << dendl;
ceph_assert(*m_state_builder != nullptr);
auto local_image_ctx = (*m_state_builder)->local_image_ctx;
ceph_assert((r >= 0 && local_image_ctx != nullptr) ||
(r < 0 && local_image_ctx == nullptr));
if (r == -ENOENT) {
dout(10) << "local image missing" << dendl;
create_local_image();
return;
} else if (r == -EREMOTEIO) {
dout(10) << "local image is primary -- skipping image replay" << dendl;
m_ret_val = r;
close_remote_image();
return;
} else if (r < 0) {
derr << "failed to open local image: " << cpp_strerror(r) << dendl;
m_ret_val = r;
close_remote_image();
return;
}
prepare_replay();
}
template <typename I>
void BootstrapRequest<I>::prepare_replay() {
dout(10) << dendl;
update_progress("PREPARE_REPLAY");
ceph_assert(*m_state_builder != nullptr);
auto ctx = create_context_callback<
BootstrapRequest<I>, &BootstrapRequest<I>::handle_prepare_replay>(this);
auto request = (*m_state_builder)->create_prepare_replay_request(
m_local_mirror_uuid, m_progress_ctx, m_do_resync, &m_syncing, ctx);
request->send();
}
template <typename I>
void BootstrapRequest<I>::handle_prepare_replay(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to prepare local replay: " << cpp_strerror(r) << dendl;
m_ret_val = r;
close_remote_image();
return;
} else if (*m_do_resync) {
dout(10) << "local image resync requested" << dendl;
close_remote_image();
return;
} else if ((*m_state_builder)->is_disconnected()) {
dout(10) << "client flagged disconnected -- skipping bootstrap" << dendl;
// The caller is expected to detect disconnect initializing remote journal.
m_ret_val = 0;
close_remote_image();
return;
} else if (m_syncing) {
dout(10) << "local image still syncing to remote image" << dendl;
image_sync();
return;
}
close_remote_image();
}
template <typename I>
void BootstrapRequest<I>::create_local_image() {
dout(10) << dendl;
update_progress("CREATE_LOCAL_IMAGE");
ceph_assert(*m_state_builder != nullptr);
auto ctx = create_context_callback<
BootstrapRequest<I>,
&BootstrapRequest<I>::handle_create_local_image>(this);
auto request = (*m_state_builder)->create_local_image_request(
m_threads, m_local_io_ctx, m_global_image_id, m_pool_meta_cache,
m_progress_ctx, ctx);
request->send();
}
template <typename I>
void BootstrapRequest<I>::handle_create_local_image(int r) {
dout(15) << "r=" << r << dendl;
if (r < 0) {
if (r == -ENOENT) {
dout(10) << "parent image does not exist" << dendl;
} else {
derr << "failed to create local image: " << cpp_strerror(r) << dendl;
}
m_ret_val = r;
close_remote_image();
return;
}
open_local_image();
}
template <typename I>
void BootstrapRequest<I>::image_sync() {
std::unique_lock locker{m_lock};
if (m_canceled) {
locker.unlock();
m_ret_val = -ECANCELED;
dout(10) << "request canceled" << dendl;
close_remote_image();
return;
}
dout(15) << dendl;
ceph_assert(m_image_sync == nullptr);
auto state_builder = *m_state_builder;
auto sync_point_handler = state_builder->create_sync_point_handler();
Context *ctx = create_context_callback<
BootstrapRequest<I>, &BootstrapRequest<I>::handle_image_sync>(this);
m_image_sync = ImageSync<I>::create(
m_threads, state_builder->local_image_ctx, state_builder->remote_image_ctx,
m_local_mirror_uuid, sync_point_handler, m_instance_watcher,
m_progress_ctx, ctx);
m_image_sync->get();
locker.unlock();
update_progress("IMAGE_SYNC");
m_image_sync->send();
}
template <typename I>
void BootstrapRequest<I>::handle_image_sync(int r) {
dout(15) << "r=" << r << dendl;
{
std::lock_guard locker{m_lock};
m_image_sync->put();
m_image_sync = nullptr;
(*m_state_builder)->destroy_sync_point_handler();
}
if (r < 0) {
if (r == -ECANCELED) {
dout(10) << "request canceled" << dendl;
} else {
derr << "failed to sync remote image: " << cpp_strerror(r) << dendl;
}
m_ret_val = r;
}
close_remote_image();
}
template <typename I>
void BootstrapRequest<I>::close_remote_image() {
if ((*m_state_builder)->replay_requires_remote_image()) {
finish(m_ret_val);
return;
}
dout(15) << dendl;
update_progress("CLOSE_REMOTE_IMAGE");
auto ctx = create_context_callback<
BootstrapRequest<I>,
&BootstrapRequest<I>::handle_close_remote_image>(this);
ceph_assert(*m_state_builder != nullptr);
(*m_state_builder)->close_remote_image(ctx);
}
template <typename I>
void BootstrapRequest<I>::handle_close_remote_image(int r) {
dout(15) << "r=" << r << dendl;
if (r < 0) {
derr << "error encountered closing remote image: " << cpp_strerror(r)
<< dendl;
}
finish(m_ret_val);
}
template <typename I>
void BootstrapRequest<I>::update_progress(const std::string &description) {
dout(15) << description << dendl;
if (m_progress_ctx) {
m_progress_ctx->update_progress(description);
}
}
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_replayer::BootstrapRequest<librbd::ImageCtx>;
| 14,292 | 28.409465 | 79 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/BootstrapRequest.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_REPLAYER_BOOTSTRAP_REQUEST_H
#define RBD_MIRROR_IMAGE_REPLAYER_BOOTSTRAP_REQUEST_H
#include "include/int_types.h"
#include "include/rados/librados.hpp"
#include "common/ceph_mutex.h"
#include "common/Timer.h"
#include "cls/rbd/cls_rbd_types.h"
#include "librbd/mirror/Types.h"
#include "tools/rbd_mirror/CancelableRequest.h"
#include "tools/rbd_mirror/Types.h"
#include <string>
class Context;
namespace journal { class CacheManagerHandler; }
namespace librbd { class ImageCtx; }
namespace rbd {
namespace mirror {
class ProgressContext;
template <typename> class ImageSync;
template <typename> class InstanceWatcher;
struct PoolMetaCache;
template <typename> struct Threads;
namespace image_replayer {
template <typename> class StateBuilder;
template <typename ImageCtxT = librbd::ImageCtx>
class BootstrapRequest : public CancelableRequest {
public:
typedef rbd::mirror::ProgressContext ProgressContext;
static BootstrapRequest* create(
Threads<ImageCtxT>* threads,
librados::IoCtx& local_io_ctx,
librados::IoCtx& remote_io_ctx,
InstanceWatcher<ImageCtxT>* instance_watcher,
const std::string& global_image_id,
const std::string& local_mirror_uuid,
const RemotePoolMeta& remote_pool_meta,
::journal::CacheManagerHandler* cache_manager_handler,
PoolMetaCache* pool_meta_cache,
ProgressContext* progress_ctx,
StateBuilder<ImageCtxT>** state_builder,
bool* do_resync,
Context* on_finish) {
return new BootstrapRequest(
threads, local_io_ctx, remote_io_ctx, instance_watcher, global_image_id,
local_mirror_uuid, remote_pool_meta, cache_manager_handler,
pool_meta_cache, progress_ctx, state_builder, do_resync, on_finish);
}
BootstrapRequest(
Threads<ImageCtxT>* threads,
librados::IoCtx& local_io_ctx,
librados::IoCtx& remote_io_ctx,
InstanceWatcher<ImageCtxT>* instance_watcher,
const std::string& global_image_id,
const std::string& local_mirror_uuid,
const RemotePoolMeta& remote_pool_meta,
::journal::CacheManagerHandler* cache_manager_handler,
PoolMetaCache* pool_meta_cache,
ProgressContext* progress_ctx,
StateBuilder<ImageCtxT>** state_builder,
bool* do_resync,
Context* on_finish);
bool is_syncing() const;
void send() override;
void cancel() override;
std::string get_local_image_name() const;
private:
/**
* @verbatim
*
* <start>
* |
* v (error)
* PREPARE_LOCAL_IMAGE * * * * * * * * * * * * * * * * * *
* | *
* v (error) *
* PREPARE_REMOTE_IMAGE * * * * * * * * * * * * * * * * * *
* | *
* v (error) *
* OPEN_REMOTE_IMAGE * * * * * * * * * * * * * * * * * * *
* | *
* | *
* \----> CREATE_LOCAL_IMAGE * * * * * * * * * * * * *
* | | ^ * *
* | | . * *
* | v . (image DNE) * *
* \----> OPEN_LOCAL_IMAGE * * * * * * * * * * * * * *
* | * *
* | * *
* v * *
* PREPARE_REPLAY * * * * * * * * * * * * * * *
* | * *
* | * *
* v (skip if not needed) * *
* IMAGE_SYNC * * * * * * * * * * * * * * * * *
* | * *
* | * *
* /---------/ * *
* | * *
* v * *
* CLOSE_REMOTE_IMAGE < * * * * * * * * * * * * * * * * *
* | *
* v *
* <finish> < * * * * * * * * * * * * * * * * * * * * * * *
*
* @endverbatim
*/
Threads<ImageCtxT>* m_threads;
librados::IoCtx &m_local_io_ctx;
librados::IoCtx &m_remote_io_ctx;
InstanceWatcher<ImageCtxT> *m_instance_watcher;
std::string m_global_image_id;
std::string m_local_mirror_uuid;
RemotePoolMeta m_remote_pool_meta;
::journal::CacheManagerHandler *m_cache_manager_handler;
PoolMetaCache* m_pool_meta_cache;
ProgressContext *m_progress_ctx;
StateBuilder<ImageCtxT>** m_state_builder;
bool *m_do_resync;
mutable ceph::mutex m_lock;
bool m_canceled = false;
int m_ret_val = 0;
std::string m_local_image_name;
std::string m_prepare_local_image_name;
bool m_syncing = false;
ImageSync<ImageCtxT> *m_image_sync = nullptr;
void prepare_local_image();
void handle_prepare_local_image(int r);
void prepare_remote_image();
void handle_prepare_remote_image(int r);
void open_remote_image();
void handle_open_remote_image(int r);
void open_local_image();
void handle_open_local_image(int r);
void create_local_image();
void handle_create_local_image(int r);
void prepare_replay();
void handle_prepare_replay(int r);
void image_sync();
void handle_image_sync(int r);
void close_remote_image();
void handle_close_remote_image(int r);
void update_progress(const std::string &description);
};
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_replayer::BootstrapRequest<librbd::ImageCtx>;
#endif // RBD_MIRROR_IMAGE_REPLAYER_BOOTSTRAP_REQUEST_H
| 6,112 | 32.587912 | 86 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/CloseImageRequest.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "CloseImageRequest.h"
#include "common/debug.h"
#include "common/errno.h"
#include "librbd/ImageCtx.h"
#include "librbd/ImageState.h"
#include "librbd/Utils.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_replayer::CloseImageRequest: " \
<< this << " " << __func__
namespace rbd {
namespace mirror {
namespace image_replayer {
using librbd::util::create_context_callback;
template <typename I>
CloseImageRequest<I>::CloseImageRequest(I **image_ctx, Context *on_finish)
: m_image_ctx(image_ctx), m_on_finish(on_finish) {
}
template <typename I>
void CloseImageRequest<I>::send() {
close_image();
}
template <typename I>
void CloseImageRequest<I>::close_image() {
dout(20) << dendl;
Context *ctx = create_context_callback<
CloseImageRequest<I>, &CloseImageRequest<I>::handle_close_image>(this);
(*m_image_ctx)->state->close(ctx);
}
template <typename I>
void CloseImageRequest<I>::handle_close_image(int r) {
dout(20) << ": r=" << r << dendl;
if (r < 0) {
derr << ": error encountered while closing image: " << cpp_strerror(r)
<< dendl;
}
*m_image_ctx = nullptr;
m_on_finish->complete(0);
delete this;
}
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_replayer::CloseImageRequest<librbd::ImageCtx>;
| 1,545 | 23.539683 | 82 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/CloseImageRequest.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_REPLAYER_CLOSE_IMAGE_REQUEST_H
#define RBD_MIRROR_IMAGE_REPLAYER_CLOSE_IMAGE_REQUEST_H
#include "include/int_types.h"
#include "librbd/ImageCtx.h"
#include <string>
class Context;
namespace librbd { class ImageCtx; }
namespace rbd {
namespace mirror {
namespace image_replayer {
template <typename ImageCtxT = librbd::ImageCtx>
class CloseImageRequest {
public:
static CloseImageRequest* create(ImageCtxT **image_ctx, Context *on_finish) {
return new CloseImageRequest(image_ctx, on_finish);
}
CloseImageRequest(ImageCtxT **image_ctx, Context *on_finish);
void send();
private:
/**
* @verbatim
*
* <start>
* |
* v
* CLOSE_IMAGE
* |
* v
* <finish>
*
* @endverbatim
*/
ImageCtxT **m_image_ctx;
Context *m_on_finish;
void close_image();
void handle_close_image(int r);
};
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_replayer::CloseImageRequest<librbd::ImageCtx>;
#endif // RBD_MIRROR_IMAGE_REPLAYER_CLOSE_IMAGE_REQUEST_H
| 1,191 | 19.912281 | 87 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/CreateImageRequest.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "CreateImageRequest.h"
#include "CloseImageRequest.h"
#include "OpenImageRequest.h"
#include "common/debug.h"
#include "common/errno.h"
#include "cls/rbd/cls_rbd_client.h"
#include "librbd/ImageCtx.h"
#include "librbd/ImageState.h"
#include "librbd/internal.h"
#include "librbd/Utils.h"
#include "librbd/asio/ContextWQ.h"
#include "librbd/image/CreateRequest.h"
#include "librbd/image/CloneRequest.h"
#include "tools/rbd_mirror/PoolMetaCache.h"
#include "tools/rbd_mirror/Types.h"
#include "tools/rbd_mirror/Threads.h"
#include "tools/rbd_mirror/image_replayer/Utils.h"
#include "tools/rbd_mirror/image_sync/Utils.h"
#include <boost/algorithm/string/predicate.hpp>
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_replayer::CreateImageRequest: " \
<< this << " " << __func__ << ": "
using librbd::util::create_async_context_callback;
using librbd::util::create_context_callback;
using librbd::util::create_rados_callback;
namespace rbd {
namespace mirror {
namespace image_replayer {
template <typename I>
CreateImageRequest<I>::CreateImageRequest(
Threads<I>* threads,
librados::IoCtx &local_io_ctx,
const std::string &global_image_id,
const std::string &remote_mirror_uuid,
const std::string &local_image_name,
const std::string &local_image_id,
I *remote_image_ctx,
PoolMetaCache* pool_meta_cache,
cls::rbd::MirrorImageMode mirror_image_mode,
Context *on_finish)
: m_threads(threads), m_local_io_ctx(local_io_ctx),
m_global_image_id(global_image_id),
m_remote_mirror_uuid(remote_mirror_uuid),
m_local_image_name(local_image_name), m_local_image_id(local_image_id),
m_remote_image_ctx(remote_image_ctx),
m_pool_meta_cache(pool_meta_cache),
m_mirror_image_mode(mirror_image_mode), m_on_finish(on_finish) {
}
template <typename I>
void CreateImageRequest<I>::send() {
int r = validate_parent();
if (r < 0) {
error(r);
return;
}
if (m_remote_parent_spec.pool_id == -1) {
create_image();
} else {
get_parent_global_image_id();
}
}
template <typename I>
void CreateImageRequest<I>::create_image() {
dout(10) << dendl;
using klass = CreateImageRequest<I>;
Context *ctx = create_context_callback<
klass, &klass::handle_create_image>(this);
std::shared_lock image_locker{m_remote_image_ctx->image_lock};
auto& config{
reinterpret_cast<CephContext*>(m_local_io_ctx.cct())->_conf};
librbd::ImageOptions image_options;
populate_image_options(&image_options);
auto req = librbd::image::CreateRequest<I>::create(
config, m_local_io_ctx, m_local_image_name, m_local_image_id,
m_remote_image_ctx->size, image_options, 0U, m_mirror_image_mode,
m_global_image_id, m_remote_mirror_uuid, m_remote_image_ctx->op_work_queue,
ctx);
req->send();
}
template <typename I>
void CreateImageRequest<I>::handle_create_image(int r) {
dout(10) << "r=" << r << dendl;
if (r == -EBADF) {
dout(5) << "image id " << m_local_image_id << " already in-use" << dendl;
finish(r);
return;
} else if (r < 0) {
derr << "failed to create local image: " << cpp_strerror(r) << dendl;
finish(r);
return;
}
finish(0);
}
template <typename I>
void CreateImageRequest<I>::get_parent_global_image_id() {
dout(10) << dendl;
librados::ObjectReadOperation op;
librbd::cls_client::mirror_image_get_start(&op,
m_remote_parent_spec.image_id);
librados::AioCompletion *aio_comp = create_rados_callback<
CreateImageRequest<I>,
&CreateImageRequest<I>::handle_get_parent_global_image_id>(this);
m_out_bl.clear();
int r = m_remote_parent_io_ctx.aio_operate(RBD_MIRRORING, aio_comp, &op,
&m_out_bl);
ceph_assert(r == 0);
aio_comp->release();
}
template <typename I>
void CreateImageRequest<I>::handle_get_parent_global_image_id(int r) {
dout(10) << "r=" << r << dendl;
if (r == 0) {
cls::rbd::MirrorImage mirror_image;
auto iter = m_out_bl.cbegin();
r = librbd::cls_client::mirror_image_get_finish(&iter, &mirror_image);
if (r == 0) {
m_parent_global_image_id = mirror_image.global_image_id;
dout(15) << "parent_global_image_id=" << m_parent_global_image_id
<< dendl;
}
}
if (r == -ENOENT) {
dout(10) << "parent image " << m_remote_parent_spec.image_id
<< " not mirrored" << dendl;
finish(r);
return;
} else if (r < 0) {
derr << "failed to retrieve global image id for parent image "
<< m_remote_parent_spec.image_id << ": " << cpp_strerror(r) << dendl;
finish(r);
return;
}
get_local_parent_image_id();
}
template <typename I>
void CreateImageRequest<I>::get_local_parent_image_id() {
dout(10) << dendl;
librados::ObjectReadOperation op;
librbd::cls_client::mirror_image_get_image_id_start(
&op, m_parent_global_image_id);
librados::AioCompletion *aio_comp = create_rados_callback<
CreateImageRequest<I>,
&CreateImageRequest<I>::handle_get_local_parent_image_id>(this);
m_out_bl.clear();
int r = m_local_parent_io_ctx.aio_operate(RBD_MIRRORING, aio_comp, &op,
&m_out_bl);
ceph_assert(r == 0);
aio_comp->release();
}
template <typename I>
void CreateImageRequest<I>::handle_get_local_parent_image_id(int r) {
dout(10) << "r=" << r << dendl;
if (r == 0) {
auto iter = m_out_bl.cbegin();
r = librbd::cls_client::mirror_image_get_image_id_finish(
&iter, &m_local_parent_spec.image_id);
}
if (r == -ENOENT) {
dout(10) << "parent image " << m_parent_global_image_id << " not "
<< "registered locally" << dendl;
finish(r);
return;
} else if (r < 0) {
derr << "failed to retrieve local image id for parent image "
<< m_parent_global_image_id << ": " << cpp_strerror(r) << dendl;
finish(r);
return;
}
open_remote_parent_image();
}
template <typename I>
void CreateImageRequest<I>::open_remote_parent_image() {
dout(10) << dendl;
Context *ctx = create_context_callback<
CreateImageRequest<I>,
&CreateImageRequest<I>::handle_open_remote_parent_image>(this);
OpenImageRequest<I> *request = OpenImageRequest<I>::create(
m_remote_parent_io_ctx, &m_remote_parent_image_ctx,
m_remote_parent_spec.image_id, true, ctx);
request->send();
}
template <typename I>
void CreateImageRequest<I>::handle_open_remote_parent_image(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to open remote parent image " << m_parent_pool_name << "/"
<< m_remote_parent_spec.image_id << dendl;
finish(r);
return;
}
clone_image();
}
template <typename I>
void CreateImageRequest<I>::clone_image() {
dout(10) << dendl;
LocalPoolMeta local_parent_pool_meta;
int r = m_pool_meta_cache->get_local_pool_meta(
m_local_parent_io_ctx.get_id(), &local_parent_pool_meta);
if (r < 0) {
derr << "failed to retrieve local parent mirror uuid for pool "
<< m_local_parent_io_ctx.get_id() << dendl;
m_ret_val = r;
close_remote_parent_image();
return;
}
// ensure no image sync snapshots for the local cluster exist in the
// remote image
bool found_parent_snap = false;
bool found_image_sync_snap = false;
std::string snap_name;
cls::rbd::SnapshotNamespace snap_namespace;
{
auto snap_prefix = image_sync::util::get_snapshot_name_prefix(
local_parent_pool_meta.mirror_uuid);
std::shared_lock remote_image_locker(m_remote_parent_image_ctx->image_lock);
for (auto snap_info : m_remote_parent_image_ctx->snap_info) {
if (snap_info.first == m_remote_parent_spec.snap_id) {
found_parent_snap = true;
snap_name = snap_info.second.name;
snap_namespace = snap_info.second.snap_namespace;
} else if (boost::starts_with(snap_info.second.name, snap_prefix)) {
found_image_sync_snap = true;
}
}
}
if (!found_parent_snap) {
dout(15) << "remote parent image snapshot not found" << dendl;
m_ret_val = -ENOENT;
close_remote_parent_image();
return;
} else if (found_image_sync_snap) {
dout(15) << "parent image not synced to local cluster" << dendl;
m_ret_val = -ENOENT;
close_remote_parent_image();
return;
}
librbd::ImageOptions opts;
populate_image_options(&opts);
auto& config{
reinterpret_cast<CephContext*>(m_local_io_ctx.cct())->_conf};
using klass = CreateImageRequest<I>;
Context *ctx = create_context_callback<
klass, &klass::handle_clone_image>(this);
librbd::image::CloneRequest<I> *req = librbd::image::CloneRequest<I>::create(
config, m_local_parent_io_ctx, m_local_parent_spec.image_id, snap_name,
snap_namespace, CEPH_NOSNAP, m_local_io_ctx, m_local_image_name,
m_local_image_id, opts, m_mirror_image_mode, m_global_image_id,
m_remote_mirror_uuid, m_remote_image_ctx->op_work_queue, ctx);
req->send();
}
template <typename I>
void CreateImageRequest<I>::handle_clone_image(int r) {
dout(10) << "r=" << r << dendl;
if (r == -EBADF) {
dout(5) << "image id " << m_local_image_id << " already in-use" << dendl;
m_ret_val = r;
} else if (r < 0) {
derr << "failed to clone image " << m_parent_pool_name << "/"
<< m_remote_parent_spec.image_id << " to "
<< m_local_image_name << dendl;
m_ret_val = r;
}
close_remote_parent_image();
}
template <typename I>
void CreateImageRequest<I>::close_remote_parent_image() {
dout(10) << dendl;
Context *ctx = create_context_callback<
CreateImageRequest<I>,
&CreateImageRequest<I>::handle_close_remote_parent_image>(this);
CloseImageRequest<I> *request = CloseImageRequest<I>::create(
&m_remote_parent_image_ctx, ctx);
request->send();
}
template <typename I>
void CreateImageRequest<I>::handle_close_remote_parent_image(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "error encountered closing remote parent image: "
<< cpp_strerror(r) << dendl;
}
finish(m_ret_val);
}
template <typename I>
void CreateImageRequest<I>::error(int r) {
dout(10) << "r=" << r << dendl;
m_threads->work_queue->queue(create_context_callback<
CreateImageRequest<I>, &CreateImageRequest<I>::finish>(this), r);
}
template <typename I>
void CreateImageRequest<I>::finish(int r) {
dout(10) << "r=" << r << dendl;
m_on_finish->complete(r);
delete this;
}
template <typename I>
int CreateImageRequest<I>::validate_parent() {
std::shared_lock owner_locker{m_remote_image_ctx->owner_lock};
std::shared_lock image_locker{m_remote_image_ctx->image_lock};
m_remote_parent_spec = m_remote_image_ctx->parent_md.spec;
// scan all remote snapshots for a linked parent
for (auto &snap_info_pair : m_remote_image_ctx->snap_info) {
auto &parent_spec = snap_info_pair.second.parent.spec;
if (parent_spec.pool_id == -1) {
continue;
} else if (m_remote_parent_spec.pool_id == -1) {
m_remote_parent_spec = parent_spec;
continue;
}
if (m_remote_parent_spec != parent_spec) {
derr << "remote image parent spec mismatch" << dendl;
return -EINVAL;
}
}
if (m_remote_parent_spec.pool_id == -1) {
return 0;
}
// map remote parent pool to local parent pool
int r = librbd::util::create_ioctx(
m_remote_image_ctx->md_ctx, "remote parent pool",
m_remote_parent_spec.pool_id, m_remote_parent_spec.pool_namespace,
&m_remote_parent_io_ctx);
if (r < 0) {
derr << "failed to open remote parent pool " << m_remote_parent_spec.pool_id
<< ": " << cpp_strerror(r) << dendl;
return r;
}
m_parent_pool_name = m_remote_parent_io_ctx.get_pool_name();
librados::Rados local_rados(m_local_io_ctx);
r = local_rados.ioctx_create(m_parent_pool_name.c_str(),
m_local_parent_io_ctx);
if (r < 0) {
derr << "failed to open local parent pool " << m_parent_pool_name << ": "
<< cpp_strerror(r) << dendl;
return r;
}
m_local_parent_io_ctx.set_namespace(m_remote_parent_io_ctx.get_namespace());
return 0;
}
template <typename I>
void CreateImageRequest<I>::populate_image_options(
librbd::ImageOptions* image_options) {
image_options->set(RBD_IMAGE_OPTION_FEATURES,
m_remote_image_ctx->features);
image_options->set(RBD_IMAGE_OPTION_ORDER, m_remote_image_ctx->order);
image_options->set(RBD_IMAGE_OPTION_STRIPE_UNIT,
m_remote_image_ctx->stripe_unit);
image_options->set(RBD_IMAGE_OPTION_STRIPE_COUNT,
m_remote_image_ctx->stripe_count);
// Determine the data pool for the local image as follows:
// 1. If the local pool has a default data pool, use it.
// 2. If the remote image has a data pool different from its metadata pool and
// a pool with the same name exists locally, use it.
// 3. Don't set the data pool explicitly.
std::string data_pool;
librados::Rados local_rados(m_local_io_ctx);
auto default_data_pool = g_ceph_context->_conf.get_val<std::string>("rbd_default_data_pool");
auto remote_md_pool = m_remote_image_ctx->md_ctx.get_pool_name();
auto remote_data_pool = m_remote_image_ctx->data_ctx.get_pool_name();
if (default_data_pool != "") {
data_pool = default_data_pool;
} else if (remote_data_pool != remote_md_pool) {
if (local_rados.pool_lookup(remote_data_pool.c_str()) >= 0) {
data_pool = remote_data_pool;
}
}
if (data_pool != "") {
image_options->set(RBD_IMAGE_OPTION_DATA_POOL, data_pool);
}
if (m_remote_parent_spec.pool_id != -1) {
uint64_t clone_format = 1;
if (m_remote_image_ctx->test_op_features(
RBD_OPERATION_FEATURE_CLONE_CHILD)) {
clone_format = 2;
}
image_options->set(RBD_IMAGE_OPTION_CLONE_FORMAT, clone_format);
}
}
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_replayer::CreateImageRequest<librbd::ImageCtx>;
| 14,206 | 30.431416 | 95 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/CreateImageRequest.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_REPLAYER_CREATE_IMAGE_REQUEST_H
#define RBD_MIRROR_IMAGE_REPLAYER_CREATE_IMAGE_REQUEST_H
#include "include/int_types.h"
#include "include/types.h"
#include "include/rados/librados.hpp"
#include "cls/rbd/cls_rbd_types.h"
#include "librbd/Types.h"
#include <string>
class Context;
namespace librbd { class ImageCtx; }
namespace librbd { class ImageOptions; }
namespace rbd {
namespace mirror {
class PoolMetaCache;
template <typename> struct Threads;
namespace image_replayer {
template <typename ImageCtxT = librbd::ImageCtx>
class CreateImageRequest {
public:
static CreateImageRequest *create(
Threads<ImageCtxT> *threads,
librados::IoCtx &local_io_ctx,
const std::string &global_image_id,
const std::string &remote_mirror_uuid,
const std::string &local_image_name,
const std::string &local_image_id,
ImageCtxT *remote_image_ctx,
PoolMetaCache* pool_meta_cache,
cls::rbd::MirrorImageMode mirror_image_mode,
Context *on_finish) {
return new CreateImageRequest(threads, local_io_ctx, global_image_id,
remote_mirror_uuid, local_image_name,
local_image_id, remote_image_ctx,
pool_meta_cache, mirror_image_mode,
on_finish);
}
CreateImageRequest(
Threads<ImageCtxT> *threads, librados::IoCtx &local_io_ctx,
const std::string &global_image_id,
const std::string &remote_mirror_uuid,
const std::string &local_image_name,
const std::string &local_image_id,
ImageCtxT *remote_image_ctx,
PoolMetaCache* pool_meta_cache,
cls::rbd::MirrorImageMode mirror_image_mode,
Context *on_finish);
void send();
private:
/**
* @verbatim
*
* <start> * * * * * * * * * * * * * * * * * * * * * * * * * * * *
* | *
* | (non-clone) *
* |\------------> CREATE_IMAGE ---------------------\ * (error)
* | | *
* | (clone) | *
* \-------------> GET_PARENT_GLOBAL_IMAGE_ID * * * | * * * *
* | | * *
* v | *
* GET_LOCAL_PARENT_IMAGE_ID * * * * | * * * *
* | | * *
* v | *
* OPEN_REMOTE_PARENT * * * * * * * | * * * *
* | | * *
* v | *
* CLONE_IMAGE | *
* | | *
* v | *
* CLOSE_REMOTE_PARENT | *
* | v *
* \------------------------> <finish> < * *
* @endverbatim
*/
Threads<ImageCtxT> *m_threads;
librados::IoCtx &m_local_io_ctx;
std::string m_global_image_id;
std::string m_remote_mirror_uuid;
std::string m_local_image_name;
std::string m_local_image_id;
ImageCtxT *m_remote_image_ctx;
PoolMetaCache* m_pool_meta_cache;
cls::rbd::MirrorImageMode m_mirror_image_mode;
Context *m_on_finish;
librados::IoCtx m_remote_parent_io_ctx;
ImageCtxT *m_remote_parent_image_ctx = nullptr;
cls::rbd::ParentImageSpec m_remote_parent_spec;
librados::IoCtx m_local_parent_io_ctx;
cls::rbd::ParentImageSpec m_local_parent_spec;
bufferlist m_out_bl;
std::string m_parent_global_image_id;
std::string m_parent_pool_name;
int m_ret_val = 0;
void create_image();
void handle_create_image(int r);
void get_parent_global_image_id();
void handle_get_parent_global_image_id(int r);
void get_local_parent_image_id();
void handle_get_local_parent_image_id(int r);
void open_remote_parent_image();
void handle_open_remote_parent_image(int r);
void clone_image();
void handle_clone_image(int r);
void close_remote_parent_image();
void handle_close_remote_parent_image(int r);
void error(int r);
void finish(int r);
int validate_parent();
void populate_image_options(librbd::ImageOptions* image_options);
};
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_replayer::CreateImageRequest<librbd::ImageCtx>;
#endif // RBD_MIRROR_IMAGE_REPLAYER_CREATE_IMAGE_REQUEST_H
| 4,941 | 33.082759 | 88 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/GetMirrorImageIdRequest.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "tools/rbd_mirror/image_replayer/GetMirrorImageIdRequest.h"
#include "include/rados/librados.hpp"
#include "cls/rbd/cls_rbd_client.h"
#include "common/debug.h"
#include "common/errno.h"
#include "librbd/ImageCtx.h"
#include "librbd/Utils.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_replayer::" \
<< "GetMirrorImageIdRequest: " << this << " " \
<< __func__ << ": "
namespace rbd {
namespace mirror {
namespace image_replayer {
using librbd::util::create_rados_callback;
template <typename I>
void GetMirrorImageIdRequest<I>::send() {
dout(20) << dendl;
get_image_id();
}
template <typename I>
void GetMirrorImageIdRequest<I>::get_image_id() {
dout(20) << dendl;
// attempt to cross-reference a image id by the global image id
librados::ObjectReadOperation op;
librbd::cls_client::mirror_image_get_image_id_start(&op, m_global_image_id);
librados::AioCompletion *aio_comp = create_rados_callback<
GetMirrorImageIdRequest<I>,
&GetMirrorImageIdRequest<I>::handle_get_image_id>(
this);
int r = m_io_ctx.aio_operate(RBD_MIRRORING, aio_comp, &op, &m_out_bl);
ceph_assert(r == 0);
aio_comp->release();
}
template <typename I>
void GetMirrorImageIdRequest<I>::handle_get_image_id(int r) {
if (r == 0) {
auto iter = m_out_bl.cbegin();
r = librbd::cls_client::mirror_image_get_image_id_finish(
&iter, m_image_id);
}
dout(20) << "r=" << r << ", "
<< "image_id=" << *m_image_id << dendl;
if (r < 0) {
if (r == -ENOENT) {
dout(10) << "global image " << m_global_image_id << " not registered"
<< dendl;
} else {
derr << "failed to retrieve image id: " << cpp_strerror(r) << dendl;
}
finish(r);
return;
}
finish(0);
}
template <typename I>
void GetMirrorImageIdRequest<I>::finish(int r) {
dout(20) << "r=" << r << dendl;
m_on_finish->complete(r);
delete this;
}
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_replayer::GetMirrorImageIdRequest<librbd::ImageCtx>;
| 2,312 | 25.895349 | 86 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/GetMirrorImageIdRequest.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_REPLAYER_GET_MIRROR_IMAGE_ID_REQUEST_H
#define RBD_MIRROR_IMAGE_REPLAYER_GET_MIRROR_IMAGE_ID_REQUEST_H
#include "include/buffer.h"
#include "include/rados/librados_fwd.hpp"
#include <string>
namespace librbd { struct ImageCtx; }
struct Context;
namespace rbd {
namespace mirror {
namespace image_replayer {
template <typename ImageCtxT = librbd::ImageCtx>
class GetMirrorImageIdRequest {
public:
static GetMirrorImageIdRequest *create(librados::IoCtx &io_ctx,
const std::string &global_image_id,
std::string *image_id,
Context *on_finish) {
return new GetMirrorImageIdRequest(io_ctx, global_image_id, image_id,
on_finish);
}
GetMirrorImageIdRequest(librados::IoCtx &io_ctx,
const std::string &global_image_id,
std::string *image_id,
Context *on_finish)
: m_io_ctx(io_ctx), m_global_image_id(global_image_id),
m_image_id(image_id), m_on_finish(on_finish) {
}
void send();
private:
/**
* @verbatim
*
* <start>
* |
* v
* GET_IMAGE_ID
* |
* v
* <finish>
* @endverbatim
*/
librados::IoCtx &m_io_ctx;
std::string m_global_image_id;
std::string *m_image_id;
Context *m_on_finish;
bufferlist m_out_bl;
void get_image_id();
void handle_get_image_id(int r);
void finish(int r);
};
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_replayer::GetMirrorImageIdRequest<librbd::ImageCtx>;
#endif // RBD_MIRROR_IMAGE_REPLAYER_GET_MIRROR_IMAGE_ID_REQUEST_H
| 1,873 | 23.657895 | 93 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/OpenImageRequest.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "OpenImageRequest.h"
#include "common/debug.h"
#include "common/errno.h"
#include "librbd/ImageCtx.h"
#include "librbd/ImageState.h"
#include "librbd/Utils.h"
#include <type_traits>
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_replayer::OpenImageRequest: " \
<< this << " " << __func__ << " "
namespace rbd {
namespace mirror {
namespace image_replayer {
using librbd::util::create_context_callback;
template <typename I>
OpenImageRequest<I>::OpenImageRequest(librados::IoCtx &io_ctx, I **image_ctx,
const std::string &image_id,
bool read_only, Context *on_finish)
: m_io_ctx(io_ctx), m_image_ctx(image_ctx), m_image_id(image_id),
m_read_only(read_only), m_on_finish(on_finish) {
}
template <typename I>
void OpenImageRequest<I>::send() {
send_open_image();
}
template <typename I>
void OpenImageRequest<I>::send_open_image() {
dout(20) << dendl;
*m_image_ctx = I::create("", m_image_id, nullptr, m_io_ctx, m_read_only);
if (!m_read_only) {
// ensure non-primary images can be modified
(*m_image_ctx)->read_only_mask = ~librbd::IMAGE_READ_ONLY_FLAG_NON_PRIMARY;
}
Context *ctx = create_context_callback<
OpenImageRequest<I>, &OpenImageRequest<I>::handle_open_image>(
this);
(*m_image_ctx)->state->open(0, ctx);
}
template <typename I>
void OpenImageRequest<I>::handle_open_image(int r) {
dout(20) << ": r=" << r << dendl;
if (r < 0) {
derr << ": failed to open image '" << m_image_id << "': "
<< cpp_strerror(r) << dendl;
*m_image_ctx = nullptr;
}
finish(r);
}
template <typename I>
void OpenImageRequest<I>::finish(int r) {
dout(20) << ": r=" << r << dendl;
m_on_finish->complete(r);
delete this;
}
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_replayer::OpenImageRequest<librbd::ImageCtx>;
| 2,150 | 25.8875 | 81 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/OpenImageRequest.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_REPLAYER_OPEN_IMAGE_REQUEST_H
#define RBD_MIRROR_IMAGE_REPLAYER_OPEN_IMAGE_REQUEST_H
#include "include/int_types.h"
#include "librbd/ImageCtx.h"
#include <string>
class Context;
namespace librbd { class ImageCtx; }
namespace rbd {
namespace mirror {
namespace image_replayer {
template <typename ImageCtxT = librbd::ImageCtx>
class OpenImageRequest {
public:
static OpenImageRequest* create(librados::IoCtx &io_ctx,
ImageCtxT **image_ctx,
const std::string &image_id,
bool read_only, Context *on_finish) {
return new OpenImageRequest(io_ctx, image_ctx, image_id, read_only,
on_finish);
}
OpenImageRequest(librados::IoCtx &io_ctx, ImageCtxT **image_ctx,
const std::string &image_id, bool read_only,
Context *on_finish);
void send();
private:
/**
* @verbatim
*
* <start>
* |
* v
* OPEN_IMAGE
* |
* v
* <finish>
*
* @endverbatim
*/
librados::IoCtx &m_io_ctx;
ImageCtxT **m_image_ctx;
std::string m_image_id;
bool m_read_only;
Context *m_on_finish;
void send_open_image();
void handle_open_image(int r);
void send_close_image(int r);
void handle_close_image(int r);
void finish(int r);
};
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_replayer::OpenImageRequest<librbd::ImageCtx>;
#endif // RBD_MIRROR_IMAGE_REPLAYER_OPEN_IMAGE_REQUEST_H
| 1,692 | 22.513889 | 86 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/OpenLocalImageRequest.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "include/compat.h"
#include "CloseImageRequest.h"
#include "OpenLocalImageRequest.h"
#include "common/debug.h"
#include "common/errno.h"
#include "librbd/ExclusiveLock.h"
#include "librbd/ImageCtx.h"
#include "librbd/ImageState.h"
#include "librbd/Journal.h"
#include "librbd/Utils.h"
#include "librbd/asio/ContextWQ.h"
#include "librbd/exclusive_lock/Policy.h"
#include "librbd/journal/Policy.h"
#include "librbd/mirror/GetInfoRequest.h"
#include <type_traits>
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_replayer::OpenLocalImageRequest: " \
<< this << " " << __func__ << " "
namespace rbd {
namespace mirror {
namespace image_replayer {
using librbd::util::create_context_callback;
namespace {
template <typename I>
struct MirrorExclusiveLockPolicy : public librbd::exclusive_lock::Policy {
I *image_ctx;
MirrorExclusiveLockPolicy(I *image_ctx) : image_ctx(image_ctx) {
}
bool may_auto_request_lock() override {
return false;
}
int lock_requested(bool force) override {
int r = -EROFS;
{
std::shared_lock owner_locker{image_ctx->owner_lock};
std::shared_lock image_locker{image_ctx->image_lock};
if (image_ctx->journal == nullptr || image_ctx->journal->is_tag_owner()) {
r = 0;
}
}
if (r == 0) {
// if the local image journal has been closed or if it was (force)
// promoted allow the lock to be released to another client
image_ctx->exclusive_lock->release_lock(nullptr);
}
return r;
}
bool accept_blocked_request(
librbd::exclusive_lock::OperationRequestType request_type) override {
switch (request_type) {
case librbd::exclusive_lock::OPERATION_REQUEST_TYPE_TRASH_SNAP_REMOVE:
case librbd::exclusive_lock::OPERATION_REQUEST_TYPE_FORCE_PROMOTION:
return true;
default:
return false;
}
}
};
struct MirrorJournalPolicy : public librbd::journal::Policy {
librbd::asio::ContextWQ *work_queue;
MirrorJournalPolicy(librbd::asio::ContextWQ *work_queue)
: work_queue(work_queue) {
}
bool append_disabled() const override {
// avoid recording any events to the local journal
return true;
}
bool journal_disabled() const override {
return false;
}
void allocate_tag_on_lock(Context *on_finish) override {
// rbd-mirror will manually create tags by copying them from the peer
work_queue->queue(on_finish, 0);
}
};
} // anonymous namespace
template <typename I>
OpenLocalImageRequest<I>::OpenLocalImageRequest(
librados::IoCtx &local_io_ctx,
I **local_image_ctx,
const std::string &local_image_id,
librbd::asio::ContextWQ *work_queue,
Context *on_finish)
: m_local_io_ctx(local_io_ctx), m_local_image_ctx(local_image_ctx),
m_local_image_id(local_image_id), m_work_queue(work_queue),
m_on_finish(on_finish) {
}
template <typename I>
void OpenLocalImageRequest<I>::send() {
send_open_image();
}
template <typename I>
void OpenLocalImageRequest<I>::send_open_image() {
dout(20) << dendl;
*m_local_image_ctx = I::create("", m_local_image_id, nullptr,
m_local_io_ctx, false);
// ensure non-primary images can be modified
(*m_local_image_ctx)->read_only_mask =
~librbd::IMAGE_READ_ONLY_FLAG_NON_PRIMARY;
{
std::scoped_lock locker{(*m_local_image_ctx)->owner_lock,
(*m_local_image_ctx)->image_lock};
(*m_local_image_ctx)->set_exclusive_lock_policy(
new MirrorExclusiveLockPolicy<I>(*m_local_image_ctx));
(*m_local_image_ctx)->set_journal_policy(
new MirrorJournalPolicy(m_work_queue));
}
Context *ctx = create_context_callback<
OpenLocalImageRequest<I>, &OpenLocalImageRequest<I>::handle_open_image>(
this);
(*m_local_image_ctx)->state->open(0, ctx);
}
template <typename I>
void OpenLocalImageRequest<I>::handle_open_image(int r) {
dout(20) << ": r=" << r << dendl;
if (r < 0) {
if (r == -ENOENT) {
dout(10) << ": local image does not exist" << dendl;
} else {
derr << ": failed to open image '" << m_local_image_id << "': "
<< cpp_strerror(r) << dendl;
}
*m_local_image_ctx = nullptr;
finish(r);
return;
}
send_get_mirror_info();
}
template <typename I>
void OpenLocalImageRequest<I>::send_get_mirror_info() {
dout(20) << dendl;
Context *ctx = create_context_callback<
OpenLocalImageRequest<I>,
&OpenLocalImageRequest<I>::handle_get_mirror_info>(
this);
auto request = librbd::mirror::GetInfoRequest<I>::create(
**m_local_image_ctx, &m_mirror_image, &m_promotion_state,
&m_primary_mirror_uuid, ctx);
request->send();
}
template <typename I>
void OpenLocalImageRequest<I>::handle_get_mirror_info(int r) {
dout(20) << ": r=" << r << dendl;
if (r == -ENOENT) {
dout(5) << ": local image is not mirrored" << dendl;
send_close_image(r);
return;
} else if (r < 0) {
derr << ": error querying local image primary status: " << cpp_strerror(r)
<< dendl;
send_close_image(r);
return;
}
if (m_mirror_image.state == cls::rbd::MIRROR_IMAGE_STATE_DISABLING) {
dout(5) << ": local image mirroring is being disabled" << dendl;
send_close_image(-ENOENT);
return;
}
// if the local image owns the tag -- don't steal the lock since
// we aren't going to mirror peer data into this image anyway
if (m_promotion_state == librbd::mirror::PROMOTION_STATE_PRIMARY) {
dout(10) << ": local image is primary -- skipping image replay" << dendl;
send_close_image(-EREMOTEIO);
return;
}
send_lock_image();
}
template <typename I>
void OpenLocalImageRequest<I>::send_lock_image() {
std::shared_lock owner_locker{(*m_local_image_ctx)->owner_lock};
if ((*m_local_image_ctx)->exclusive_lock == nullptr) {
owner_locker.unlock();
if (m_mirror_image.mode == cls::rbd::MIRROR_IMAGE_MODE_SNAPSHOT) {
finish(0);
} else {
derr << ": image does not support exclusive lock" << dendl;
send_close_image(-EINVAL);
}
return;
}
dout(20) << dendl;
// disallow any proxied maintenance operations before grabbing lock
(*m_local_image_ctx)->exclusive_lock->block_requests(-EROFS);
Context *ctx = create_context_callback<
OpenLocalImageRequest<I>, &OpenLocalImageRequest<I>::handle_lock_image>(
this);
(*m_local_image_ctx)->exclusive_lock->acquire_lock(ctx);
}
template <typename I>
void OpenLocalImageRequest<I>::handle_lock_image(int r) {
dout(20) << ": r=" << r << dendl;
if (r < 0) {
derr << ": failed to lock image '" << m_local_image_id << "': "
<< cpp_strerror(r) << dendl;
send_close_image(r);
return;
}
{
std::shared_lock owner_locker{(*m_local_image_ctx)->owner_lock};
if ((*m_local_image_ctx)->exclusive_lock == nullptr ||
!(*m_local_image_ctx)->exclusive_lock->is_lock_owner()) {
derr << ": image is not locked" << dendl;
send_close_image(-EBUSY);
return;
}
}
finish(0);
}
template <typename I>
void OpenLocalImageRequest<I>::send_close_image(int r) {
dout(20) << dendl;
if (m_ret_val == 0 && r < 0) {
m_ret_val = r;
}
Context *ctx = create_context_callback<
OpenLocalImageRequest<I>, &OpenLocalImageRequest<I>::handle_close_image>(
this);
CloseImageRequest<I> *request = CloseImageRequest<I>::create(
m_local_image_ctx, ctx);
request->send();
}
template <typename I>
void OpenLocalImageRequest<I>::handle_close_image(int r) {
dout(20) << dendl;
ceph_assert(r == 0);
finish(m_ret_val);
}
template <typename I>
void OpenLocalImageRequest<I>::finish(int r) {
dout(20) << ": r=" << r << dendl;
m_on_finish->complete(r);
delete this;
}
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_replayer::OpenLocalImageRequest<librbd::ImageCtx>;
| 8,058 | 26.505119 | 86 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/OpenLocalImageRequest.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_REPLAYER_OPEN_LOCAL_IMAGE_REQUEST_H
#define RBD_MIRROR_IMAGE_REPLAYER_OPEN_LOCAL_IMAGE_REQUEST_H
#include "include/int_types.h"
#include "cls/rbd/cls_rbd_types.h"
#include "librbd/ImageCtx.h"
#include "librbd/mirror/Types.h"
#include <string>
class Context;
namespace librbd {
class ImageCtx;
namespace asio { struct ContextWQ; }
} // namespace librbd
namespace rbd {
namespace mirror {
namespace image_replayer {
template <typename ImageCtxT = librbd::ImageCtx>
class OpenLocalImageRequest {
public:
static OpenLocalImageRequest* create(librados::IoCtx &local_io_ctx,
ImageCtxT **local_image_ctx,
const std::string &local_image_id,
librbd::asio::ContextWQ *work_queue,
Context *on_finish) {
return new OpenLocalImageRequest(local_io_ctx, local_image_ctx,
local_image_id, work_queue, on_finish);
}
OpenLocalImageRequest(librados::IoCtx &local_io_ctx,
ImageCtxT **local_image_ctx,
const std::string &local_image_id,
librbd::asio::ContextWQ *work_queue,
Context *on_finish);
void send();
private:
/**
* @verbatim
*
* <start>
* |
* v
* OPEN_IMAGE * * * * * * * *
* | *
* v *
* GET_MIRROR_INFO * * * * *
* | *
* v (skip if primary) v
* LOCK_IMAGE * * * > CLOSE_IMAGE
* | |
* v |
* <finish> <---------------/
*
* @endverbatim
*/
librados::IoCtx &m_local_io_ctx;
ImageCtxT **m_local_image_ctx;
std::string m_local_image_id;
librbd::asio::ContextWQ *m_work_queue;
Context *m_on_finish;
cls::rbd::MirrorImage m_mirror_image;
librbd::mirror::PromotionState m_promotion_state =
librbd::mirror::PROMOTION_STATE_NON_PRIMARY;
std::string m_primary_mirror_uuid;
int m_ret_val = 0;
void send_open_image();
void handle_open_image(int r);
void send_get_mirror_info();
void handle_get_mirror_info(int r);
void send_lock_image();
void handle_lock_image(int r);
void send_close_image(int r);
void handle_close_image(int r);
void finish(int r);
};
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_replayer::OpenLocalImageRequest<librbd::ImageCtx>;
#endif // RBD_MIRROR_IMAGE_REPLAYER_OPEN_LOCAL_IMAGE_REQUEST_H
| 2,718 | 26.744898 | 91 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/PrepareLocalImageRequest.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "tools/rbd_mirror/image_replayer/PrepareLocalImageRequest.h"
#include "include/rados/librados.hpp"
#include "cls/rbd/cls_rbd_client.h"
#include "common/debug.h"
#include "common/errno.h"
#include "librbd/ImageCtx.h"
#include "librbd/Journal.h"
#include "librbd/Utils.h"
#include "librbd/mirror/GetInfoRequest.h"
#include "tools/rbd_mirror/ImageDeleter.h"
#include "tools/rbd_mirror/Threads.h"
#include "tools/rbd_mirror/image_replayer/GetMirrorImageIdRequest.h"
#include "tools/rbd_mirror/image_replayer/journal/StateBuilder.h"
#include "tools/rbd_mirror/image_replayer/snapshot/StateBuilder.h"
#include <type_traits>
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_replayer::" \
<< "PrepareLocalImageRequest: " << this << " " \
<< __func__ << ": "
namespace rbd {
namespace mirror {
namespace image_replayer {
using librbd::util::create_context_callback;
using librbd::util::create_rados_callback;
template <typename I>
void PrepareLocalImageRequest<I>::send() {
dout(10) << dendl;
get_local_image_id();
}
template <typename I>
void PrepareLocalImageRequest<I>::get_local_image_id() {
dout(10) << dendl;
Context *ctx = create_context_callback<
PrepareLocalImageRequest<I>,
&PrepareLocalImageRequest<I>::handle_get_local_image_id>(this);
auto req = GetMirrorImageIdRequest<I>::create(m_io_ctx, m_global_image_id,
&m_local_image_id, ctx);
req->send();
}
template <typename I>
void PrepareLocalImageRequest<I>::handle_get_local_image_id(int r) {
dout(10) << "r=" << r << ", "
<< "local_image_id=" << m_local_image_id << dendl;
if (r < 0) {
finish(r);
return;
}
get_local_image_name();
}
template <typename I>
void PrepareLocalImageRequest<I>::get_local_image_name() {
dout(10) << dendl;
librados::ObjectReadOperation op;
librbd::cls_client::dir_get_name_start(&op, m_local_image_id);
m_out_bl.clear();
librados::AioCompletion *aio_comp = create_rados_callback<
PrepareLocalImageRequest<I>,
&PrepareLocalImageRequest<I>::handle_get_local_image_name>(this);
int r = m_io_ctx.aio_operate(RBD_DIRECTORY, aio_comp, &op, &m_out_bl);
ceph_assert(r == 0);
aio_comp->release();
}
template <typename I>
void PrepareLocalImageRequest<I>::handle_get_local_image_name(int r) {
dout(10) << "r=" << r << dendl;
if (r == 0) {
auto it = m_out_bl.cbegin();
r = librbd::cls_client::dir_get_name_finish(&it, m_local_image_name);
}
if (r == -ENOENT) {
// proceed we should have a mirror image record if we got this far
dout(10) << "image does not exist for local image id " << m_local_image_id
<< dendl;
*m_local_image_name = "";
} else if (r < 0) {
derr << "failed to retrieve image name: " << cpp_strerror(r) << dendl;
finish(r);
return;
}
get_mirror_info();
}
template <typename I>
void PrepareLocalImageRequest<I>::get_mirror_info() {
dout(10) << dendl;
auto ctx = create_context_callback<
PrepareLocalImageRequest<I>,
&PrepareLocalImageRequest<I>::handle_get_mirror_info>(this);
auto req = librbd::mirror::GetInfoRequest<I>::create(
m_io_ctx, m_work_queue, m_local_image_id, &m_mirror_image,
&m_promotion_state, &m_primary_mirror_uuid, ctx);
req->send();
}
template <typename I>
void PrepareLocalImageRequest<I>::handle_get_mirror_info(int r) {
dout(10) << ": r=" << r << dendl;
if (r < 0) {
derr << "failed to retrieve local mirror image info: " << cpp_strerror(r)
<< dendl;
finish(r);
return;
}
if (m_mirror_image.state == cls::rbd::MIRROR_IMAGE_STATE_CREATING) {
dout(5) << "local image is still in creating state, issuing a removal"
<< dendl;
move_to_trash();
return;
} else if (m_mirror_image.state == cls::rbd::MIRROR_IMAGE_STATE_DISABLING) {
dout(5) << "local image mirroring is in disabling state" << dendl;
finish(-ERESTART);
return;
}
switch (m_mirror_image.mode) {
case cls::rbd::MIRROR_IMAGE_MODE_JOURNAL:
// journal-based local image exists
{
auto state_builder = journal::StateBuilder<I>::create(m_global_image_id);
state_builder->local_primary_mirror_uuid = m_primary_mirror_uuid;
*m_state_builder = state_builder;
}
break;
case cls::rbd::MIRROR_IMAGE_MODE_SNAPSHOT:
// snapshot-based local image exists
*m_state_builder = snapshot::StateBuilder<I>::create(m_global_image_id);
break;
default:
derr << "unsupported mirror image mode " << m_mirror_image.mode << " "
<< "for image " << m_global_image_id << dendl;
finish(-EOPNOTSUPP);
break;
}
dout(10) << "local_image_id=" << m_local_image_id << ", "
<< "local_promotion_state=" << m_promotion_state << ", "
<< "local_primary_mirror_uuid=" << m_primary_mirror_uuid << dendl;
(*m_state_builder)->local_image_id = m_local_image_id;
(*m_state_builder)->local_promotion_state = m_promotion_state;
finish(0);
}
template <typename I>
void PrepareLocalImageRequest<I>::move_to_trash() {
dout(10) << dendl;
Context *ctx = create_context_callback<
PrepareLocalImageRequest<I>,
&PrepareLocalImageRequest<I>::handle_move_to_trash>(this);
ImageDeleter<I>::trash_move(m_io_ctx, m_global_image_id,
false, m_work_queue, ctx);
}
template <typename I>
void PrepareLocalImageRequest<I>::handle_move_to_trash(int r) {
dout(10) << ": r=" << r << dendl;
finish(-ENOENT);
}
template <typename I>
void PrepareLocalImageRequest<I>::finish(int r) {
dout(10) << "r=" << r << dendl;
m_on_finish->complete(r);
delete this;
}
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_replayer::PrepareLocalImageRequest<librbd::ImageCtx>;
| 6,027 | 29.444444 | 87 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/PrepareLocalImageRequest.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_REPLAYER_PREPARE_LOCAL_IMAGE_REQUEST_H
#define RBD_MIRROR_IMAGE_REPLAYER_PREPARE_LOCAL_IMAGE_REQUEST_H
#include "include/buffer.h"
#include "include/rados/librados_fwd.hpp"
#include "cls/rbd/cls_rbd_types.h"
#include "librbd/mirror/Types.h"
#include <string>
struct Context;
namespace librbd {
struct ImageCtx;
namespace asio { struct ContextWQ; }
} // namespace librbd
namespace rbd {
namespace mirror {
namespace image_replayer {
template <typename> class StateBuilder;
template <typename ImageCtxT = librbd::ImageCtx>
class PrepareLocalImageRequest {
public:
static PrepareLocalImageRequest *create(
librados::IoCtx &io_ctx,
const std::string &global_image_id,
std::string *local_image_name,
StateBuilder<ImageCtxT>** state_builder,
librbd::asio::ContextWQ *work_queue,
Context *on_finish) {
return new PrepareLocalImageRequest(io_ctx, global_image_id,
local_image_name, state_builder,
work_queue, on_finish);
}
PrepareLocalImageRequest(
librados::IoCtx &io_ctx,
const std::string &global_image_id,
std::string *local_image_name,
StateBuilder<ImageCtxT>** state_builder,
librbd::asio::ContextWQ *work_queue,
Context *on_finish)
: m_io_ctx(io_ctx), m_global_image_id(global_image_id),
m_local_image_name(local_image_name), m_state_builder(state_builder),
m_work_queue(work_queue), m_on_finish(on_finish) {
}
void send();
private:
/**
* @verbatim
*
* <start>
* |
* v
* GET_LOCAL_IMAGE_ID
* |
* v
* GET_LOCAL_IMAGE_NAME
* |
* v
* GET_MIRROR_INFO
* |
* | (if the image mirror state is CREATING)
* v
* TRASH_MOVE
* |
* v
* <finish>
*
* @endverbatim
*/
librados::IoCtx &m_io_ctx;
std::string m_global_image_id;
std::string *m_local_image_name;
StateBuilder<ImageCtxT>** m_state_builder;
librbd::asio::ContextWQ *m_work_queue;
Context *m_on_finish;
bufferlist m_out_bl;
std::string m_local_image_id;
cls::rbd::MirrorImage m_mirror_image;
librbd::mirror::PromotionState m_promotion_state;
std::string m_primary_mirror_uuid;
void get_local_image_id();
void handle_get_local_image_id(int r);
void get_local_image_name();
void handle_get_local_image_name(int r);
void get_mirror_info();
void handle_get_mirror_info(int r);
void move_to_trash();
void handle_move_to_trash(int r);
void finish(int r);
};
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_replayer::PrepareLocalImageRequest<librbd::ImageCtx>;
#endif // RBD_MIRROR_IMAGE_REPLAYER_PREPARE_LOCAL_IMAGE_REQUEST_H
| 2,893 | 23.948276 | 94 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/PrepareRemoteImageRequest.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "tools/rbd_mirror/image_replayer/PrepareRemoteImageRequest.h"
#include "include/rados/librados.hpp"
#include "cls/rbd/cls_rbd_client.h"
#include "common/debug.h"
#include "common/errno.h"
#include "journal/Journaler.h"
#include "journal/Settings.h"
#include "librbd/ImageCtx.h"
#include "librbd/Journal.h"
#include "librbd/Utils.h"
#include "librbd/asio/ContextWQ.h"
#include "librbd/mirror/GetInfoRequest.h"
#include "tools/rbd_mirror/Threads.h"
#include "tools/rbd_mirror/image_replayer/GetMirrorImageIdRequest.h"
#include "tools/rbd_mirror/image_replayer/Utils.h"
#include "tools/rbd_mirror/image_replayer/journal/StateBuilder.h"
#include "tools/rbd_mirror/image_replayer/snapshot/StateBuilder.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_replayer::" \
<< "PrepareRemoteImageRequest: " << this << " " \
<< __func__ << ": "
namespace rbd {
namespace mirror {
namespace image_replayer {
using librbd::util::create_async_context_callback;
using librbd::util::create_context_callback;
using librbd::util::create_rados_callback;
template <typename I>
void PrepareRemoteImageRequest<I>::send() {
if (*m_state_builder != nullptr) {
(*m_state_builder)->remote_mirror_uuid = m_remote_pool_meta.mirror_uuid;
auto state_builder = dynamic_cast<snapshot::StateBuilder<I>*>(*m_state_builder);
if (state_builder) {
state_builder->remote_mirror_peer_uuid = m_remote_pool_meta.mirror_peer_uuid;
}
}
get_remote_image_id();
}
template <typename I>
void PrepareRemoteImageRequest<I>::get_remote_image_id() {
dout(10) << dendl;
Context *ctx = create_context_callback<
PrepareRemoteImageRequest<I>,
&PrepareRemoteImageRequest<I>::handle_get_remote_image_id>(this);
auto req = GetMirrorImageIdRequest<I>::create(m_remote_io_ctx,
m_global_image_id,
&m_remote_image_id, ctx);
req->send();
}
template <typename I>
void PrepareRemoteImageRequest<I>::handle_get_remote_image_id(int r) {
dout(10) << "r=" << r << ", "
<< "remote_image_id=" << m_remote_image_id << dendl;
if (r < 0) {
finish(r);
return;
}
get_mirror_info();
}
template <typename I>
void PrepareRemoteImageRequest<I>::get_mirror_info() {
dout(10) << dendl;
auto ctx = create_context_callback<
PrepareRemoteImageRequest<I>,
&PrepareRemoteImageRequest<I>::handle_get_mirror_info>(this);
auto req = librbd::mirror::GetInfoRequest<I>::create(
m_remote_io_ctx, m_threads->work_queue, m_remote_image_id,
&m_mirror_image, &m_promotion_state, &m_primary_mirror_uuid,
ctx);
req->send();
}
template <typename I>
void PrepareRemoteImageRequest<I>::handle_get_mirror_info(int r) {
dout(10) << "r=" << r << dendl;
if (r == -ENOENT) {
dout(10) << "image " << m_global_image_id << " not mirrored" << dendl;
finish(r);
return;
} else if (r < 0) {
derr << "failed to retrieve mirror image details for image "
<< m_global_image_id << ": " << cpp_strerror(r) << dendl;
finish(r);
return;
}
auto state_builder = *m_state_builder;
if (state_builder != nullptr &&
state_builder->get_mirror_image_mode() != m_mirror_image.mode) {
derr << "local and remote mirror image using different mirroring modes "
<< "for image " << m_global_image_id << ": split-brain" << dendl;
finish(-EEXIST);
return;
} else if (m_mirror_image.state == cls::rbd::MIRROR_IMAGE_STATE_DISABLING) {
dout(5) << "remote image mirroring is being disabled" << dendl;
finish(-ENOENT);
return;
}
switch (m_mirror_image.mode) {
case cls::rbd::MIRROR_IMAGE_MODE_JOURNAL:
get_client();
break;
case cls::rbd::MIRROR_IMAGE_MODE_SNAPSHOT:
finalize_snapshot_state_builder();
finish(0);
break;
default:
derr << "unsupported mirror image mode " << m_mirror_image.mode << " "
<< "for image " << m_global_image_id << dendl;
finish(-EOPNOTSUPP);
break;
}
}
template <typename I>
void PrepareRemoteImageRequest<I>::get_client() {
dout(10) << dendl;
auto cct = static_cast<CephContext *>(m_local_io_ctx.cct());
::journal::Settings journal_settings;
journal_settings.commit_interval = cct->_conf.get_val<double>(
"rbd_mirror_journal_commit_age");
// TODO use Journal thread pool for journal ops until converted to ASIO
ContextWQ* context_wq;
librbd::Journal<>::get_work_queue(cct, &context_wq);
ceph_assert(m_remote_journaler == nullptr);
m_remote_journaler = new Journaler(context_wq, m_threads->timer,
&m_threads->timer_lock, m_remote_io_ctx,
m_remote_image_id, m_local_mirror_uuid,
journal_settings, m_cache_manager_handler);
Context *ctx = create_async_context_callback(
m_threads->work_queue, create_context_callback<
PrepareRemoteImageRequest<I>,
&PrepareRemoteImageRequest<I>::handle_get_client>(this));
m_remote_journaler->get_client(m_local_mirror_uuid, &m_client, ctx);
}
template <typename I>
void PrepareRemoteImageRequest<I>::handle_get_client(int r) {
dout(10) << "r=" << r << dendl;
MirrorPeerClientMeta client_meta;
if (r == -ENOENT) {
dout(10) << "client not registered" << dendl;
register_client();
} else if (r < 0) {
derr << "failed to retrieve client: " << cpp_strerror(r) << dendl;
finish(r);
} else if (!util::decode_client_meta(m_client, &client_meta)) {
// require operator intervention since the data is corrupt
finish(-EBADMSG);
} else {
// skip registration if it already exists
finalize_journal_state_builder(m_client.state, client_meta);
finish(0);
}
}
template <typename I>
void PrepareRemoteImageRequest<I>::register_client() {
dout(10) << dendl;
auto state_builder = *m_state_builder;
librbd::journal::MirrorPeerClientMeta client_meta{
(state_builder == nullptr ? "" : state_builder->local_image_id)};
client_meta.state = librbd::journal::MIRROR_PEER_STATE_REPLAYING;
librbd::journal::ClientData client_data{client_meta};
bufferlist client_data_bl;
encode(client_data, client_data_bl);
Context *ctx = create_async_context_callback(
m_threads->work_queue, create_context_callback<
PrepareRemoteImageRequest<I>,
&PrepareRemoteImageRequest<I>::handle_register_client>(this));
m_remote_journaler->register_client(client_data_bl, ctx);
}
template <typename I>
void PrepareRemoteImageRequest<I>::handle_register_client(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to register with remote journal: " << cpp_strerror(r)
<< dendl;
finish(r);
return;
}
auto state_builder = *m_state_builder;
librbd::journal::MirrorPeerClientMeta client_meta{
(state_builder == nullptr ? "" : state_builder->local_image_id)};
client_meta.state = librbd::journal::MIRROR_PEER_STATE_REPLAYING;
finalize_journal_state_builder(cls::journal::CLIENT_STATE_CONNECTED,
client_meta);
finish(0);
}
template <typename I>
void PrepareRemoteImageRequest<I>::finalize_journal_state_builder(
cls::journal::ClientState client_state,
const MirrorPeerClientMeta& client_meta) {
journal::StateBuilder<I>* state_builder = nullptr;
if (*m_state_builder != nullptr) {
// already verified that it's a matching builder in
// 'handle_get_mirror_info'
state_builder = dynamic_cast<journal::StateBuilder<I>*>(*m_state_builder);
ceph_assert(state_builder != nullptr);
} else {
state_builder = journal::StateBuilder<I>::create(m_global_image_id);
*m_state_builder = state_builder;
}
state_builder->remote_mirror_uuid = m_remote_pool_meta.mirror_uuid;
state_builder->remote_image_id = m_remote_image_id;
state_builder->remote_promotion_state = m_promotion_state;
state_builder->remote_journaler = m_remote_journaler;
state_builder->remote_client_state = client_state;
state_builder->remote_client_meta = client_meta;
}
template <typename I>
void PrepareRemoteImageRequest<I>::finalize_snapshot_state_builder() {
snapshot::StateBuilder<I>* state_builder = nullptr;
if (*m_state_builder != nullptr) {
state_builder = dynamic_cast<snapshot::StateBuilder<I>*>(*m_state_builder);
ceph_assert(state_builder != nullptr);
} else {
state_builder = snapshot::StateBuilder<I>::create(m_global_image_id);
*m_state_builder = state_builder;
}
dout(10) << "remote_mirror_uuid=" << m_remote_pool_meta.mirror_uuid << ", "
<< "remote_mirror_peer_uuid="
<< m_remote_pool_meta.mirror_peer_uuid << ", "
<< "remote_image_id=" << m_remote_image_id << ", "
<< "remote_promotion_state=" << m_promotion_state << dendl;
state_builder->remote_mirror_uuid = m_remote_pool_meta.mirror_uuid;
state_builder->remote_mirror_peer_uuid = m_remote_pool_meta.mirror_peer_uuid;
state_builder->remote_image_id = m_remote_image_id;
state_builder->remote_promotion_state = m_promotion_state;
}
template <typename I>
void PrepareRemoteImageRequest<I>::finish(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
delete m_remote_journaler;
m_remote_journaler = nullptr;
}
m_on_finish->complete(r);
delete this;
}
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_replayer::PrepareRemoteImageRequest<librbd::ImageCtx>;
| 9,702 | 33.165493 | 88 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/PrepareRemoteImageRequest.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_REPLAYER_PREPARE_REMOTE_IMAGE_REQUEST_H
#define RBD_MIRROR_IMAGE_REPLAYER_PREPARE_REMOTE_IMAGE_REQUEST_H
#include "include/buffer_fwd.h"
#include "include/rados/librados_fwd.hpp"
#include "cls/journal/cls_journal_types.h"
#include "cls/rbd/cls_rbd_types.h"
#include "librbd/journal/Types.h"
#include "librbd/journal/TypeTraits.h"
#include "librbd/mirror/Types.h"
#include "tools/rbd_mirror/Types.h"
#include <string>
namespace journal { class Journaler; }
namespace journal { struct CacheManagerHandler; }
namespace librbd { struct ImageCtx; }
namespace librbd { namespace journal { struct MirrorPeerClientMeta; } }
struct Context;
namespace rbd {
namespace mirror {
template <typename> struct Threads;
namespace image_replayer {
template <typename> class StateBuilder;
template <typename ImageCtxT = librbd::ImageCtx>
class PrepareRemoteImageRequest {
public:
typedef librbd::journal::TypeTraits<ImageCtxT> TypeTraits;
typedef typename TypeTraits::Journaler Journaler;
typedef librbd::journal::MirrorPeerClientMeta MirrorPeerClientMeta;
static PrepareRemoteImageRequest *create(
Threads<ImageCtxT> *threads,
librados::IoCtx &local_io_ctx,
librados::IoCtx &remote_io_ctx,
const std::string &global_image_id,
const std::string &local_mirror_uuid,
const RemotePoolMeta& remote_pool_meta,
::journal::CacheManagerHandler *cache_manager_handler,
StateBuilder<ImageCtxT>** state_builder,
Context *on_finish) {
return new PrepareRemoteImageRequest(threads, local_io_ctx, remote_io_ctx,
global_image_id, local_mirror_uuid,
remote_pool_meta,
cache_manager_handler, state_builder,
on_finish);
}
PrepareRemoteImageRequest(
Threads<ImageCtxT> *threads,
librados::IoCtx &local_io_ctx,
librados::IoCtx &remote_io_ctx,
const std::string &global_image_id,
const std::string &local_mirror_uuid,
const RemotePoolMeta& remote_pool_meta,
::journal::CacheManagerHandler *cache_manager_handler,
StateBuilder<ImageCtxT>** state_builder,
Context *on_finish)
: m_threads(threads),
m_local_io_ctx(local_io_ctx),
m_remote_io_ctx(remote_io_ctx),
m_global_image_id(global_image_id),
m_local_mirror_uuid(local_mirror_uuid),
m_remote_pool_meta(remote_pool_meta),
m_cache_manager_handler(cache_manager_handler),
m_state_builder(state_builder),
m_on_finish(on_finish) {
}
void send();
private:
/**
* @verbatim
*
* <start>
* |
* v
* GET_REMOTE_IMAGE_ID
* |
* v
* GET_REMOTE_MIRROR_INFO
* |
* | (journal)
* \-----------> GET_CLIENT
* | |
* | v (skip if not needed)
* | REGISTER_CLIENT
* | |
* | |
* |/----------------/
* |
* v
* <finish>
*
* @endverbatim
*/
Threads<ImageCtxT> *m_threads;
librados::IoCtx &m_local_io_ctx;
librados::IoCtx &m_remote_io_ctx;
std::string m_global_image_id;
std::string m_local_mirror_uuid;
RemotePoolMeta m_remote_pool_meta;
::journal::CacheManagerHandler *m_cache_manager_handler;
StateBuilder<ImageCtxT>** m_state_builder;
Context *m_on_finish;
bufferlist m_out_bl;
std::string m_remote_image_id;
cls::rbd::MirrorImage m_mirror_image;
librbd::mirror::PromotionState m_promotion_state =
librbd::mirror::PROMOTION_STATE_UNKNOWN;
std::string m_primary_mirror_uuid;
// journal-based mirroring
Journaler *m_remote_journaler = nullptr;
cls::journal::Client m_client;
void get_remote_image_id();
void handle_get_remote_image_id(int r);
void get_mirror_info();
void handle_get_mirror_info(int r);
void get_client();
void handle_get_client(int r);
void register_client();
void handle_register_client(int r);
void finalize_journal_state_builder(cls::journal::ClientState client_state,
const MirrorPeerClientMeta& client_meta);
void finalize_snapshot_state_builder();
void finish(int r);
};
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_replayer::PrepareRemoteImageRequest<librbd::ImageCtx>;
#endif // RBD_MIRROR_IMAGE_REPLAYER_PREPARE_REMOTE_IMAGE_REQUEST_H
| 4,596 | 28.850649 | 95 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/Replayer.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_REPLAYER_REPLAYER_H
#define RBD_MIRROR_IMAGE_REPLAYER_REPLAYER_H
#include <string>
struct Context;
namespace rbd {
namespace mirror {
namespace image_replayer {
struct Replayer {
virtual ~Replayer() {}
virtual void destroy() = 0;
virtual void init(Context* on_finish) = 0;
virtual void shut_down(Context* on_finish) = 0;
virtual void flush(Context* on_finish) = 0;
virtual bool get_replay_status(std::string* description,
Context* on_finish) = 0;
virtual bool is_replaying() const = 0;
virtual bool is_resync_requested() const = 0;
virtual int get_error_code() const = 0;
virtual std::string get_error_description() const = 0;
};
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
#endif // RBD_MIRROR_IMAGE_REPLAYER_REPLAYER_H
| 937 | 22.45 | 70 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/ReplayerListener.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_REPLAYER_REPLAYER_LISTENER_H
#define RBD_MIRROR_IMAGE_REPLAYER_REPLAYER_LISTENER_H
namespace rbd {
namespace mirror {
namespace image_replayer {
struct ReplayerListener {
virtual ~ReplayerListener() {}
virtual void handle_notification() = 0;
};
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
#endif // RBD_MIRROR_IMAGE_REPLAYER_REPLAYER_LISTENER_H
| 505 | 22 | 70 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/StateBuilder.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "StateBuilder.h"
#include "include/ceph_assert.h"
#include "include/Context.h"
#include "common/debug.h"
#include "common/errno.h"
#include "journal/Journaler.h"
#include "librbd/ImageCtx.h"
#include "tools/rbd_mirror/image_replayer/CloseImageRequest.h"
#include "tools/rbd_mirror/image_sync/Types.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_replayer::" \
<< "StateBuilder: " << this << " " \
<< __func__ << ": "
namespace rbd {
namespace mirror {
namespace image_replayer {
template <typename I>
StateBuilder<I>::StateBuilder(const std::string& global_image_id)
: global_image_id(global_image_id) {
dout(10) << "global_image_id=" << global_image_id << dendl;
}
template <typename I>
StateBuilder<I>::~StateBuilder() {
ceph_assert(local_image_ctx == nullptr);
ceph_assert(remote_image_ctx == nullptr);
ceph_assert(m_sync_point_handler == nullptr);
}
template <typename I>
bool StateBuilder<I>::is_local_primary() const {
if (local_promotion_state == librbd::mirror::PROMOTION_STATE_PRIMARY) {
ceph_assert(!local_image_id.empty());
return true;
}
return false;
}
template <typename I>
bool StateBuilder<I>::is_remote_primary() const {
if (remote_promotion_state == librbd::mirror::PROMOTION_STATE_PRIMARY) {
ceph_assert(!remote_image_id.empty());
return true;
}
return false;
}
template <typename I>
bool StateBuilder<I>::is_linked() const {
if (local_promotion_state == librbd::mirror::PROMOTION_STATE_NON_PRIMARY) {
ceph_assert(!local_image_id.empty());
return is_linked_impl();
}
return false;
}
template <typename I>
void StateBuilder<I>::close_local_image(Context* on_finish) {
if (local_image_ctx == nullptr) {
on_finish->complete(0);
return;
}
dout(10) << dendl;
auto ctx = new LambdaContext([this, on_finish](int r) {
handle_close_local_image(r, on_finish);
});
auto request = image_replayer::CloseImageRequest<I>::create(
&local_image_ctx, ctx);
request->send();
}
template <typename I>
void StateBuilder<I>::handle_close_local_image(int r, Context* on_finish) {
dout(10) << "r=" << r << dendl;
ceph_assert(local_image_ctx == nullptr);
if (r < 0) {
derr << "failed to close local image for image " << global_image_id << ": "
<< cpp_strerror(r) << dendl;
}
on_finish->complete(r);
}
template <typename I>
void StateBuilder<I>::close_remote_image(Context* on_finish) {
if (remote_image_ctx == nullptr) {
on_finish->complete(0);
return;
}
dout(10) << dendl;
auto ctx = new LambdaContext([this, on_finish](int r) {
handle_close_remote_image(r, on_finish);
});
auto request = image_replayer::CloseImageRequest<I>::create(
&remote_image_ctx, ctx);
request->send();
}
template <typename I>
void StateBuilder<I>::handle_close_remote_image(int r, Context* on_finish) {
dout(10) << "r=" << r << dendl;
ceph_assert(remote_image_ctx == nullptr);
if (r < 0) {
derr << "failed to close remote image for image " << global_image_id << ": "
<< cpp_strerror(r) << dendl;
}
on_finish->complete(r);
}
template <typename I>
void StateBuilder<I>::destroy_sync_point_handler() {
if (m_sync_point_handler == nullptr) {
return;
}
dout(15) << dendl;
m_sync_point_handler->destroy();
m_sync_point_handler = nullptr;
}
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_replayer::StateBuilder<librbd::ImageCtx>;
| 3,716 | 25.741007 | 80 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/StateBuilder.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef CEPH_RBD_MIRROR_IMAGE_REPLAYER_STATE_BUILDER_H
#define CEPH_RBD_MIRROR_IMAGE_REPLAYER_STATE_BUILDER_H
#include "include/rados/librados_fwd.hpp"
#include "cls/rbd/cls_rbd_types.h"
#include "librbd/mirror/Types.h"
struct Context;
namespace librbd { struct ImageCtx; }
namespace rbd {
namespace mirror {
struct BaseRequest;
template <typename> class InstanceWatcher;
struct PoolMetaCache;
struct ProgressContext;
template <typename> class Threads;
namespace image_sync { struct SyncPointHandler; }
namespace image_replayer {
struct Replayer;
struct ReplayerListener;
template <typename ImageCtxT>
class StateBuilder {
public:
StateBuilder(const StateBuilder&) = delete;
StateBuilder& operator=(const StateBuilder&) = delete;
virtual ~StateBuilder();
virtual void destroy() {
delete this;
}
virtual void close(Context* on_finish) = 0;
virtual bool is_disconnected() const = 0;
bool is_local_primary() const;
bool is_remote_primary() const;
bool is_linked() const;
virtual cls::rbd::MirrorImageMode get_mirror_image_mode() const = 0;
virtual image_sync::SyncPointHandler* create_sync_point_handler() = 0;
void destroy_sync_point_handler();
virtual bool replay_requires_remote_image() const = 0;
void close_remote_image(Context* on_finish);
virtual BaseRequest* create_local_image_request(
Threads<ImageCtxT>* threads,
librados::IoCtx& local_io_ctx,
const std::string& global_image_id,
PoolMetaCache* pool_meta_cache,
ProgressContext* progress_ctx,
Context* on_finish) = 0;
virtual BaseRequest* create_prepare_replay_request(
const std::string& local_mirror_uuid,
ProgressContext* progress_ctx,
bool* resync_requested,
bool* syncing,
Context* on_finish) = 0;
virtual Replayer* create_replayer(
Threads<ImageCtxT>* threads,
InstanceWatcher<ImageCtxT>* instance_watcher,
const std::string& local_mirror_uuid,
PoolMetaCache* pool_meta_cache,
ReplayerListener* replayer_listener) = 0;
std::string global_image_id;
std::string local_image_id;
librbd::mirror::PromotionState local_promotion_state =
librbd::mirror::PROMOTION_STATE_UNKNOWN;
ImageCtxT* local_image_ctx = nullptr;
std::string remote_mirror_uuid;
std::string remote_image_id;
librbd::mirror::PromotionState remote_promotion_state =
librbd::mirror::PROMOTION_STATE_UNKNOWN;
ImageCtxT* remote_image_ctx = nullptr;
protected:
image_sync::SyncPointHandler* m_sync_point_handler = nullptr;
StateBuilder(const std::string& global_image_id);
void close_local_image(Context* on_finish);
private:
virtual bool is_linked_impl() const = 0;
void handle_close_local_image(int r, Context* on_finish);
void handle_close_remote_image(int r, Context* on_finish);
};
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_replayer::StateBuilder<librbd::ImageCtx>;
#endif // CEPH_RBD_MIRROR_IMAGE_REPLAYER_STATE_BUILDER_H
| 3,121 | 26.147826 | 82 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/TimeRollingMean.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "tools/rbd_mirror/image_replayer/TimeRollingMean.h"
#include "common/Clock.h"
namespace rbd {
namespace mirror {
namespace image_replayer {
void TimeRollingMean::operator()(uint32_t value) {
auto time = ceph_clock_now();
if (m_last_time.is_zero()) {
m_last_time = time;
} else if (m_last_time.sec() < time.sec()) {
auto sec = m_last_time.sec();
while (sec++ < time.sec()) {
m_rolling_mean(m_sum);
m_sum = 0;
}
m_last_time = time;
}
m_sum += value;
}
double TimeRollingMean::get_average() const {
return boost::accumulators::rolling_mean(m_rolling_mean);
}
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
| 785 | 21.457143 | 70 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/TimeRollingMean.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_REPLAYER_TIME_ROLLING_MEAN_H
#define RBD_MIRROR_IMAGE_REPLAYER_TIME_ROLLING_MEAN_H
#include "include/utime.h"
#include <boost/accumulators/accumulators.hpp>
#include <boost/accumulators/statistics/stats.hpp>
#include <boost/accumulators/statistics/rolling_mean.hpp>
namespace rbd {
namespace mirror {
namespace image_replayer {
class TimeRollingMean {
public:
void operator()(uint32_t value);
double get_average() const;
private:
typedef boost::accumulators::accumulator_set<
uint64_t, boost::accumulators::stats<
boost::accumulators::tag::rolling_mean>> RollingMean;
utime_t m_last_time;
uint64_t m_sum = 0;
RollingMean m_rolling_mean{
boost::accumulators::tag::rolling_window::window_size = 30};
};
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
#endif // RBD_MIRROR_IMAGE_REPLAYER_TIME_ROLLING_MEAN_H
| 989 | 23.146341 | 70 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/Types.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef CEPH_RBD_MIRROR_IMAGE_REPLAYER_TYPES_H
#define CEPH_RBD_MIRROR_IMAGE_REPLAYER_TYPES_H
namespace rbd {
namespace mirror {
namespace image_replayer {
enum HealthState {
HEALTH_STATE_OK,
HEALTH_STATE_WARNING,
HEALTH_STATE_ERROR
};
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
#endif // CEPH_RBD_MIRROR_IMAGE_REPLAYER_TYPES_H
| 465 | 20.181818 | 70 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/Utils.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "tools/rbd_mirror/image_replayer/Utils.h"
#include "include/rados/librados.hpp"
#include "common/debug.h"
#include "common/errno.h"
#include "cls/journal/cls_journal_types.h"
#include "librbd/journal/Types.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_replayer::util::" \
<< __func__ << ": "
namespace rbd {
namespace mirror {
namespace image_replayer {
namespace util {
std::string compute_image_spec(librados::IoCtx& io_ctx,
const std::string& image_name) {
std::string name = io_ctx.get_namespace();
if (!name.empty()) {
name += "/";
}
return io_ctx.get_pool_name() + "/" + name + image_name;
}
bool decode_client_meta(const cls::journal::Client& client,
librbd::journal::MirrorPeerClientMeta* client_meta) {
dout(15) << dendl;
librbd::journal::ClientData client_data;
auto it = client.data.cbegin();
try {
decode(client_data, it);
} catch (const buffer::error &err) {
derr << "failed to decode client meta data: " << err.what() << dendl;
return false;
}
auto local_client_meta = boost::get<librbd::journal::MirrorPeerClientMeta>(
&client_data.client_meta);
if (local_client_meta == nullptr) {
derr << "unknown peer registration" << dendl;
return false;
}
*client_meta = *local_client_meta;
dout(15) << "client found: client_meta=" << *client_meta << dendl;
return true;
}
} // namespace util
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
| 1,732 | 26.951613 | 77 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/Utils.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_REPLAYER_UTILS_H
#define RBD_MIRROR_IMAGE_REPLAYER_UTILS_H
#include "include/rados/librados_fwd.hpp"
#include <string>
namespace cls { namespace journal { struct Client; } }
namespace librbd { namespace journal { struct MirrorPeerClientMeta; } }
namespace rbd {
namespace mirror {
namespace image_replayer {
namespace util {
std::string compute_image_spec(librados::IoCtx& io_ctx,
const std::string& image_name);
bool decode_client_meta(const cls::journal::Client& client,
librbd::journal::MirrorPeerClientMeta* client_meta);
} // namespace util
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
#endif // RBD_MIRROR_IMAGE_REPLAYER_UTILS_H
| 847 | 27.266667 | 76 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/journal/CreateLocalImageRequest.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "CreateLocalImageRequest.h"
#include "include/rados/librados.hpp"
#include "common/debug.h"
#include "common/dout.h"
#include "common/errno.h"
#include "journal/Journaler.h"
#include "librbd/ImageCtx.h"
#include "librbd/Utils.h"
#include "librbd/journal/Types.h"
#include "tools/rbd_mirror/PoolMetaCache.h"
#include "tools/rbd_mirror/ProgressContext.h"
#include "tools/rbd_mirror/Threads.h"
#include "tools/rbd_mirror/image_replayer/CreateImageRequest.h"
#include "tools/rbd_mirror/image_replayer/Utils.h"
#include "tools/rbd_mirror/image_replayer/journal/StateBuilder.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_replayer::journal::" \
<< "CreateLocalImageRequest: " << this << " " \
<< __func__ << ": "
namespace rbd {
namespace mirror {
namespace image_replayer {
namespace journal {
using librbd::util::create_async_context_callback;
using librbd::util::create_context_callback;
template <typename I>
void CreateLocalImageRequest<I>::send() {
unregister_client();
}
template <typename I>
void CreateLocalImageRequest<I>::unregister_client() {
dout(10) << dendl;
update_progress("UNREGISTER_CLIENT");
auto ctx = create_context_callback<
CreateLocalImageRequest<I>,
&CreateLocalImageRequest<I>::handle_unregister_client>(this);
m_state_builder->remote_journaler->unregister_client(ctx);
}
template <typename I>
void CreateLocalImageRequest<I>::handle_unregister_client(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0 && r != -ENOENT) {
derr << "failed to unregister with remote journal: " << cpp_strerror(r)
<< dendl;
finish(r);
return;
}
m_state_builder->local_image_id = "";
m_state_builder->remote_client_meta = {};
register_client();
}
template <typename I>
void CreateLocalImageRequest<I>::register_client() {
ceph_assert(m_state_builder->local_image_id.empty());
m_state_builder->local_image_id =
librbd::util::generate_image_id<I>(m_local_io_ctx);
dout(10) << "local_image_id=" << m_state_builder->local_image_id << dendl;
update_progress("REGISTER_CLIENT");
librbd::journal::MirrorPeerClientMeta client_meta{
m_state_builder->local_image_id};
client_meta.state = librbd::journal::MIRROR_PEER_STATE_SYNCING;
librbd::journal::ClientData client_data{client_meta};
bufferlist client_data_bl;
encode(client_data, client_data_bl);
auto ctx = create_context_callback<
CreateLocalImageRequest<I>,
&CreateLocalImageRequest<I>::handle_register_client>(this);
m_state_builder->remote_journaler->register_client(client_data_bl, ctx);
}
template <typename I>
void CreateLocalImageRequest<I>::handle_register_client(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to register with remote journal: " << cpp_strerror(r)
<< dendl;
finish(r);
return;
}
m_state_builder->remote_client_state = cls::journal::CLIENT_STATE_CONNECTED;
m_state_builder->remote_client_meta = {m_state_builder->local_image_id};
m_state_builder->remote_client_meta.state =
librbd::journal::MIRROR_PEER_STATE_SYNCING;
create_local_image();
}
template <typename I>
void CreateLocalImageRequest<I>::create_local_image() {
dout(10) << "local_image_id=" << m_state_builder->local_image_id << dendl;
update_progress("CREATE_LOCAL_IMAGE");
m_remote_image_ctx->image_lock.lock_shared();
std::string image_name = m_remote_image_ctx->name;
m_remote_image_ctx->image_lock.unlock_shared();
auto ctx = create_context_callback<
CreateLocalImageRequest<I>,
&CreateLocalImageRequest<I>::handle_create_local_image>(this);
auto request = CreateImageRequest<I>::create(
m_threads, m_local_io_ctx, m_global_image_id,
m_state_builder->remote_mirror_uuid, image_name,
m_state_builder->local_image_id, m_remote_image_ctx,
m_pool_meta_cache, cls::rbd::MIRROR_IMAGE_MODE_JOURNAL, ctx);
request->send();
}
template <typename I>
void CreateLocalImageRequest<I>::handle_create_local_image(int r) {
dout(10) << "r=" << r << dendl;
if (r == -EBADF) {
dout(5) << "image id " << m_state_builder->local_image_id << " "
<< "already in-use" << dendl;
unregister_client();
return;
} else if (r < 0) {
if (r == -ENOENT) {
dout(10) << "parent image does not exist" << dendl;
} else {
derr << "failed to create local image: " << cpp_strerror(r) << dendl;
}
finish(r);
return;
}
finish(0);
}
template <typename I>
void CreateLocalImageRequest<I>::update_progress(
const std::string& description) {
dout(15) << description << dendl;
if (m_progress_ctx != nullptr) {
m_progress_ctx->update_progress(description);
}
}
} // namespace journal
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_replayer::journal::CreateLocalImageRequest<librbd::ImageCtx>;
| 5,096 | 30.269939 | 95 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/journal/CreateLocalImageRequest.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_REPLAYER_JOURNAL_CREATE_LOCAL_IMAGE_REQUEST_H
#define RBD_MIRROR_IMAGE_REPLAYER_JOURNAL_CREATE_LOCAL_IMAGE_REQUEST_H
#include "include/rados/librados_fwd.hpp"
#include "tools/rbd_mirror/BaseRequest.h"
#include <string>
struct Context;
namespace journal { class Journaler; }
namespace librbd { class ImageCtx; }
namespace rbd {
namespace mirror {
class PoolMetaCache;
class ProgressContext;
template <typename> struct Threads;
namespace image_replayer {
namespace journal {
template <typename> class StateBuilder;
template <typename ImageCtxT>
class CreateLocalImageRequest : public BaseRequest {
public:
typedef rbd::mirror::ProgressContext ProgressContext;
static CreateLocalImageRequest* create(
Threads<ImageCtxT>* threads,
librados::IoCtx& local_io_ctx,
ImageCtxT* remote_image_ctx,
const std::string& global_image_id,
PoolMetaCache* pool_meta_cache,
ProgressContext* progress_ctx,
StateBuilder<ImageCtxT>* state_builder,
Context* on_finish) {
return new CreateLocalImageRequest(threads, local_io_ctx, remote_image_ctx,
global_image_id, pool_meta_cache,
progress_ctx, state_builder, on_finish);
}
CreateLocalImageRequest(
Threads<ImageCtxT>* threads,
librados::IoCtx& local_io_ctx,
ImageCtxT* remote_image_ctx,
const std::string& global_image_id,
PoolMetaCache* pool_meta_cache,
ProgressContext* progress_ctx,
StateBuilder<ImageCtxT>* state_builder,
Context* on_finish)
: BaseRequest(on_finish),
m_threads(threads),
m_local_io_ctx(local_io_ctx),
m_remote_image_ctx(remote_image_ctx),
m_global_image_id(global_image_id),
m_pool_meta_cache(pool_meta_cache),
m_progress_ctx(progress_ctx),
m_state_builder(state_builder) {
}
void send();
private:
/**
* @verbatim
*
* <start>
* |
* v
* UNREGISTER_CLIENT < * * * * * * * *
* | *
* v *
* REGISTER_CLIENT *
* | *
* v (id exists) *
* CREATE_LOCAL_IMAGE * * * * * * * * *
* |
* v
* <finish>
*
* @endverbatim
*/
Threads<ImageCtxT>* m_threads;
librados::IoCtx& m_local_io_ctx;
ImageCtxT* m_remote_image_ctx;
std::string m_global_image_id;
PoolMetaCache* m_pool_meta_cache;
ProgressContext* m_progress_ctx;
StateBuilder<ImageCtxT>* m_state_builder;
void unregister_client();
void handle_unregister_client(int r);
void register_client();
void handle_register_client(int r);
void create_local_image();
void handle_create_local_image(int r);
void update_progress(const std::string& description);
};
} // namespace journal
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_replayer::journal::CreateLocalImageRequest<librbd::ImageCtx>;
#endif // RBD_MIRROR_IMAGE_REPLAYER_JOURNAL_CREATE_LOCAL_IMAGE_REQUEST_H
| 3,223 | 26.555556 | 102 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/journal/EventPreprocessor.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "EventPreprocessor.h"
#include "common/debug.h"
#include "common/dout.h"
#include "common/errno.h"
#include "journal/Journaler.h"
#include "librbd/ImageCtx.h"
#include "librbd/ImageState.h"
#include "librbd/Utils.h"
#include "librbd/asio/ContextWQ.h"
#include "librbd/journal/Types.h"
#include <boost/variant.hpp>
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_replayer::journal::" \
<< "EventPreprocessor: " << this << " " << __func__ \
<< ": "
namespace rbd {
namespace mirror {
namespace image_replayer {
namespace journal {
using librbd::util::create_context_callback;
template <typename I>
EventPreprocessor<I>::EventPreprocessor(I &local_image_ctx,
Journaler &remote_journaler,
const std::string &local_mirror_uuid,
MirrorPeerClientMeta *client_meta,
librbd::asio::ContextWQ *work_queue)
: m_local_image_ctx(local_image_ctx), m_remote_journaler(remote_journaler),
m_local_mirror_uuid(local_mirror_uuid), m_client_meta(client_meta),
m_work_queue(work_queue) {
}
template <typename I>
EventPreprocessor<I>::~EventPreprocessor() {
ceph_assert(!m_in_progress);
}
template <typename I>
bool EventPreprocessor<I>::is_required(const EventEntry &event_entry) {
SnapSeqs snap_seqs(m_client_meta->snap_seqs);
return (prune_snap_map(&snap_seqs) ||
event_entry.get_event_type() ==
librbd::journal::EVENT_TYPE_SNAP_RENAME);
}
template <typename I>
void EventPreprocessor<I>::preprocess(EventEntry *event_entry,
Context *on_finish) {
ceph_assert(!m_in_progress);
m_in_progress = true;
m_event_entry = event_entry;
m_on_finish = on_finish;
refresh_image();
}
template <typename I>
void EventPreprocessor<I>::refresh_image() {
dout(20) << dendl;
Context *ctx = create_context_callback<
EventPreprocessor<I>, &EventPreprocessor<I>::handle_refresh_image>(this);
m_local_image_ctx.state->refresh(ctx);
}
template <typename I>
void EventPreprocessor<I>::handle_refresh_image(int r) {
dout(20) << "r=" << r << dendl;
if (r < 0) {
derr << "error encountered during image refresh: " << cpp_strerror(r)
<< dendl;
finish(r);
return;
}
preprocess_event();
}
template <typename I>
void EventPreprocessor<I>::preprocess_event() {
dout(20) << dendl;
m_snap_seqs = m_client_meta->snap_seqs;
m_snap_seqs_updated = prune_snap_map(&m_snap_seqs);
int r = boost::apply_visitor(PreprocessEventVisitor(this),
m_event_entry->event);
if (r < 0) {
finish(r);
return;
}
update_client();
}
template <typename I>
int EventPreprocessor<I>::preprocess_snap_rename(
librbd::journal::SnapRenameEvent &event) {
dout(20) << "remote_snap_id=" << event.snap_id << ", "
<< "src_snap_name=" << event.src_snap_name << ", "
<< "dest_snap_name=" << event.dst_snap_name << dendl;
auto snap_seq_it = m_snap_seqs.find(event.snap_id);
if (snap_seq_it != m_snap_seqs.end()) {
dout(20) << "remapping remote snap id " << snap_seq_it->first << " "
<< "to local snap id " << snap_seq_it->second << dendl;
event.snap_id = snap_seq_it->second;
return 0;
}
auto snap_id_it = m_local_image_ctx.snap_ids.find({cls::rbd::UserSnapshotNamespace(),
event.src_snap_name});
if (snap_id_it == m_local_image_ctx.snap_ids.end()) {
dout(20) << "cannot map remote snapshot '" << event.src_snap_name << "' "
<< "to local snapshot" << dendl;
event.snap_id = CEPH_NOSNAP;
return -ENOENT;
}
dout(20) << "mapping remote snap id " << event.snap_id << " "
<< "to local snap id " << snap_id_it->second << dendl;
m_snap_seqs_updated = true;
m_snap_seqs[event.snap_id] = snap_id_it->second;
event.snap_id = snap_id_it->second;
return 0;
}
template <typename I>
void EventPreprocessor<I>::update_client() {
if (!m_snap_seqs_updated) {
finish(0);
return;
}
dout(20) << dendl;
librbd::journal::MirrorPeerClientMeta client_meta(*m_client_meta);
client_meta.snap_seqs = m_snap_seqs;
librbd::journal::ClientData client_data(client_meta);
bufferlist data_bl;
encode(client_data, data_bl);
Context *ctx = create_context_callback<
EventPreprocessor<I>, &EventPreprocessor<I>::handle_update_client>(
this);
m_remote_journaler.update_client(data_bl, ctx);
}
template <typename I>
void EventPreprocessor<I>::handle_update_client(int r) {
dout(20) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to update mirror peer journal client: "
<< cpp_strerror(r) << dendl;
finish(r);
return;
}
m_client_meta->snap_seqs = m_snap_seqs;
finish(0);
}
template <typename I>
bool EventPreprocessor<I>::prune_snap_map(SnapSeqs *snap_seqs) {
bool pruned = false;
std::shared_lock image_locker{m_local_image_ctx.image_lock};
for (auto it = snap_seqs->begin(); it != snap_seqs->end(); ) {
auto current_it(it++);
if (m_local_image_ctx.snap_info.count(current_it->second) == 0) {
snap_seqs->erase(current_it);
pruned = true;
}
}
return pruned;
}
template <typename I>
void EventPreprocessor<I>::finish(int r) {
dout(20) << "r=" << r << dendl;
Context *on_finish = m_on_finish;
m_on_finish = nullptr;
m_event_entry = nullptr;
m_in_progress = false;
m_snap_seqs_updated = false;
m_work_queue->queue(on_finish, r);
}
} // namespace journal
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_replayer::journal::EventPreprocessor<librbd::ImageCtx>;
| 5,959 | 27.792271 | 89 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/journal/EventPreprocessor.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_REPLAYER_EVENT_PREPROCESSOR_H
#define RBD_MIRROR_IMAGE_REPLAYER_EVENT_PREPROCESSOR_H
#include "include/int_types.h"
#include "librbd/journal/Types.h"
#include "librbd/journal/TypeTraits.h"
#include <map>
#include <string>
#include <boost/variant/static_visitor.hpp>
struct Context;
namespace journal { class Journaler; }
namespace librbd {
class ImageCtx;
namespace asio { struct ContextWQ; }
} // namespace librbd
namespace rbd {
namespace mirror {
namespace image_replayer {
namespace journal {
template <typename ImageCtxT = librbd::ImageCtx>
class EventPreprocessor {
public:
using Journaler = typename librbd::journal::TypeTraits<ImageCtxT>::Journaler;
using EventEntry = librbd::journal::EventEntry;
using MirrorPeerClientMeta = librbd::journal::MirrorPeerClientMeta;
static EventPreprocessor *create(ImageCtxT &local_image_ctx,
Journaler &remote_journaler,
const std::string &local_mirror_uuid,
MirrorPeerClientMeta *client_meta,
librbd::asio::ContextWQ *work_queue) {
return new EventPreprocessor(local_image_ctx, remote_journaler,
local_mirror_uuid, client_meta, work_queue);
}
static void destroy(EventPreprocessor* processor) {
delete processor;
}
EventPreprocessor(ImageCtxT &local_image_ctx, Journaler &remote_journaler,
const std::string &local_mirror_uuid,
MirrorPeerClientMeta *client_meta,
librbd::asio::ContextWQ *work_queue);
~EventPreprocessor();
bool is_required(const EventEntry &event_entry);
void preprocess(EventEntry *event_entry, Context *on_finish);
private:
/**
* @verbatim
*
* <start>
* |
* v (skip if not required)
* REFRESH_IMAGE
* |
* v (skip if not required)
* PREPROCESS_EVENT
* |
* v (skip if not required)
* UPDATE_CLIENT
*
* @endverbatim
*/
typedef std::map<uint64_t, uint64_t> SnapSeqs;
class PreprocessEventVisitor : public boost::static_visitor<int> {
public:
EventPreprocessor *event_preprocessor;
PreprocessEventVisitor(EventPreprocessor *event_preprocessor)
: event_preprocessor(event_preprocessor) {
}
template <typename T>
inline int operator()(T&) const {
return 0;
}
inline int operator()(librbd::journal::SnapRenameEvent &event) const {
return event_preprocessor->preprocess_snap_rename(event);
}
};
ImageCtxT &m_local_image_ctx;
Journaler &m_remote_journaler;
std::string m_local_mirror_uuid;
MirrorPeerClientMeta *m_client_meta;
librbd::asio::ContextWQ *m_work_queue;
bool m_in_progress = false;
EventEntry *m_event_entry = nullptr;
Context *m_on_finish = nullptr;
SnapSeqs m_snap_seqs;
bool m_snap_seqs_updated = false;
bool prune_snap_map(SnapSeqs *snap_seqs);
void refresh_image();
void handle_refresh_image(int r);
void preprocess_event();
int preprocess_snap_rename(librbd::journal::SnapRenameEvent &event);
void update_client();
void handle_update_client(int r);
void finish(int r);
};
} // namespace journal
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_replayer::journal::EventPreprocessor<librbd::ImageCtx>;
#endif // RBD_MIRROR_IMAGE_REPLAYER_EVENT_PREPROCESSOR_H
| 3,562 | 26.835938 | 96 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/journal/PrepareReplayRequest.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "PrepareReplayRequest.h"
#include "common/debug.h"
#include "common/dout.h"
#include "common/errno.h"
#include "journal/Journaler.h"
#include "librbd/ImageCtx.h"
#include "librbd/Journal.h"
#include "librbd/Utils.h"
#include "tools/rbd_mirror/ProgressContext.h"
#include "tools/rbd_mirror/image_replayer/journal/StateBuilder.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_replayer::journal::" \
<< "PrepareReplayRequest: " << this << " " \
<< __func__ << ": "
namespace rbd {
namespace mirror {
namespace image_replayer {
namespace journal {
using librbd::util::create_context_callback;
template <typename I>
void PrepareReplayRequest<I>::send() {
*m_resync_requested = false;
*m_syncing = false;
if (m_state_builder->local_image_id !=
m_state_builder->remote_client_meta.image_id) {
// somehow our local image has a different image id than the image id
// registered in the remote image
derr << "split-brain detected: local_image_id="
<< m_state_builder->local_image_id << ", "
<< "registered local_image_id="
<< m_state_builder->remote_client_meta.image_id << dendl;
finish(-EEXIST);
return;
}
std::shared_lock image_locker(m_state_builder->local_image_ctx->image_lock);
if (m_state_builder->local_image_ctx->journal == nullptr) {
image_locker.unlock();
derr << "local image does not support journaling" << dendl;
finish(-EINVAL);
return;
}
int r = m_state_builder->local_image_ctx->journal->is_resync_requested(
m_resync_requested);
if (r < 0) {
image_locker.unlock();
derr << "failed to check if a resync was requested" << dendl;
finish(r);
return;
}
m_local_tag_tid = m_state_builder->local_image_ctx->journal->get_tag_tid();
m_local_tag_data = m_state_builder->local_image_ctx->journal->get_tag_data();
dout(10) << "local tag=" << m_local_tag_tid << ", "
<< "local tag data=" << m_local_tag_data << dendl;
image_locker.unlock();
if (*m_resync_requested) {
finish(0);
return;
} else if (m_state_builder->remote_client_meta.state ==
librbd::journal::MIRROR_PEER_STATE_SYNCING &&
m_local_tag_data.mirror_uuid ==
m_state_builder->remote_mirror_uuid) {
// if the initial sync hasn't completed, we cannot replay
*m_syncing = true;
finish(0);
return;
}
update_client_state();
}
template <typename I>
void PrepareReplayRequest<I>::update_client_state() {
if (m_state_builder->remote_client_meta.state !=
librbd::journal::MIRROR_PEER_STATE_SYNCING ||
m_local_tag_data.mirror_uuid == m_state_builder->remote_mirror_uuid) {
get_remote_tag_class();
return;
}
// our local image is not primary, is flagged as syncing on the remote side,
// but is no longer tied to the remote -- this implies we were forced
// promoted and then demoted at some point
dout(15) << dendl;
update_progress("UPDATE_CLIENT_STATE");
auto client_meta = m_state_builder->remote_client_meta;
client_meta.state = librbd::journal::MIRROR_PEER_STATE_REPLAYING;
librbd::journal::ClientData client_data(client_meta);
bufferlist data_bl;
encode(client_data, data_bl);
auto ctx = create_context_callback<
PrepareReplayRequest<I>,
&PrepareReplayRequest<I>::handle_update_client_state>(this);
m_state_builder->remote_journaler->update_client(data_bl, ctx);
}
template <typename I>
void PrepareReplayRequest<I>::handle_update_client_state(int r) {
dout(15) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to update client: " << cpp_strerror(r) << dendl;
finish(r);
return;
}
m_state_builder->remote_client_meta.state =
librbd::journal::MIRROR_PEER_STATE_REPLAYING;
get_remote_tag_class();
}
template <typename I>
void PrepareReplayRequest<I>::get_remote_tag_class() {
dout(10) << dendl;
update_progress("GET_REMOTE_TAG_CLASS");
auto ctx = create_context_callback<
PrepareReplayRequest<I>,
&PrepareReplayRequest<I>::handle_get_remote_tag_class>(this);
m_state_builder->remote_journaler->get_client(
librbd::Journal<>::IMAGE_CLIENT_ID, &m_client, ctx);
}
template <typename I>
void PrepareReplayRequest<I>::handle_get_remote_tag_class(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to retrieve remote client: " << cpp_strerror(r) << dendl;
finish(r);
return;
}
librbd::journal::ClientData client_data;
auto it = m_client.data.cbegin();
try {
decode(client_data, it);
} catch (const buffer::error &err) {
derr << "failed to decode remote client meta data: " << err.what()
<< dendl;
finish(-EBADMSG);
return;
}
librbd::journal::ImageClientMeta *client_meta =
boost::get<librbd::journal::ImageClientMeta>(&client_data.client_meta);
if (client_meta == nullptr) {
derr << "unknown remote client registration" << dendl;
finish(-EINVAL);
return;
}
m_remote_tag_class = client_meta->tag_class;
dout(10) << "remote tag class=" << m_remote_tag_class << dendl;
get_remote_tags();
}
template <typename I>
void PrepareReplayRequest<I>::get_remote_tags() {
dout(10) << dendl;
update_progress("GET_REMOTE_TAGS");
auto ctx = create_context_callback<
PrepareReplayRequest<I>,
&PrepareReplayRequest<I>::handle_get_remote_tags>(this);
m_state_builder->remote_journaler->get_tags(m_remote_tag_class,
&m_remote_tags, ctx);
}
template <typename I>
void PrepareReplayRequest<I>::handle_get_remote_tags(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to retrieve remote tags: " << cpp_strerror(r) << dendl;
finish(r);
return;
}
// At this point, the local image was existing, non-primary, and replaying;
// and the remote image is primary. Attempt to link the local image's most
// recent tag to the remote image's tag chain.
bool remote_tag_data_valid = false;
librbd::journal::TagData remote_tag_data;
boost::optional<uint64_t> remote_orphan_tag_tid =
boost::make_optional<uint64_t>(false, 0U);
bool reconnect_orphan = false;
// decode the remote tags
for (auto &remote_tag : m_remote_tags) {
if (m_local_tag_data.predecessor.commit_valid &&
m_local_tag_data.predecessor.mirror_uuid ==
m_state_builder->remote_mirror_uuid &&
m_local_tag_data.predecessor.tag_tid > remote_tag.tid) {
dout(10) << "skipping processed predecessor remote tag "
<< remote_tag.tid << dendl;
continue;
}
try {
auto it = remote_tag.data.cbegin();
decode(remote_tag_data, it);
remote_tag_data_valid = true;
} catch (const buffer::error &err) {
derr << "failed to decode remote tag " << remote_tag.tid << ": "
<< err.what() << dendl;
finish(-EBADMSG);
return;
}
dout(10) << "decoded remote tag " << remote_tag.tid << ": "
<< remote_tag_data << dendl;
if (!m_local_tag_data.predecessor.commit_valid) {
// newly synced local image (no predecessor) replays from the first tag
if (remote_tag_data.mirror_uuid != librbd::Journal<>::LOCAL_MIRROR_UUID) {
dout(10) << "skipping non-primary remote tag" << dendl;
continue;
}
dout(10) << "using initial primary remote tag" << dendl;
break;
}
if (m_local_tag_data.mirror_uuid == librbd::Journal<>::ORPHAN_MIRROR_UUID) {
// demotion last available local epoch
if (remote_tag_data.mirror_uuid == m_local_tag_data.mirror_uuid &&
remote_tag_data.predecessor.commit_valid &&
remote_tag_data.predecessor.tag_tid ==
m_local_tag_data.predecessor.tag_tid) {
// demotion matches remote epoch
if (remote_tag_data.predecessor.mirror_uuid == m_local_mirror_uuid &&
m_local_tag_data.predecessor.mirror_uuid ==
librbd::Journal<>::LOCAL_MIRROR_UUID) {
// local demoted and remote has matching event
dout(10) << "found matching local demotion tag" << dendl;
remote_orphan_tag_tid = remote_tag.tid;
continue;
}
if (m_local_tag_data.predecessor.mirror_uuid ==
m_state_builder->remote_mirror_uuid &&
remote_tag_data.predecessor.mirror_uuid ==
librbd::Journal<>::LOCAL_MIRROR_UUID) {
// remote demoted and local has matching event
dout(10) << "found matching remote demotion tag" << dendl;
remote_orphan_tag_tid = remote_tag.tid;
continue;
}
}
if (remote_tag_data.mirror_uuid == librbd::Journal<>::LOCAL_MIRROR_UUID &&
remote_tag_data.predecessor.mirror_uuid ==
librbd::Journal<>::ORPHAN_MIRROR_UUID &&
remote_tag_data.predecessor.commit_valid && remote_orphan_tag_tid &&
remote_tag_data.predecessor.tag_tid == *remote_orphan_tag_tid) {
// remote promotion tag chained to remote/local demotion tag
dout(10) << "found chained remote promotion tag" << dendl;
reconnect_orphan = true;
break;
}
// promotion must follow demotion
remote_orphan_tag_tid = boost::none;
}
}
if (remote_tag_data_valid &&
m_local_tag_data.mirror_uuid == m_state_builder->remote_mirror_uuid) {
dout(10) << "local image is in clean replay state" << dendl;
} else if (reconnect_orphan) {
dout(10) << "remote image was demoted/promoted" << dendl;
} else {
derr << "split-brain detected -- skipping image replay" << dendl;
finish(-EEXIST);
return;
}
finish(0);
}
template <typename I>
void PrepareReplayRequest<I>::update_progress(const std::string &description) {
dout(10) << description << dendl;
if (m_progress_ctx != nullptr) {
m_progress_ctx->update_progress(description);
}
}
} // namespace journal
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_replayer::journal::PrepareReplayRequest<librbd::ImageCtx>;
| 10,297 | 31.485804 | 92 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/journal/PrepareReplayRequest.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_REPLAYER_JOURNAL_PREPARE_REPLAY_REQUEST_H
#define RBD_MIRROR_IMAGE_REPLAYER_JOURNAL_PREPARE_REPLAY_REQUEST_H
#include "include/int_types.h"
#include "cls/journal/cls_journal_types.h"
#include "librbd/journal/Types.h"
#include "librbd/mirror/Types.h"
#include "tools/rbd_mirror/BaseRequest.h"
#include <list>
#include <string>
struct Context;
namespace librbd { struct ImageCtx; }
namespace rbd {
namespace mirror {
class ProgressContext;
namespace image_replayer {
namespace journal {
template <typename> class StateBuilder;
template <typename ImageCtxT>
class PrepareReplayRequest : public BaseRequest {
public:
static PrepareReplayRequest* create(
const std::string& local_mirror_uuid,
ProgressContext* progress_ctx,
StateBuilder<ImageCtxT>* state_builder,
bool* resync_requested,
bool* syncing,
Context* on_finish) {
return new PrepareReplayRequest(
local_mirror_uuid, progress_ctx, state_builder, resync_requested,
syncing, on_finish);
}
PrepareReplayRequest(
const std::string& local_mirror_uuid,
ProgressContext* progress_ctx,
StateBuilder<ImageCtxT>* state_builder,
bool* resync_requested,
bool* syncing,
Context* on_finish)
: BaseRequest(on_finish),
m_local_mirror_uuid(local_mirror_uuid),
m_progress_ctx(progress_ctx),
m_state_builder(state_builder),
m_resync_requested(resync_requested),
m_syncing(syncing) {
}
void send() override;
private:
/**
* @verbatim
*
* <start>
* |
* v
* UPDATE_CLIENT_STATE
* |
* v
* GET_REMOTE_TAG_CLASS
* |
* v
* GET_REMOTE_TAGS
* |
* v
* <finish>
*
* @endverbatim
*/
typedef std::list<cls::journal::Tag> Tags;
std::string m_local_mirror_uuid;
ProgressContext* m_progress_ctx;
StateBuilder<ImageCtxT>* m_state_builder;
bool* m_resync_requested;
bool* m_syncing;
uint64_t m_local_tag_tid = 0;
librbd::journal::TagData m_local_tag_data;
uint64_t m_remote_tag_class = 0;
Tags m_remote_tags;
cls::journal::Client m_client;
void update_client_state();
void handle_update_client_state(int r);
void get_remote_tag_class();
void handle_get_remote_tag_class(int r);
void get_remote_tags();
void handle_get_remote_tags(int r);
void update_progress(const std::string& description);
};
} // namespace journal
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_replayer::journal::PrepareReplayRequest<librbd::ImageCtx>;
#endif // RBD_MIRROR_IMAGE_REPLAYER_JOURNAL_PREPARE_REPLAY_REQUEST_H
| 2,768 | 22.87069 | 99 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/journal/ReplayStatusFormatter.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "ReplayStatusFormatter.h"
#include "common/debug.h"
#include "common/dout.h"
#include "common/errno.h"
#include "journal/Journaler.h"
#include "json_spirit/json_spirit.h"
#include "librbd/ImageCtx.h"
#include "librbd/Journal.h"
#include "librbd/Utils.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_replayer::journal::" \
<< "ReplayStatusFormatter: " << this << " " \
<< __func__ << ": "
namespace rbd {
namespace mirror {
namespace image_replayer {
namespace journal {
using librbd::util::unique_lock_name;
namespace {
double round_to_two_places(double value) {
return abs(round(value * 100) / 100);
}
json_spirit::mObject to_json_object(
const cls::journal::ObjectPosition& position) {
json_spirit::mObject object;
if (position != cls::journal::ObjectPosition{}) {
object["object_number"] = position.object_number;
object["tag_tid"] = position.tag_tid;
object["entry_tid"] = position.entry_tid;
}
return object;
}
} // anonymous namespace
template <typename I>
ReplayStatusFormatter<I>::ReplayStatusFormatter(Journaler *journaler,
const std::string &mirror_uuid)
: m_journaler(journaler),
m_mirror_uuid(mirror_uuid),
m_lock(ceph::make_mutex(unique_lock_name("ReplayStatusFormatter::m_lock", this))) {
}
template <typename I>
void ReplayStatusFormatter<I>::handle_entry_processed(uint32_t bytes) {
dout(20) << dendl;
m_bytes_per_second(bytes);
m_entries_per_second(1);
}
template <typename I>
bool ReplayStatusFormatter<I>::get_or_send_update(std::string *description,
Context *on_finish) {
dout(20) << dendl;
bool in_progress = false;
{
std::lock_guard locker{m_lock};
if (m_on_finish) {
in_progress = true;
} else {
m_on_finish = on_finish;
}
}
if (in_progress) {
dout(10) << "previous request is still in progress, ignoring" << dendl;
on_finish->complete(-EAGAIN);
return false;
}
m_master_position = cls::journal::ObjectPosition();
m_mirror_position = cls::journal::ObjectPosition();
cls::journal::Client master_client, mirror_client;
int r;
r = m_journaler->get_cached_client(librbd::Journal<>::IMAGE_CLIENT_ID,
&master_client);
if (r < 0) {
derr << "error retrieving registered master client: "
<< cpp_strerror(r) << dendl;
} else {
r = m_journaler->get_cached_client(m_mirror_uuid, &mirror_client);
if (r < 0) {
derr << "error retrieving registered mirror client: "
<< cpp_strerror(r) << dendl;
}
}
if (!master_client.commit_position.object_positions.empty()) {
m_master_position =
*(master_client.commit_position.object_positions.begin());
}
if (!mirror_client.commit_position.object_positions.empty()) {
m_mirror_position =
*(mirror_client.commit_position.object_positions.begin());
}
if (!calculate_behind_master_or_send_update()) {
dout(20) << "need to update tag cache" << dendl;
return false;
}
format(description);
{
std::lock_guard locker{m_lock};
ceph_assert(m_on_finish == on_finish);
m_on_finish = nullptr;
}
on_finish->complete(-EEXIST);
return true;
}
template <typename I>
bool ReplayStatusFormatter<I>::calculate_behind_master_or_send_update() {
dout(20) << "m_master_position=" << m_master_position
<< ", m_mirror_position=" << m_mirror_position << dendl;
m_entries_behind_master = 0;
if (m_master_position == cls::journal::ObjectPosition() ||
m_master_position.tag_tid < m_mirror_position.tag_tid) {
return true;
}
cls::journal::ObjectPosition master = m_master_position;
uint64_t mirror_tag_tid = m_mirror_position.tag_tid;
while (master.tag_tid > mirror_tag_tid) {
auto tag_it = m_tag_cache.find(master.tag_tid);
if (tag_it == m_tag_cache.end()) {
send_update_tag_cache(master.tag_tid, mirror_tag_tid);
return false;
}
librbd::journal::TagData &tag_data = tag_it->second;
m_entries_behind_master += master.entry_tid;
master = {0, tag_data.predecessor.tag_tid, tag_data.predecessor.entry_tid};
}
if (master.tag_tid == mirror_tag_tid &&
master.entry_tid > m_mirror_position.entry_tid) {
m_entries_behind_master += master.entry_tid - m_mirror_position.entry_tid;
}
dout(20) << "clearing tags not needed any more (below mirror position)"
<< dendl;
uint64_t tag_tid = mirror_tag_tid;
size_t old_size = m_tag_cache.size();
while (tag_tid != 0) {
auto tag_it = m_tag_cache.find(tag_tid);
if (tag_it == m_tag_cache.end()) {
break;
}
librbd::journal::TagData &tag_data = tag_it->second;
dout(20) << "erasing tag " << tag_data << "for tag_tid " << tag_tid
<< dendl;
tag_tid = tag_data.predecessor.tag_tid;
m_tag_cache.erase(tag_it);
}
dout(20) << old_size - m_tag_cache.size() << " entries cleared" << dendl;
return true;
}
template <typename I>
void ReplayStatusFormatter<I>::send_update_tag_cache(uint64_t master_tag_tid,
uint64_t mirror_tag_tid) {
if (master_tag_tid <= mirror_tag_tid ||
m_tag_cache.find(master_tag_tid) != m_tag_cache.end()) {
Context *on_finish = nullptr;
{
std::lock_guard locker{m_lock};
std::swap(m_on_finish, on_finish);
}
ceph_assert(on_finish);
on_finish->complete(0);
return;
}
dout(20) << "master_tag_tid=" << master_tag_tid << ", mirror_tag_tid="
<< mirror_tag_tid << dendl;
auto ctx = new LambdaContext(
[this, master_tag_tid, mirror_tag_tid](int r) {
handle_update_tag_cache(master_tag_tid, mirror_tag_tid, r);
});
m_journaler->get_tag(master_tag_tid, &m_tag, ctx);
}
template <typename I>
void ReplayStatusFormatter<I>::handle_update_tag_cache(uint64_t master_tag_tid,
uint64_t mirror_tag_tid,
int r) {
librbd::journal::TagData tag_data;
if (r < 0) {
derr << "error retrieving tag " << master_tag_tid << ": " << cpp_strerror(r)
<< dendl;
} else {
dout(20) << "retrieved tag " << master_tag_tid << ": " << m_tag << dendl;
auto it = m_tag.data.cbegin();
try {
decode(tag_data, it);
} catch (const buffer::error &err) {
derr << "error decoding tag " << master_tag_tid << ": " << err.what()
<< dendl;
}
}
if (tag_data.predecessor.mirror_uuid !=
librbd::Journal<>::LOCAL_MIRROR_UUID &&
tag_data.predecessor.mirror_uuid !=
librbd::Journal<>::ORPHAN_MIRROR_UUID) {
dout(20) << "hit remote image non-primary epoch" << dendl;
tag_data.predecessor = {};
}
dout(20) << "decoded tag " << master_tag_tid << ": " << tag_data << dendl;
m_tag_cache[master_tag_tid] = tag_data;
send_update_tag_cache(tag_data.predecessor.tag_tid, mirror_tag_tid);
}
template <typename I>
void ReplayStatusFormatter<I>::format(std::string *description) {
dout(20) << "m_master_position=" << m_master_position
<< ", m_mirror_position=" << m_mirror_position
<< ", m_entries_behind_master=" << m_entries_behind_master << dendl;
json_spirit::mObject root_obj;
root_obj["primary_position"] = to_json_object(m_master_position);
root_obj["non_primary_position"] = to_json_object(m_mirror_position);
root_obj["entries_behind_primary"] = (
m_entries_behind_master > 0 ? m_entries_behind_master : 0);
m_bytes_per_second(0);
root_obj["bytes_per_second"] = round_to_two_places(
m_bytes_per_second.get_average());
m_entries_per_second(0);
auto entries_per_second = m_entries_per_second.get_average();
root_obj["entries_per_second"] = round_to_two_places(entries_per_second);
if (m_entries_behind_master > 0 && entries_per_second > 0) {
std::uint64_t seconds_until_synced = round_to_two_places(
m_entries_behind_master / entries_per_second);
if (seconds_until_synced >= std::numeric_limits<uint64_t>::max()) {
seconds_until_synced = std::numeric_limits<uint64_t>::max();
}
root_obj["seconds_until_synced"] = seconds_until_synced;
}
*description = json_spirit::write(
root_obj, json_spirit::remove_trailing_zeros);
}
} // namespace journal
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_replayer::journal::ReplayStatusFormatter<librbd::ImageCtx>;
| 8,484 | 28.77193 | 93 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/journal/ReplayStatusFormatter.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_REPLAYER_REPLAY_STATUS_FORMATTER_H
#define RBD_MIRROR_IMAGE_REPLAYER_REPLAY_STATUS_FORMATTER_H
#include "include/Context.h"
#include "common/ceph_mutex.h"
#include "cls/journal/cls_journal_types.h"
#include "librbd/journal/Types.h"
#include "librbd/journal/TypeTraits.h"
#include "tools/rbd_mirror/image_replayer/TimeRollingMean.h"
namespace journal { class Journaler; }
namespace librbd { class ImageCtx; }
namespace rbd {
namespace mirror {
namespace image_replayer {
namespace journal {
template <typename ImageCtxT = librbd::ImageCtx>
class ReplayStatusFormatter {
public:
typedef typename librbd::journal::TypeTraits<ImageCtxT>::Journaler Journaler;
static ReplayStatusFormatter* create(Journaler *journaler,
const std::string &mirror_uuid) {
return new ReplayStatusFormatter(journaler, mirror_uuid);
}
static void destroy(ReplayStatusFormatter* formatter) {
delete formatter;
}
ReplayStatusFormatter(Journaler *journaler, const std::string &mirror_uuid);
void handle_entry_processed(uint32_t bytes);
bool get_or_send_update(std::string *description, Context *on_finish);
private:
Journaler *m_journaler;
std::string m_mirror_uuid;
ceph::mutex m_lock;
Context *m_on_finish = nullptr;
cls::journal::ObjectPosition m_master_position;
cls::journal::ObjectPosition m_mirror_position;
int64_t m_entries_behind_master = 0;
cls::journal::Tag m_tag;
std::map<uint64_t, librbd::journal::TagData> m_tag_cache;
TimeRollingMean m_bytes_per_second;
TimeRollingMean m_entries_per_second;
bool calculate_behind_master_or_send_update();
void send_update_tag_cache(uint64_t master_tag_tid, uint64_t mirror_tag_tid);
void handle_update_tag_cache(uint64_t master_tag_tid, uint64_t mirror_tag_tid,
int r);
void format(std::string *description);
};
} // namespace journal
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_replayer::journal::ReplayStatusFormatter<librbd::ImageCtx>;
#endif // RBD_MIRROR_IMAGE_REPLAYER_REPLAY_STATUS_FORMATTER_H
| 2,204 | 30.056338 | 100 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/journal/Replayer.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "Replayer.h"
#include "common/debug.h"
#include "common/errno.h"
#include "common/perf_counters.h"
#include "common/perf_counters_key.h"
#include "common/Timer.h"
#include "librbd/Journal.h"
#include "librbd/Utils.h"
#include "librbd/asio/ContextWQ.h"
#include "librbd/journal/Replay.h"
#include "journal/Journaler.h"
#include "journal/JournalMetadataListener.h"
#include "journal/ReplayHandler.h"
#include "tools/rbd_mirror/Threads.h"
#include "tools/rbd_mirror/Types.h"
#include "tools/rbd_mirror/image_replayer/CloseImageRequest.h"
#include "tools/rbd_mirror/image_replayer/ReplayerListener.h"
#include "tools/rbd_mirror/image_replayer/Utils.h"
#include "tools/rbd_mirror/image_replayer/journal/EventPreprocessor.h"
#include "tools/rbd_mirror/image_replayer/journal/ReplayStatusFormatter.h"
#include "tools/rbd_mirror/image_replayer/journal/StateBuilder.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_replayer::journal::" \
<< "Replayer: " << this << " " << __func__ << ": "
extern PerfCounters *g_journal_perf_counters;
namespace rbd {
namespace mirror {
namespace image_replayer {
namespace journal {
namespace {
uint32_t calculate_replay_delay(const utime_t &event_time,
int mirroring_replay_delay) {
if (mirroring_replay_delay <= 0) {
return 0;
}
utime_t now = ceph_clock_now();
if (event_time + mirroring_replay_delay <= now) {
return 0;
}
// ensure it is rounded up when converting to integer
return (event_time + mirroring_replay_delay - now) + 1;
}
} // anonymous namespace
using librbd::util::create_async_context_callback;
using librbd::util::create_context_callback;
template <typename I>
struct Replayer<I>::C_ReplayCommitted : public Context {
Replayer* replayer;
ReplayEntry replay_entry;
uint64_t replay_bytes;
utime_t replay_start_time;
C_ReplayCommitted(Replayer* replayer, ReplayEntry &&replay_entry,
uint64_t replay_bytes, const utime_t &replay_start_time)
: replayer(replayer), replay_entry(std::move(replay_entry)),
replay_bytes(replay_bytes), replay_start_time(replay_start_time) {
}
void finish(int r) override {
replayer->handle_process_entry_safe(replay_entry, replay_bytes,
replay_start_time, r);
}
};
template <typename I>
struct Replayer<I>::RemoteJournalerListener
: public ::journal::JournalMetadataListener {
Replayer* replayer;
RemoteJournalerListener(Replayer* replayer) : replayer(replayer) {}
void handle_update(::journal::JournalMetadata*) override {
auto ctx = new C_TrackedOp(
replayer->m_in_flight_op_tracker,
new LambdaContext([this](int r) {
replayer->handle_remote_journal_metadata_updated();
}));
replayer->m_threads->work_queue->queue(ctx, 0);
}
};
template <typename I>
struct Replayer<I>::RemoteReplayHandler : public ::journal::ReplayHandler {
Replayer* replayer;
RemoteReplayHandler(Replayer* replayer) : replayer(replayer) {}
~RemoteReplayHandler() override {};
void handle_entries_available() override {
replayer->handle_replay_ready();
}
void handle_complete(int r) override {
std::string error;
if (r == -ENOMEM) {
error = "not enough memory in autotune cache";
} else if (r < 0) {
error = "replay completed with error: " + cpp_strerror(r);
}
replayer->handle_replay_complete(r, error);
}
};
template <typename I>
struct Replayer<I>::LocalJournalListener
: public librbd::journal::Listener {
Replayer* replayer;
LocalJournalListener(Replayer* replayer) : replayer(replayer) {
}
void handle_close() override {
replayer->handle_replay_complete(0, "");
}
void handle_promoted() override {
replayer->handle_replay_complete(0, "force promoted");
}
void handle_resync() override {
replayer->handle_resync_image();
}
};
template <typename I>
Replayer<I>::Replayer(
Threads<I>* threads,
const std::string& local_mirror_uuid,
StateBuilder<I>* state_builder,
ReplayerListener* replayer_listener)
: m_threads(threads),
m_local_mirror_uuid(local_mirror_uuid),
m_state_builder(state_builder),
m_replayer_listener(replayer_listener),
m_lock(ceph::make_mutex(librbd::util::unique_lock_name(
"rbd::mirror::image_replayer::journal::Replayer", this))) {
dout(10) << dendl;
}
template <typename I>
Replayer<I>::~Replayer() {
dout(10) << dendl;
{
std::unique_lock locker{m_lock};
unregister_perf_counters();
}
ceph_assert(m_remote_listener == nullptr);
ceph_assert(m_local_journal_listener == nullptr);
ceph_assert(m_local_journal_replay == nullptr);
ceph_assert(m_remote_replay_handler == nullptr);
ceph_assert(m_event_preprocessor == nullptr);
ceph_assert(m_replay_status_formatter == nullptr);
ceph_assert(m_delayed_preprocess_task == nullptr);
ceph_assert(m_flush_local_replay_task == nullptr);
ceph_assert(m_state_builder->local_image_ctx == nullptr);
}
template <typename I>
void Replayer<I>::init(Context* on_finish) {
dout(10) << dendl;
{
auto local_image_ctx = m_state_builder->local_image_ctx;
std::shared_lock image_locker{local_image_ctx->image_lock};
m_image_spec = util::compute_image_spec(local_image_ctx->md_ctx,
local_image_ctx->name);
}
{
std::unique_lock locker{m_lock};
register_perf_counters();
}
ceph_assert(m_on_init_shutdown == nullptr);
m_on_init_shutdown = on_finish;
init_remote_journaler();
}
template <typename I>
void Replayer<I>::shut_down(Context* on_finish) {
dout(10) << dendl;
std::unique_lock locker{m_lock};
ceph_assert(m_on_init_shutdown == nullptr);
m_on_init_shutdown = on_finish;
if (m_state == STATE_INIT) {
// raced with the last piece of the init state machine
return;
} else if (m_state == STATE_REPLAYING) {
m_state = STATE_COMPLETE;
}
// if shutting down due to an error notification, we don't
// need to propagate the same error again
m_error_code = 0;
m_error_description = "";
cancel_delayed_preprocess_task();
cancel_flush_local_replay_task();
wait_for_flush();
}
template <typename I>
void Replayer<I>::flush(Context* on_finish) {
dout(10) << dendl;
flush_local_replay(new C_TrackedOp(m_in_flight_op_tracker, on_finish));
}
template <typename I>
bool Replayer<I>::get_replay_status(std::string* description,
Context* on_finish) {
dout(10) << dendl;
std::unique_lock locker{m_lock};
if (m_replay_status_formatter == nullptr) {
derr << "replay not running" << dendl;
locker.unlock();
on_finish->complete(-EAGAIN);
return false;
}
on_finish = new C_TrackedOp(m_in_flight_op_tracker, on_finish);
return m_replay_status_formatter->get_or_send_update(description,
on_finish);
}
template <typename I>
void Replayer<I>::init_remote_journaler() {
dout(10) << dendl;
Context *ctx = create_context_callback<
Replayer, &Replayer<I>::handle_init_remote_journaler>(this);
m_state_builder->remote_journaler->init(ctx);
}
template <typename I>
void Replayer<I>::handle_init_remote_journaler(int r) {
dout(10) << "r=" << r << dendl;
std::unique_lock locker{m_lock};
if (r < 0) {
derr << "failed to initialize remote journal: " << cpp_strerror(r) << dendl;
handle_replay_complete(locker, r, "error initializing remote journal");
close_local_image();
return;
}
// listen for metadata updates to check for disconnect events
ceph_assert(m_remote_listener == nullptr);
m_remote_listener = new RemoteJournalerListener(this);
m_state_builder->remote_journaler->add_listener(m_remote_listener);
cls::journal::Client remote_client;
r = m_state_builder->remote_journaler->get_cached_client(m_local_mirror_uuid,
&remote_client);
if (r < 0) {
derr << "error retrieving remote journal client: " << cpp_strerror(r)
<< dendl;
handle_replay_complete(locker, r, "error retrieving remote journal client");
close_local_image();
return;
}
std::string error;
r = validate_remote_client_state(remote_client,
&m_state_builder->remote_client_meta,
&m_resync_requested, &error);
if (r < 0) {
handle_replay_complete(locker, r, error);
close_local_image();
return;
}
start_external_replay(locker);
}
template <typename I>
void Replayer<I>::start_external_replay(std::unique_lock<ceph::mutex>& locker) {
dout(10) << dendl;
auto local_image_ctx = m_state_builder->local_image_ctx;
std::shared_lock local_image_locker{local_image_ctx->image_lock};
ceph_assert(m_local_journal == nullptr);
m_local_journal = local_image_ctx->journal;
if (m_local_journal == nullptr) {
local_image_locker.unlock();
derr << "local image journal closed" << dendl;
handle_replay_complete(locker, -EINVAL, "error accessing local journal");
close_local_image();
return;
}
// safe to hold pointer to journal after external playback starts
Context *start_ctx = create_context_callback<
Replayer, &Replayer<I>::handle_start_external_replay>(this);
m_local_journal->start_external_replay(&m_local_journal_replay, start_ctx);
}
template <typename I>
void Replayer<I>::handle_start_external_replay(int r) {
dout(10) << "r=" << r << dendl;
std::unique_lock locker{m_lock};
if (r < 0) {
ceph_assert(m_local_journal_replay == nullptr);
derr << "error starting external replay on local image "
<< m_state_builder->local_image_ctx->id << ": "
<< cpp_strerror(r) << dendl;
handle_replay_complete(locker, r, "error starting replay on local image");
close_local_image();
return;
}
if (!notify_init_complete(locker)) {
return;
}
m_state = STATE_REPLAYING;
// check for resync/promotion state after adding listener
if (!add_local_journal_listener(locker)) {
return;
}
// start remote journal replay
m_event_preprocessor = EventPreprocessor<I>::create(
*m_state_builder->local_image_ctx, *m_state_builder->remote_journaler,
m_local_mirror_uuid, &m_state_builder->remote_client_meta,
m_threads->work_queue);
m_replay_status_formatter = ReplayStatusFormatter<I>::create(
m_state_builder->remote_journaler, m_local_mirror_uuid);
auto cct = static_cast<CephContext *>(m_state_builder->local_image_ctx->cct);
double poll_seconds = cct->_conf.get_val<double>(
"rbd_mirror_journal_poll_age");
m_remote_replay_handler = new RemoteReplayHandler(this);
m_state_builder->remote_journaler->start_live_replay(m_remote_replay_handler,
poll_seconds);
notify_status_updated();
}
template <typename I>
bool Replayer<I>::add_local_journal_listener(
std::unique_lock<ceph::mutex>& locker) {
dout(10) << dendl;
// listen for promotion and resync requests against local journal
ceph_assert(m_local_journal_listener == nullptr);
m_local_journal_listener = new LocalJournalListener(this);
m_local_journal->add_listener(m_local_journal_listener);
// verify that the local image wasn't force-promoted and that a resync hasn't
// been requested now that we are listening for events
if (m_local_journal->is_tag_owner()) {
dout(10) << "local image force-promoted" << dendl;
handle_replay_complete(locker, 0, "force promoted");
return false;
}
bool resync_requested = false;
int r = m_local_journal->is_resync_requested(&resync_requested);
if (r < 0) {
dout(10) << "failed to determine resync state: " << cpp_strerror(r)
<< dendl;
handle_replay_complete(locker, r, "error parsing resync state");
return false;
} else if (resync_requested) {
dout(10) << "local image resync requested" << dendl;
handle_replay_complete(locker, 0, "resync requested");
return false;
}
return true;
}
template <typename I>
bool Replayer<I>::notify_init_complete(std::unique_lock<ceph::mutex>& locker) {
dout(10) << dendl;
ceph_assert(ceph_mutex_is_locked_by_me(m_lock));
ceph_assert(m_state == STATE_INIT);
// notify that init has completed
Context *on_finish = nullptr;
std::swap(m_on_init_shutdown, on_finish);
locker.unlock();
on_finish->complete(0);
locker.lock();
if (m_on_init_shutdown != nullptr) {
// shut down requested after we notified init complete but before we
// grabbed the lock
close_local_image();
return false;
}
return true;
}
template <typename I>
void Replayer<I>::wait_for_flush() {
ceph_assert(ceph_mutex_is_locked_by_me(m_lock));
// ensure that we don't have two concurrent local journal replay shut downs
dout(10) << dendl;
auto ctx = create_async_context_callback(
m_threads->work_queue, create_context_callback<
Replayer<I>, &Replayer<I>::handle_wait_for_flush>(this));
m_flush_tracker.wait_for_ops(ctx);
}
template <typename I>
void Replayer<I>::handle_wait_for_flush(int r) {
dout(10) << "r=" << r << dendl;
shut_down_local_journal_replay();
}
template <typename I>
void Replayer<I>::shut_down_local_journal_replay() {
std::unique_lock locker{m_lock};
if (m_local_journal_replay == nullptr) {
wait_for_event_replay();
return;
}
// It's required to stop the local journal replay state machine prior to
// waiting for the events to complete. This is to ensure that IO is properly
// flushed (it might be batched), wait for any running ops to complete, and
// to cancel any ops waiting for their associated OnFinish events.
dout(10) << dendl;
auto ctx = create_context_callback<
Replayer<I>, &Replayer<I>::handle_shut_down_local_journal_replay>(this);
m_local_journal_replay->shut_down(true, ctx);
}
template <typename I>
void Replayer<I>::handle_shut_down_local_journal_replay(int r) {
dout(10) << "r=" << r << dendl;
std::unique_lock locker{m_lock};
if (r < 0) {
derr << "error shutting down journal replay: " << cpp_strerror(r) << dendl;
handle_replay_error(r, "failed to shut down local journal replay");
}
wait_for_event_replay();
}
template <typename I>
void Replayer<I>::wait_for_event_replay() {
ceph_assert(ceph_mutex_is_locked_by_me(m_lock));
dout(10) << dendl;
auto ctx = create_async_context_callback(
m_threads->work_queue, create_context_callback<
Replayer<I>, &Replayer<I>::handle_wait_for_event_replay>(this));
m_event_replay_tracker.wait_for_ops(ctx);
}
template <typename I>
void Replayer<I>::handle_wait_for_event_replay(int r) {
dout(10) << "r=" << r << dendl;
std::unique_lock locker{m_lock};
close_local_image();
}
template <typename I>
void Replayer<I>::close_local_image() {
ceph_assert(ceph_mutex_is_locked_by_me(m_lock));
if (m_state_builder->local_image_ctx == nullptr) {
stop_remote_journaler_replay();
return;
}
dout(10) << dendl;
if (m_local_journal_listener != nullptr) {
// blocks if listener notification is in-progress
m_local_journal->remove_listener(m_local_journal_listener);
delete m_local_journal_listener;
m_local_journal_listener = nullptr;
}
if (m_local_journal_replay != nullptr) {
m_local_journal->stop_external_replay();
m_local_journal_replay = nullptr;
}
if (m_event_preprocessor != nullptr) {
image_replayer::journal::EventPreprocessor<I>::destroy(
m_event_preprocessor);
m_event_preprocessor = nullptr;
}
m_local_journal.reset();
// NOTE: it's important to ensure that the local image is fully
// closed before attempting to close the remote journal in
// case the remote cluster is unreachable
ceph_assert(m_state_builder->local_image_ctx != nullptr);
auto ctx = create_context_callback<
Replayer<I>, &Replayer<I>::handle_close_local_image>(this);
auto request = image_replayer::CloseImageRequest<I>::create(
&m_state_builder->local_image_ctx, ctx);
request->send();
}
template <typename I>
void Replayer<I>::handle_close_local_image(int r) {
dout(10) << "r=" << r << dendl;
std::unique_lock locker{m_lock};
if (r < 0) {
derr << "error closing local image: " << cpp_strerror(r) << dendl;
handle_replay_error(r, "failed to close local image");
}
ceph_assert(m_state_builder->local_image_ctx == nullptr);
stop_remote_journaler_replay();
}
template <typename I>
void Replayer<I>::stop_remote_journaler_replay() {
ceph_assert(ceph_mutex_is_locked_by_me(m_lock));
if (m_state_builder->remote_journaler == nullptr) {
wait_for_in_flight_ops();
return;
} else if (m_remote_replay_handler == nullptr) {
wait_for_in_flight_ops();
return;
}
dout(10) << dendl;
auto ctx = create_async_context_callback(
m_threads->work_queue, create_context_callback<
Replayer<I>, &Replayer<I>::handle_stop_remote_journaler_replay>(this));
m_state_builder->remote_journaler->stop_replay(ctx);
}
template <typename I>
void Replayer<I>::handle_stop_remote_journaler_replay(int r) {
dout(10) << "r=" << r << dendl;
std::unique_lock locker{m_lock};
if (r < 0) {
derr << "failed to stop remote journaler replay : " << cpp_strerror(r)
<< dendl;
handle_replay_error(r, "failed to stop remote journaler replay");
}
delete m_remote_replay_handler;
m_remote_replay_handler = nullptr;
wait_for_in_flight_ops();
}
template <typename I>
void Replayer<I>::wait_for_in_flight_ops() {
dout(10) << dendl;
if (m_remote_listener != nullptr) {
m_state_builder->remote_journaler->remove_listener(m_remote_listener);
delete m_remote_listener;
m_remote_listener = nullptr;
}
auto ctx = create_async_context_callback(
m_threads->work_queue, create_context_callback<
Replayer<I>, &Replayer<I>::handle_wait_for_in_flight_ops>(this));
m_in_flight_op_tracker.wait_for_ops(ctx);
}
template <typename I>
void Replayer<I>::handle_wait_for_in_flight_ops(int r) {
dout(10) << "r=" << r << dendl;
ReplayStatusFormatter<I>::destroy(m_replay_status_formatter);
m_replay_status_formatter = nullptr;
Context* on_init_shutdown = nullptr;
{
std::unique_lock locker{m_lock};
ceph_assert(m_on_init_shutdown != nullptr);
std::swap(m_on_init_shutdown, on_init_shutdown);
m_state = STATE_COMPLETE;
}
on_init_shutdown->complete(m_error_code);
}
template <typename I>
void Replayer<I>::handle_remote_journal_metadata_updated() {
dout(20) << dendl;
std::unique_lock locker{m_lock};
if (m_state != STATE_REPLAYING) {
return;
}
cls::journal::Client remote_client;
int r = m_state_builder->remote_journaler->get_cached_client(
m_local_mirror_uuid, &remote_client);
if (r < 0) {
derr << "failed to retrieve client: " << cpp_strerror(r) << dendl;
return;
}
librbd::journal::MirrorPeerClientMeta remote_client_meta;
std::string error;
r = validate_remote_client_state(remote_client, &remote_client_meta,
&m_resync_requested, &error);
if (r < 0) {
dout(0) << "client flagged disconnected, stopping image replay" << dendl;
handle_replay_complete(locker, r, error);
}
}
template <typename I>
void Replayer<I>::schedule_flush_local_replay_task() {
ceph_assert(ceph_mutex_is_locked_by_me(m_lock));
std::unique_lock timer_locker{m_threads->timer_lock};
if (m_state != STATE_REPLAYING || m_flush_local_replay_task != nullptr) {
return;
}
dout(15) << dendl;
m_flush_local_replay_task = create_async_context_callback(
m_threads->work_queue, create_context_callback<
Replayer<I>, &Replayer<I>::handle_flush_local_replay_task>(this));
m_threads->timer->add_event_after(30, m_flush_local_replay_task);
}
template <typename I>
void Replayer<I>::cancel_flush_local_replay_task() {
ceph_assert(ceph_mutex_is_locked_by_me(m_lock));
std::unique_lock timer_locker{m_threads->timer_lock};
if (m_flush_local_replay_task != nullptr) {
dout(10) << dendl;
m_threads->timer->cancel_event(m_flush_local_replay_task);
m_flush_local_replay_task = nullptr;
}
}
template <typename I>
void Replayer<I>::handle_flush_local_replay_task(int) {
dout(15) << dendl;
m_in_flight_op_tracker.start_op();
auto on_finish = new LambdaContext([this](int) {
std::unique_lock locker{m_lock};
{
std::unique_lock timer_locker{m_threads->timer_lock};
m_flush_local_replay_task = nullptr;
}
notify_status_updated();
m_in_flight_op_tracker.finish_op();
});
flush_local_replay(on_finish);
}
template <typename I>
void Replayer<I>::flush_local_replay(Context* on_flush) {
std::unique_lock locker{m_lock};
if (m_state != STATE_REPLAYING) {
locker.unlock();
on_flush->complete(0);
return;
} else if (m_local_journal_replay == nullptr) {
// raced w/ a tag creation stop/start, which implies that
// the replay is flushed
locker.unlock();
flush_commit_position(on_flush);
return;
}
dout(15) << dendl;
auto ctx = new LambdaContext(
[this, on_flush](int r) {
handle_flush_local_replay(on_flush, r);
});
m_local_journal_replay->flush(ctx);
}
template <typename I>
void Replayer<I>::handle_flush_local_replay(Context* on_flush, int r) {
dout(15) << "r=" << r << dendl;
if (r < 0) {
derr << "error flushing local replay: " << cpp_strerror(r) << dendl;
on_flush->complete(r);
return;
}
flush_commit_position(on_flush);
}
template <typename I>
void Replayer<I>::flush_commit_position(Context* on_flush) {
std::unique_lock locker{m_lock};
if (m_state != STATE_REPLAYING) {
locker.unlock();
on_flush->complete(0);
return;
}
dout(15) << dendl;
auto ctx = new LambdaContext(
[this, on_flush](int r) {
handle_flush_commit_position(on_flush, r);
});
m_state_builder->remote_journaler->flush_commit_position(ctx);
}
template <typename I>
void Replayer<I>::handle_flush_commit_position(Context* on_flush, int r) {
dout(15) << "r=" << r << dendl;
if (r < 0) {
derr << "error flushing remote journal commit position: "
<< cpp_strerror(r) << dendl;
}
on_flush->complete(r);
}
template <typename I>
void Replayer<I>::handle_replay_error(int r, const std::string &error) {
ceph_assert(ceph_mutex_is_locked_by_me(m_lock));
if (m_error_code == 0) {
m_error_code = r;
m_error_description = error;
}
}
template <typename I>
bool Replayer<I>::is_replay_complete() const {
std::unique_lock locker{m_lock};
return is_replay_complete(locker);
}
template <typename I>
bool Replayer<I>::is_replay_complete(
const std::unique_lock<ceph::mutex>&) const {
ceph_assert(ceph_mutex_is_locked_by_me(m_lock));
return (m_state == STATE_COMPLETE);
}
template <typename I>
void Replayer<I>::handle_replay_complete(int r, const std::string &error) {
std::unique_lock locker{m_lock};
handle_replay_complete(locker, r, error);
}
template <typename I>
void Replayer<I>::handle_replay_complete(
const std::unique_lock<ceph::mutex>&, int r, const std::string &error) {
ceph_assert(ceph_mutex_is_locked_by_me(m_lock));
dout(10) << "r=" << r << ", error=" << error << dendl;
if (r < 0) {
derr << "replay encountered an error: " << cpp_strerror(r) << dendl;
handle_replay_error(r, error);
}
if (m_state != STATE_REPLAYING) {
return;
}
m_state = STATE_COMPLETE;
notify_status_updated();
}
template <typename I>
void Replayer<I>::handle_replay_ready() {
std::unique_lock locker{m_lock};
handle_replay_ready(locker);
}
template <typename I>
void Replayer<I>::handle_replay_ready(
std::unique_lock<ceph::mutex>& locker) {
ceph_assert(ceph_mutex_is_locked_by_me(m_lock));
dout(20) << dendl;
if (is_replay_complete(locker)) {
return;
}
if (!m_state_builder->remote_journaler->try_pop_front(&m_replay_entry,
&m_replay_tag_tid)) {
dout(20) << "no entries ready for replay" << dendl;
return;
}
// can safely drop lock once the entry is tracked
m_event_replay_tracker.start_op();
locker.unlock();
dout(20) << "entry tid=" << m_replay_entry.get_commit_tid()
<< "tag_tid=" << m_replay_tag_tid << dendl;
if (!m_replay_tag_valid || m_replay_tag.tid != m_replay_tag_tid) {
// must allocate a new local journal tag prior to processing
replay_flush();
return;
}
preprocess_entry();
}
template <typename I>
void Replayer<I>::replay_flush() {
dout(10) << dendl;
m_flush_tracker.start_op();
// shut down the replay to flush all IO and ops and create a new
// replayer to handle the new tag epoch
auto ctx = create_context_callback<
Replayer<I>, &Replayer<I>::handle_replay_flush_shut_down>(this);
ceph_assert(m_local_journal_replay != nullptr);
m_local_journal_replay->shut_down(false, ctx);
}
template <typename I>
void Replayer<I>::handle_replay_flush_shut_down(int r) {
std::unique_lock locker{m_lock};
dout(10) << "r=" << r << dendl;
ceph_assert(m_local_journal != nullptr);
ceph_assert(m_local_journal_listener != nullptr);
// blocks if listener notification is in-progress
m_local_journal->remove_listener(m_local_journal_listener);
delete m_local_journal_listener;
m_local_journal_listener = nullptr;
m_local_journal->stop_external_replay();
m_local_journal_replay = nullptr;
m_local_journal.reset();
if (r < 0) {
locker.unlock();
handle_replay_flush(r);
return;
}
// journal might have been closed now that we stopped external replay
auto local_image_ctx = m_state_builder->local_image_ctx;
std::shared_lock local_image_locker{local_image_ctx->image_lock};
m_local_journal = local_image_ctx->journal;
if (m_local_journal == nullptr) {
local_image_locker.unlock();
locker.unlock();
derr << "local image journal closed" << dendl;
handle_replay_flush(-EINVAL);
return;
}
auto ctx = create_context_callback<
Replayer<I>, &Replayer<I>::handle_replay_flush>(this);
m_local_journal->start_external_replay(&m_local_journal_replay, ctx);
}
template <typename I>
void Replayer<I>::handle_replay_flush(int r) {
std::unique_lock locker{m_lock};
dout(10) << "r=" << r << dendl;
m_flush_tracker.finish_op();
if (r < 0) {
derr << "replay flush encountered an error: " << cpp_strerror(r) << dendl;
handle_replay_complete(locker, r, "replay flush encountered an error");
m_event_replay_tracker.finish_op();
return;
} else if (is_replay_complete(locker)) {
m_event_replay_tracker.finish_op();
return;
}
// check for resync/promotion state after adding listener
if (!add_local_journal_listener(locker)) {
m_event_replay_tracker.finish_op();
return;
}
locker.unlock();
get_remote_tag();
}
template <typename I>
void Replayer<I>::get_remote_tag() {
dout(15) << "tag_tid: " << m_replay_tag_tid << dendl;
Context *ctx = create_context_callback<
Replayer, &Replayer<I>::handle_get_remote_tag>(this);
m_state_builder->remote_journaler->get_tag(m_replay_tag_tid, &m_replay_tag,
ctx);
}
template <typename I>
void Replayer<I>::handle_get_remote_tag(int r) {
dout(15) << "r=" << r << dendl;
if (r == 0) {
try {
auto it = m_replay_tag.data.cbegin();
decode(m_replay_tag_data, it);
} catch (const buffer::error &err) {
r = -EBADMSG;
}
}
if (r < 0) {
derr << "failed to retrieve remote tag " << m_replay_tag_tid << ": "
<< cpp_strerror(r) << dendl;
handle_replay_complete(r, "failed to retrieve remote tag");
m_event_replay_tracker.finish_op();
return;
}
m_replay_tag_valid = true;
dout(15) << "decoded remote tag " << m_replay_tag_tid << ": "
<< m_replay_tag_data << dendl;
allocate_local_tag();
}
template <typename I>
void Replayer<I>::allocate_local_tag() {
dout(15) << dendl;
std::string mirror_uuid = m_replay_tag_data.mirror_uuid;
if (mirror_uuid == librbd::Journal<>::LOCAL_MIRROR_UUID) {
mirror_uuid = m_state_builder->remote_mirror_uuid;
} else if (mirror_uuid == m_local_mirror_uuid) {
mirror_uuid = librbd::Journal<>::LOCAL_MIRROR_UUID;
} else if (mirror_uuid == librbd::Journal<>::ORPHAN_MIRROR_UUID) {
// handle possible edge condition where daemon can failover and
// the local image has already been promoted/demoted
auto local_tag_data = m_local_journal->get_tag_data();
if (local_tag_data.mirror_uuid == librbd::Journal<>::ORPHAN_MIRROR_UUID &&
(local_tag_data.predecessor.commit_valid &&
local_tag_data.predecessor.mirror_uuid ==
librbd::Journal<>::LOCAL_MIRROR_UUID)) {
dout(15) << "skipping stale demotion event" << dendl;
handle_process_entry_safe(m_replay_entry, m_replay_bytes,
m_replay_start_time, 0);
handle_replay_ready();
return;
} else {
dout(5) << "encountered image demotion: stopping" << dendl;
handle_replay_complete(0, "");
}
}
librbd::journal::TagPredecessor predecessor(m_replay_tag_data.predecessor);
if (predecessor.mirror_uuid == librbd::Journal<>::LOCAL_MIRROR_UUID) {
predecessor.mirror_uuid = m_state_builder->remote_mirror_uuid;
} else if (predecessor.mirror_uuid == m_local_mirror_uuid) {
predecessor.mirror_uuid = librbd::Journal<>::LOCAL_MIRROR_UUID;
}
dout(15) << "mirror_uuid=" << mirror_uuid << ", "
<< "predecessor=" << predecessor << ", "
<< "replay_tag_tid=" << m_replay_tag_tid << dendl;
Context *ctx = create_context_callback<
Replayer, &Replayer<I>::handle_allocate_local_tag>(this);
m_local_journal->allocate_tag(mirror_uuid, predecessor, ctx);
}
template <typename I>
void Replayer<I>::handle_allocate_local_tag(int r) {
dout(15) << "r=" << r << ", "
<< "tag_tid=" << m_local_journal->get_tag_tid() << dendl;
if (r < 0) {
derr << "failed to allocate journal tag: " << cpp_strerror(r) << dendl;
handle_replay_complete(r, "failed to allocate journal tag");
m_event_replay_tracker.finish_op();
return;
}
preprocess_entry();
}
template <typename I>
void Replayer<I>::preprocess_entry() {
dout(20) << "preprocessing entry tid=" << m_replay_entry.get_commit_tid()
<< dendl;
bufferlist data = m_replay_entry.get_data();
auto it = data.cbegin();
int r = m_local_journal_replay->decode(&it, &m_event_entry);
if (r < 0) {
derr << "failed to decode journal event" << dendl;
handle_replay_complete(r, "failed to decode journal event");
m_event_replay_tracker.finish_op();
return;
}
m_replay_bytes = data.length();
uint32_t delay = calculate_replay_delay(
m_event_entry.timestamp,
m_state_builder->local_image_ctx->mirroring_replay_delay);
if (delay == 0) {
handle_preprocess_entry_ready(0);
return;
}
std::unique_lock locker{m_lock};
if (is_replay_complete(locker)) {
// don't schedule a delayed replay task if a shut-down is in-progress
m_event_replay_tracker.finish_op();
return;
}
dout(20) << "delaying replay by " << delay << " sec" << dendl;
std::unique_lock timer_locker{m_threads->timer_lock};
ceph_assert(m_delayed_preprocess_task == nullptr);
m_delayed_preprocess_task = create_context_callback<
Replayer<I>, &Replayer<I>::handle_delayed_preprocess_task>(this);
m_threads->timer->add_event_after(delay, m_delayed_preprocess_task);
}
template <typename I>
void Replayer<I>::handle_delayed_preprocess_task(int r) {
dout(20) << "r=" << r << dendl;
ceph_assert(ceph_mutex_is_locked_by_me(m_threads->timer_lock));
m_delayed_preprocess_task = nullptr;
m_threads->work_queue->queue(create_context_callback<
Replayer, &Replayer<I>::handle_preprocess_entry_ready>(this), 0);
}
template <typename I>
void Replayer<I>::handle_preprocess_entry_ready(int r) {
dout(20) << "r=" << r << dendl;
ceph_assert(r == 0);
m_replay_start_time = ceph_clock_now();
if (!m_event_preprocessor->is_required(m_event_entry)) {
process_entry();
return;
}
Context *ctx = create_context_callback<
Replayer, &Replayer<I>::handle_preprocess_entry_safe>(this);
m_event_preprocessor->preprocess(&m_event_entry, ctx);
}
template <typename I>
void Replayer<I>::handle_preprocess_entry_safe(int r) {
dout(20) << "r=" << r << dendl;
if (r < 0) {
if (r == -ECANCELED) {
handle_replay_complete(0, "lost exclusive lock");
} else {
derr << "failed to preprocess journal event" << dendl;
handle_replay_complete(r, "failed to preprocess journal event");
}
m_event_replay_tracker.finish_op();
return;
}
process_entry();
}
template <typename I>
void Replayer<I>::process_entry() {
dout(20) << "processing entry tid=" << m_replay_entry.get_commit_tid()
<< dendl;
Context *on_ready = create_context_callback<
Replayer, &Replayer<I>::handle_process_entry_ready>(this);
Context *on_commit = new C_ReplayCommitted(this, std::move(m_replay_entry),
m_replay_bytes,
m_replay_start_time);
m_local_journal_replay->process(m_event_entry, on_ready, on_commit);
}
template <typename I>
void Replayer<I>::handle_process_entry_ready(int r) {
std::unique_lock locker{m_lock};
dout(20) << dendl;
ceph_assert(r == 0);
bool update_status = false;
{
auto local_image_ctx = m_state_builder->local_image_ctx;
std::shared_lock image_locker{local_image_ctx->image_lock};
auto image_spec = util::compute_image_spec(local_image_ctx->md_ctx,
local_image_ctx->name);
if (m_image_spec != image_spec) {
m_image_spec = image_spec;
update_status = true;
}
}
m_replay_status_formatter->handle_entry_processed(m_replay_bytes);
if (update_status) {
unregister_perf_counters();
register_perf_counters();
notify_status_updated();
}
// attempt to process the next event
handle_replay_ready(locker);
}
template <typename I>
void Replayer<I>::handle_process_entry_safe(
const ReplayEntry &replay_entry, uint64_t replay_bytes,
const utime_t &replay_start_time, int r) {
dout(20) << "commit_tid=" << replay_entry.get_commit_tid() << ", r=" << r
<< dendl;
if (r < 0) {
derr << "failed to commit journal event: " << cpp_strerror(r) << dendl;
handle_replay_complete(r, "failed to commit journal event");
} else {
ceph_assert(m_state_builder->remote_journaler != nullptr);
m_state_builder->remote_journaler->committed(replay_entry);
}
auto latency = ceph_clock_now() - replay_start_time;
if (g_journal_perf_counters) {
g_journal_perf_counters->inc(l_rbd_mirror_journal_entries);
g_journal_perf_counters->inc(l_rbd_mirror_journal_replay_bytes,
replay_bytes);
g_journal_perf_counters->tinc(l_rbd_mirror_journal_replay_latency,
latency);
}
auto ctx = new LambdaContext(
[this, replay_bytes, latency](int r) {
std::unique_lock locker{m_lock};
schedule_flush_local_replay_task();
if (m_perf_counters) {
m_perf_counters->inc(l_rbd_mirror_journal_entries);
m_perf_counters->inc(l_rbd_mirror_journal_replay_bytes, replay_bytes);
m_perf_counters->tinc(l_rbd_mirror_journal_replay_latency, latency);
}
m_event_replay_tracker.finish_op();
});
m_threads->work_queue->queue(ctx, 0);
}
template <typename I>
void Replayer<I>::handle_resync_image() {
dout(10) << dendl;
std::unique_lock locker{m_lock};
m_resync_requested = true;
handle_replay_complete(locker, 0, "resync requested");
}
template <typename I>
void Replayer<I>::notify_status_updated() {
ceph_assert(ceph_mutex_is_locked_by_me(m_lock));
dout(10) << dendl;
auto ctx = new C_TrackedOp(m_in_flight_op_tracker, new LambdaContext(
[this](int) {
m_replayer_listener->handle_notification();
}));
m_threads->work_queue->queue(ctx, 0);
}
template <typename I>
void Replayer<I>::cancel_delayed_preprocess_task() {
ceph_assert(ceph_mutex_is_locked_by_me(m_lock));
bool canceled_delayed_preprocess_task = false;
{
std::unique_lock timer_locker{m_threads->timer_lock};
if (m_delayed_preprocess_task != nullptr) {
dout(10) << dendl;
canceled_delayed_preprocess_task = m_threads->timer->cancel_event(
m_delayed_preprocess_task);
ceph_assert(canceled_delayed_preprocess_task);
m_delayed_preprocess_task = nullptr;
}
}
if (canceled_delayed_preprocess_task) {
// wake up sleeping replay
m_event_replay_tracker.finish_op();
}
}
template <typename I>
int Replayer<I>::validate_remote_client_state(
const cls::journal::Client& remote_client,
librbd::journal::MirrorPeerClientMeta* remote_client_meta,
bool* resync_requested, std::string* error) {
ceph_assert(ceph_mutex_is_locked_by_me(m_lock));
if (!util::decode_client_meta(remote_client, remote_client_meta)) {
// require operator intervention since the data is corrupt
*error = "error retrieving remote journal client";
return -EBADMSG;
}
auto local_image_ctx = m_state_builder->local_image_ctx;
dout(5) << "image_id=" << local_image_ctx->id << ", "
<< "remote_client_meta.image_id="
<< remote_client_meta->image_id << ", "
<< "remote_client.state=" << remote_client.state << dendl;
if (remote_client_meta->image_id == local_image_ctx->id &&
remote_client.state != cls::journal::CLIENT_STATE_CONNECTED) {
dout(5) << "client flagged disconnected, stopping image replay" << dendl;
if (local_image_ctx->config.template get_val<bool>(
"rbd_mirroring_resync_after_disconnect")) {
dout(10) << "disconnected: automatic resync" << dendl;
*resync_requested = true;
*error = "disconnected: automatic resync";
return -ENOTCONN;
} else {
dout(10) << "disconnected" << dendl;
*error = "disconnected";
return -ENOTCONN;
}
}
return 0;
}
template <typename I>
void Replayer<I>::register_perf_counters() {
dout(5) << dendl;
ceph_assert(ceph_mutex_is_locked_by_me(m_lock));
ceph_assert(m_perf_counters == nullptr);
auto cct = static_cast<CephContext *>(m_state_builder->local_image_ctx->cct);
auto prio = cct->_conf.get_val<int64_t>("rbd_mirror_image_perf_stats_prio");
auto local_image_ctx = m_state_builder->local_image_ctx;
std::string labels = ceph::perf_counters::key_create(
"rbd_mirror_journal_image",
{{"pool", local_image_ctx->md_ctx.get_pool_name()},
{"namespace", local_image_ctx->md_ctx.get_namespace()},
{"image", local_image_ctx->name}});
PerfCountersBuilder plb(g_ceph_context, labels, l_rbd_mirror_journal_first,
l_rbd_mirror_journal_last);
plb.add_u64_counter(l_rbd_mirror_journal_entries, "entries",
"Number of entries replayed", nullptr, prio);
plb.add_u64_counter(l_rbd_mirror_journal_replay_bytes, "replay_bytes",
"Total bytes replayed", nullptr, prio,
unit_t(UNIT_BYTES));
plb.add_time_avg(l_rbd_mirror_journal_replay_latency, "replay_latency",
"Replay latency", nullptr, prio);
m_perf_counters = plb.create_perf_counters();
g_ceph_context->get_perfcounters_collection()->add(m_perf_counters);
}
template <typename I>
void Replayer<I>::unregister_perf_counters() {
dout(5) << dendl;
ceph_assert(ceph_mutex_is_locked_by_me(m_lock));
PerfCounters *perf_counters = nullptr;
std::swap(perf_counters, m_perf_counters);
if (perf_counters != nullptr) {
g_ceph_context->get_perfcounters_collection()->remove(perf_counters);
delete perf_counters;
}
}
} // namespace journal
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_replayer::journal::Replayer<librbd::ImageCtx>;
| 40,251 | 29.540212 | 80 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/journal/Replayer.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_REPLAYER_JOURNAL_REPLAYER_H
#define RBD_MIRROR_IMAGE_REPLAYER_JOURNAL_REPLAYER_H
#include "tools/rbd_mirror/image_replayer/Replayer.h"
#include "include/utime.h"
#include "common/AsyncOpTracker.h"
#include "common/ceph_mutex.h"
#include "common/RefCountedObj.h"
#include "cls/journal/cls_journal_types.h"
#include "journal/ReplayEntry.h"
#include "librbd/ImageCtx.h"
#include "librbd/journal/Types.h"
#include "librbd/journal/TypeTraits.h"
#include <string>
#include <type_traits>
namespace journal { class Journaler; }
namespace librbd {
struct ImageCtx;
namespace journal { template <typename I> class Replay; }
} // namespace librbd
namespace rbd {
namespace mirror {
template <typename> struct Threads;
namespace image_replayer {
struct ReplayerListener;
namespace journal {
template <typename> class EventPreprocessor;
template <typename> class ReplayStatusFormatter;
template <typename> class StateBuilder;
template <typename ImageCtxT>
class Replayer : public image_replayer::Replayer {
public:
typedef typename librbd::journal::TypeTraits<ImageCtxT>::Journaler Journaler;
static Replayer* create(
Threads<ImageCtxT>* threads,
const std::string& local_mirror_uuid,
StateBuilder<ImageCtxT>* state_builder,
ReplayerListener* replayer_listener) {
return new Replayer(threads, local_mirror_uuid, state_builder,
replayer_listener);
}
Replayer(
Threads<ImageCtxT>* threads,
const std::string& local_mirror_uuid,
StateBuilder<ImageCtxT>* state_builder,
ReplayerListener* replayer_listener);
~Replayer();
void destroy() override {
delete this;
}
void init(Context* on_finish) override;
void shut_down(Context* on_finish) override;
void flush(Context* on_finish) override;
bool get_replay_status(std::string* description, Context* on_finish) override;
bool is_replaying() const override {
std::unique_lock locker{m_lock};
return (m_state == STATE_REPLAYING);
}
bool is_resync_requested() const override {
std::unique_lock locker(m_lock);
return m_resync_requested;
}
int get_error_code() const override {
std::unique_lock locker(m_lock);
return m_error_code;
}
std::string get_error_description() const override {
std::unique_lock locker(m_lock);
return m_error_description;
}
std::string get_image_spec() const {
std::unique_lock locker(m_lock);
return m_image_spec;
}
private:
/**
* @verbatim
*
* <init>
* |
* v (error)
* INIT_REMOTE_JOURNALER * * * * * * * * * * * * * * * * * * *
* | *
* v (error) *
* START_EXTERNAL_REPLAY * * * * * * * * * * * * * * * * * * *
* | *
* | /--------------------------------------------\ *
* | | | *
* v v (asok flush) | *
* REPLAYING -------------> LOCAL_REPLAY_FLUSH | *
* | \ | | *
* | | v | *
* | | FLUSH_COMMIT_POSITION | *
* | | | | *
* | | \--------------------/| *
* | | | *
* | | (entries available) | *
* | \-----------> REPLAY_READY | *
* | | | *
* | | (skip if not | *
* | v needed) (error) *
* | REPLAY_FLUSH * * * * * * * * * *
* | | | * *
* | | (skip if not | * *
* | v needed) (error) * *
* | GET_REMOTE_TAG * * * * * * * * *
* | | | * *
* | | (skip if not | * *
* | v needed) (error) * *
* | ALLOCATE_LOCAL_TAG * * * * * * *
* | | | * *
* | v (error) * *
* | PREPROCESS_ENTRY * * * * * * * *
* | | | * *
* | v (error) * *
* | PROCESS_ENTRY * * * * * * * * * *
* | | | * *
* | \---------------------/ * *
* v (shutdown) * *
* REPLAY_COMPLETE < * * * * * * * * * * * * * * * * * * * *
* | *
* v *
* WAIT_FOR_FLUSH *
* | *
* v *
* SHUT_DOWN_LOCAL_JOURNAL_REPLAY *
* | *
* v *
* WAIT_FOR_REPLAY *
* | *
* v *
* CLOSE_LOCAL_IMAGE < * * * * * * * * * * * * * * * * * * * *
* |
* v (skip if not started)
* STOP_REMOTE_JOURNALER_REPLAY
* |
* v
* WAIT_FOR_IN_FLIGHT_OPS
* |
* v
* <shutdown>
*
* @endverbatim
*/
typedef typename librbd::journal::TypeTraits<ImageCtxT>::ReplayEntry ReplayEntry;
enum State {
STATE_INIT,
STATE_REPLAYING,
STATE_COMPLETE
};
struct C_ReplayCommitted;
struct RemoteJournalerListener;
struct RemoteReplayHandler;
struct LocalJournalListener;
Threads<ImageCtxT>* m_threads;
std::string m_local_mirror_uuid;
StateBuilder<ImageCtxT>* m_state_builder;
ReplayerListener* m_replayer_listener;
mutable ceph::mutex m_lock;
std::string m_image_spec;
Context* m_on_init_shutdown = nullptr;
State m_state = STATE_INIT;
int m_error_code = 0;
std::string m_error_description;
bool m_resync_requested = false;
ceph::ref_t<typename std::remove_pointer<decltype(ImageCtxT::journal)>::type>
m_local_journal;
RemoteJournalerListener* m_remote_listener = nullptr;
librbd::journal::Replay<ImageCtxT>* m_local_journal_replay = nullptr;
EventPreprocessor<ImageCtxT>* m_event_preprocessor = nullptr;
ReplayStatusFormatter<ImageCtxT>* m_replay_status_formatter = nullptr;
RemoteReplayHandler* m_remote_replay_handler = nullptr;
LocalJournalListener* m_local_journal_listener = nullptr;
PerfCounters *m_perf_counters = nullptr;
ReplayEntry m_replay_entry;
uint64_t m_replay_bytes = 0;
utime_t m_replay_start_time;
bool m_replay_tag_valid = false;
uint64_t m_replay_tag_tid = 0;
cls::journal::Tag m_replay_tag;
librbd::journal::TagData m_replay_tag_data;
librbd::journal::EventEntry m_event_entry;
AsyncOpTracker m_flush_tracker;
AsyncOpTracker m_event_replay_tracker;
Context *m_delayed_preprocess_task = nullptr;
AsyncOpTracker m_in_flight_op_tracker;
Context *m_flush_local_replay_task = nullptr;
void handle_remote_journal_metadata_updated();
void schedule_flush_local_replay_task();
void cancel_flush_local_replay_task();
void handle_flush_local_replay_task(int r);
void flush_local_replay(Context* on_flush);
void handle_flush_local_replay(Context* on_flush, int r);
void flush_commit_position(Context* on_flush);
void handle_flush_commit_position(Context* on_flush, int r);
void init_remote_journaler();
void handle_init_remote_journaler(int r);
void start_external_replay(std::unique_lock<ceph::mutex>& locker);
void handle_start_external_replay(int r);
bool add_local_journal_listener(std::unique_lock<ceph::mutex>& locker);
bool notify_init_complete(std::unique_lock<ceph::mutex>& locker);
void wait_for_flush();
void handle_wait_for_flush(int r);
void shut_down_local_journal_replay();
void handle_shut_down_local_journal_replay(int r);
void wait_for_event_replay();
void handle_wait_for_event_replay(int r);
void close_local_image();
void handle_close_local_image(int r);
void stop_remote_journaler_replay();
void handle_stop_remote_journaler_replay(int r);
void wait_for_in_flight_ops();
void handle_wait_for_in_flight_ops(int r);
void replay_flush();
void handle_replay_flush_shut_down(int r);
void handle_replay_flush(int r);
void get_remote_tag();
void handle_get_remote_tag(int r);
void allocate_local_tag();
void handle_allocate_local_tag(int r);
void handle_replay_error(int r, const std::string &error);
bool is_replay_complete() const;
bool is_replay_complete(const std::unique_lock<ceph::mutex>& locker) const;
void handle_replay_complete(int r, const std::string &error_desc);
void handle_replay_complete(const std::unique_lock<ceph::mutex>&,
int r, const std::string &error_desc);
void handle_replay_ready();
void handle_replay_ready(std::unique_lock<ceph::mutex>& locker);
void preprocess_entry();
void handle_delayed_preprocess_task(int r);
void handle_preprocess_entry_ready(int r);
void handle_preprocess_entry_safe(int r);
void process_entry();
void handle_process_entry_ready(int r);
void handle_process_entry_safe(const ReplayEntry& replay_entry,
uint64_t relay_bytes,
const utime_t &replay_start_time, int r);
void handle_resync_image();
void notify_status_updated();
void cancel_delayed_preprocess_task();
int validate_remote_client_state(
const cls::journal::Client& remote_client,
librbd::journal::MirrorPeerClientMeta* remote_client_meta,
bool* resync_requested, std::string* error);
void register_perf_counters();
void unregister_perf_counters();
};
} // namespace journal
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_replayer::journal::Replayer<librbd::ImageCtx>;
#endif // RBD_MIRROR_IMAGE_REPLAYER_JOURNAL_REPLAYER_H
| 10,900 | 32.645062 | 87 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/journal/StateBuilder.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "StateBuilder.h"
#include "include/ceph_assert.h"
#include "include/Context.h"
#include "common/debug.h"
#include "common/errno.h"
#include "journal/Journaler.h"
#include "librbd/ImageCtx.h"
#include "librbd/Journal.h"
#include "tools/rbd_mirror/image_replayer/journal/CreateLocalImageRequest.h"
#include "tools/rbd_mirror/image_replayer/journal/PrepareReplayRequest.h"
#include "tools/rbd_mirror/image_replayer/journal/Replayer.h"
#include "tools/rbd_mirror/image_replayer/journal/SyncPointHandler.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_replayer::journal::" \
<< "StateBuilder: " << this << " " \
<< __func__ << ": "
namespace rbd {
namespace mirror {
namespace image_replayer {
namespace journal {
template <typename I>
StateBuilder<I>::StateBuilder(const std::string& global_image_id)
: image_replayer::StateBuilder<I>(global_image_id) {
}
template <typename I>
StateBuilder<I>::~StateBuilder() {
ceph_assert(remote_journaler == nullptr);
}
template <typename I>
void StateBuilder<I>::close(Context* on_finish) {
dout(10) << dendl;
// close the remote journaler after closing the local image
// in case we have lost contact w/ the remote cluster and
// will block
on_finish = new LambdaContext([this, on_finish](int) {
shut_down_remote_journaler(on_finish);
});
on_finish = new LambdaContext([this, on_finish](int) {
this->close_local_image(on_finish);
});
this->close_remote_image(on_finish);
}
template <typename I>
bool StateBuilder<I>::is_disconnected() const {
return (remote_client_state == cls::journal::CLIENT_STATE_DISCONNECTED);
}
template <typename I>
bool StateBuilder<I>::is_linked_impl() const {
ceph_assert(!this->remote_mirror_uuid.empty());
return (local_primary_mirror_uuid == this->remote_mirror_uuid);
}
template <typename I>
cls::rbd::MirrorImageMode StateBuilder<I>::get_mirror_image_mode() const {
return cls::rbd::MIRROR_IMAGE_MODE_JOURNAL;
}
template <typename I>
image_sync::SyncPointHandler* StateBuilder<I>::create_sync_point_handler() {
dout(10) << dendl;
this->m_sync_point_handler = SyncPointHandler<I>::create(this);
return this->m_sync_point_handler;
}
template <typename I>
BaseRequest* StateBuilder<I>::create_local_image_request(
Threads<I>* threads,
librados::IoCtx& local_io_ctx,
const std::string& global_image_id,
PoolMetaCache* pool_meta_cache,
ProgressContext* progress_ctx,
Context* on_finish) {
return CreateLocalImageRequest<I>::create(
threads, local_io_ctx, this->remote_image_ctx, this->global_image_id,
pool_meta_cache, progress_ctx, this, on_finish);
}
template <typename I>
BaseRequest* StateBuilder<I>::create_prepare_replay_request(
const std::string& local_mirror_uuid,
ProgressContext* progress_ctx,
bool* resync_requested,
bool* syncing,
Context* on_finish) {
return PrepareReplayRequest<I>::create(
local_mirror_uuid, progress_ctx, this, resync_requested, syncing,
on_finish);
}
template <typename I>
image_replayer::Replayer* StateBuilder<I>::create_replayer(
Threads<I>* threads,
InstanceWatcher<I>* instance_watcher,
const std::string& local_mirror_uuid,
PoolMetaCache* pool_meta_cache,
ReplayerListener* replayer_listener) {
return Replayer<I>::create(
threads, local_mirror_uuid, this, replayer_listener);
}
template <typename I>
void StateBuilder<I>::shut_down_remote_journaler(Context* on_finish) {
if (remote_journaler == nullptr) {
on_finish->complete(0);
return;
}
dout(10) << dendl;
auto ctx = new LambdaContext([this, on_finish](int r) {
handle_shut_down_remote_journaler(r, on_finish);
});
remote_journaler->shut_down(ctx);
}
template <typename I>
void StateBuilder<I>::handle_shut_down_remote_journaler(int r,
Context* on_finish) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to shut down remote journaler: " << cpp_strerror(r)
<< dendl;
}
delete remote_journaler;
remote_journaler = nullptr;
on_finish->complete(r);
}
} // namespace journal
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_replayer::journal::StateBuilder<librbd::ImageCtx>;
| 4,535 | 29.24 | 84 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/journal/StateBuilder.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef CEPH_RBD_MIRROR_IMAGE_REPLAYER_JOURNAL_STATE_BUILDER_H
#define CEPH_RBD_MIRROR_IMAGE_REPLAYER_JOURNAL_STATE_BUILDER_H
#include "tools/rbd_mirror/image_replayer/StateBuilder.h"
#include "cls/journal/cls_journal_types.h"
#include "librbd/journal/Types.h"
#include "librbd/journal/TypeTraits.h"
#include <string>
struct Context;
namespace librbd { struct ImageCtx; }
namespace rbd {
namespace mirror {
namespace image_replayer {
namespace journal {
template <typename> class SyncPointHandler;
template <typename ImageCtxT>
class StateBuilder : public image_replayer::StateBuilder<ImageCtxT> {
public:
typedef librbd::journal::TypeTraits<ImageCtxT> TypeTraits;
typedef typename TypeTraits::Journaler Journaler;
static StateBuilder* create(const std::string& global_image_id) {
return new StateBuilder(global_image_id);
}
StateBuilder(const std::string& global_image_id);
~StateBuilder() override;
void close(Context* on_finish) override;
bool is_disconnected() const override;
cls::rbd::MirrorImageMode get_mirror_image_mode() const override;
image_sync::SyncPointHandler* create_sync_point_handler() override;
bool replay_requires_remote_image() const override {
return false;
}
BaseRequest* create_local_image_request(
Threads<ImageCtxT>* threads,
librados::IoCtx& local_io_ctx,
const std::string& global_image_id,
PoolMetaCache* pool_meta_cache,
ProgressContext* progress_ctx,
Context* on_finish) override;
BaseRequest* create_prepare_replay_request(
const std::string& local_mirror_uuid,
ProgressContext* progress_ctx,
bool* resync_requested,
bool* syncing,
Context* on_finish) override;
image_replayer::Replayer* create_replayer(
Threads<ImageCtxT>* threads,
InstanceWatcher<ImageCtxT>* instance_watcher,
const std::string& local_mirror_uuid,
PoolMetaCache* pool_meta_cache,
ReplayerListener* replayer_listener) override;
std::string local_primary_mirror_uuid;
Journaler* remote_journaler = nullptr;
cls::journal::ClientState remote_client_state =
cls::journal::CLIENT_STATE_CONNECTED;
librbd::journal::MirrorPeerClientMeta remote_client_meta;
SyncPointHandler<ImageCtxT>* sync_point_handler = nullptr;
private:
bool is_linked_impl() const override;
void shut_down_remote_journaler(Context* on_finish);
void handle_shut_down_remote_journaler(int r, Context* on_finish);
};
} // namespace journal
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_replayer::journal::StateBuilder<librbd::ImageCtx>;
#endif // CEPH_RBD_MIRROR_IMAGE_REPLAYER_JOURNAL_STATE_BUILDER_H
| 2,810 | 28.589474 | 91 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/journal/SyncPointHandler.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "SyncPointHandler.h"
#include "StateBuilder.h"
#include "include/ceph_assert.h"
#include "include/Context.h"
#include "common/debug.h"
#include "common/errno.h"
#include "journal/Journaler.h"
#include "librbd/ImageCtx.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_replayer::journal::" \
<< "SyncPointHandler: " << this << " " \
<< __func__ << ": "
namespace rbd {
namespace mirror {
namespace image_replayer {
namespace journal {
template <typename I>
SyncPointHandler<I>::SyncPointHandler(StateBuilder<I>* state_builder)
: m_state_builder(state_builder),
m_client_meta_copy(state_builder->remote_client_meta) {
}
template <typename I>
typename SyncPointHandler<I>::SyncPoints
SyncPointHandler<I>::get_sync_points() const {
SyncPoints sync_points;
for (auto& sync_point : m_client_meta_copy.sync_points) {
sync_points.emplace_back(
sync_point.snap_namespace,
sync_point.snap_name,
sync_point.from_snap_name,
sync_point.object_number);
}
return sync_points;
}
template <typename I>
librbd::SnapSeqs SyncPointHandler<I>::get_snap_seqs() const {
return m_client_meta_copy.snap_seqs;
}
template <typename I>
void SyncPointHandler<I>::update_sync_points(
const librbd::SnapSeqs& snap_seqs, const SyncPoints& sync_points,
bool sync_complete, Context* on_finish) {
dout(10) << dendl;
if (sync_complete && sync_points.empty()) {
m_client_meta_copy.state = librbd::journal::MIRROR_PEER_STATE_REPLAYING;
}
m_client_meta_copy.snap_seqs = snap_seqs;
m_client_meta_copy.sync_points.clear();
for (auto& sync_point : sync_points) {
m_client_meta_copy.sync_points.emplace_back(
sync_point.snap_namespace,
sync_point.snap_name,
sync_point.from_snap_name,
sync_point.object_number);
if (sync_point.object_number) {
m_client_meta_copy.sync_object_count = std::max(
m_client_meta_copy.sync_object_count, *sync_point.object_number + 1);
}
}
dout(20) << "client_meta=" << m_client_meta_copy << dendl;
bufferlist client_data_bl;
librbd::journal::ClientData client_data{m_client_meta_copy};
encode(client_data, client_data_bl);
auto ctx = new LambdaContext([this, on_finish](int r) {
handle_update_sync_points(r, on_finish);
});
m_state_builder->remote_journaler->update_client(client_data_bl, ctx);
}
template <typename I>
void SyncPointHandler<I>::handle_update_sync_points(int r, Context* on_finish) {
dout(10) << "r=" << r << dendl;
if (r >= 0) {
m_state_builder->remote_client_meta.snap_seqs =
m_client_meta_copy.snap_seqs;
m_state_builder->remote_client_meta.sync_points =
m_client_meta_copy.sync_points;
} else {
derr << "failed to update remote journal client meta for image "
<< m_state_builder->global_image_id << ": " << cpp_strerror(r)
<< dendl;
}
on_finish->complete(r);
}
} // namespace journal
} // namespace image_sync
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_replayer::journal::SyncPointHandler<librbd::ImageCtx>;
| 3,318 | 29.172727 | 88 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/journal/SyncPointHandler.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_REPLAYER_JOURNAL_SYNC_POINT_HANDLER_H
#define RBD_MIRROR_IMAGE_REPLAYER_JOURNAL_SYNC_POINT_HANDLER_H
#include "tools/rbd_mirror/image_sync/Types.h"
#include "librbd/journal/Types.h"
struct Context;
namespace librbd { struct ImageCtx; }
namespace rbd {
namespace mirror {
namespace image_replayer {
namespace journal {
template <typename> class StateBuilder;
template <typename ImageCtxT>
class SyncPointHandler : public image_sync::SyncPointHandler {
public:
using SyncPoint = image_sync::SyncPoint;
using SyncPoints = image_sync::SyncPoints;
static SyncPointHandler* create(StateBuilder<ImageCtxT>* state_builder) {
return new SyncPointHandler(state_builder);
}
SyncPointHandler(StateBuilder<ImageCtxT>* state_builder);
SyncPoints get_sync_points() const override;
librbd::SnapSeqs get_snap_seqs() const override;
void update_sync_points(const librbd::SnapSeqs& snap_seqs,
const SyncPoints& sync_points,
bool sync_complete,
Context* on_finish) override;
private:
StateBuilder<ImageCtxT>* m_state_builder;
librbd::journal::MirrorPeerClientMeta m_client_meta_copy;
void handle_update_sync_points(int r, Context* on_finish);
};
} // namespace journal
} // namespace image_sync
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_replayer::journal::SyncPointHandler<librbd::ImageCtx>;
#endif // RBD_MIRROR_IMAGE_REPLAYER_JOURNAL_SYNC_POINT_HANDLER_H
| 1,620 | 27.946429 | 95 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/snapshot/ApplyImageStateRequest.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "ApplyImageStateRequest.h"
#include "common/debug.h"
#include "common/errno.h"
#include "cls/rbd/cls_rbd_client.h"
#include "librbd/ImageCtx.h"
#include "librbd/Operations.h"
#include "librbd/Utils.h"
#include "librbd/image/GetMetadataRequest.h"
#include "tools/rbd_mirror/image_replayer/snapshot/Utils.h"
#include <boost/algorithm/string/predicate.hpp>
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_replayer::snapshot::" \
<< "ApplyImageStateRequest: " << this << " " \
<< __func__ << ": "
namespace rbd {
namespace mirror {
namespace image_replayer {
namespace snapshot {
using librbd::util::create_context_callback;
using librbd::util::create_rados_callback;
template <typename I>
ApplyImageStateRequest<I>::ApplyImageStateRequest(
const std::string& local_mirror_uuid,
const std::string& remote_mirror_uuid,
I* local_image_ctx,
I* remote_image_ctx,
librbd::mirror::snapshot::ImageState image_state,
Context* on_finish)
: m_local_mirror_uuid(local_mirror_uuid),
m_remote_mirror_uuid(remote_mirror_uuid),
m_local_image_ctx(local_image_ctx),
m_remote_image_ctx(remote_image_ctx),
m_image_state(image_state),
m_on_finish(on_finish) {
dout(15) << "image_state=" << m_image_state << dendl;
std::shared_lock image_locker{m_local_image_ctx->image_lock};
m_features = m_local_image_ctx->features & ~RBD_FEATURES_IMPLICIT_ENABLE;
compute_local_to_remote_snap_ids();
}
template <typename I>
void ApplyImageStateRequest<I>::send() {
rename_image();
}
template <typename I>
void ApplyImageStateRequest<I>::rename_image() {
std::shared_lock owner_locker{m_local_image_ctx->owner_lock};
std::shared_lock image_locker{m_local_image_ctx->image_lock};
if (m_local_image_ctx->name == m_image_state.name) {
image_locker.unlock();
owner_locker.unlock();
update_features();
return;
}
image_locker.unlock();
dout(15) << "local_image_name=" << m_local_image_ctx->name << ", "
<< "remote_image_name=" << m_image_state.name << dendl;
auto ctx = create_context_callback<
ApplyImageStateRequest<I>,
&ApplyImageStateRequest<I>::handle_rename_image>(this);
m_local_image_ctx->operations->execute_rename(m_image_state.name, ctx);
}
template <typename I>
void ApplyImageStateRequest<I>::handle_rename_image(int r) {
dout(15) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to rename image to '" << m_image_state.name << "': "
<< cpp_strerror(r) << dendl;
finish(r);
return;
}
update_features();
}
template <typename I>
void ApplyImageStateRequest<I>::update_features() {
uint64_t feature_updates = 0UL;
bool enabled = false;
auto image_state_features =
m_image_state.features & ~RBD_FEATURES_IMPLICIT_ENABLE;
feature_updates = (m_features & ~image_state_features);
if (feature_updates == 0UL) {
feature_updates = (image_state_features & ~m_features);
enabled = (feature_updates != 0UL);
}
if (feature_updates == 0UL) {
get_image_meta();
return;
}
dout(15) << "image_features=" << m_features << ", "
<< "state_features=" << image_state_features << ", "
<< "feature_updates=" << feature_updates << ", "
<< "enabled=" << enabled << dendl;
if (enabled) {
m_features |= feature_updates;
} else {
m_features &= ~feature_updates;
}
std::shared_lock owner_lock{m_local_image_ctx->owner_lock};
auto ctx = create_context_callback<
ApplyImageStateRequest<I>,
&ApplyImageStateRequest<I>::handle_update_features>(this);
m_local_image_ctx->operations->execute_update_features(
feature_updates, enabled, ctx, 0U);
}
template <typename I>
void ApplyImageStateRequest<I>::handle_update_features(int r) {
dout(15) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to update image features: " << cpp_strerror(r) << dendl;
finish(r);
return;
}
update_features();
}
template <typename I>
void ApplyImageStateRequest<I>::get_image_meta() {
dout(15) << dendl;
auto ctx = create_context_callback<
ApplyImageStateRequest<I>,
&ApplyImageStateRequest<I>::handle_get_image_meta>(this);
auto req = librbd::image::GetMetadataRequest<I>::create(
m_local_image_ctx->md_ctx, m_local_image_ctx->header_oid, true, "", "", 0U,
&m_metadata, ctx);
req->send();
}
template <typename I>
void ApplyImageStateRequest<I>::handle_get_image_meta(int r) {
dout(15) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to fetch local image metadata: " << cpp_strerror(r)
<< dendl;
finish(r);
return;
}
update_image_meta();
}
template <typename I>
void ApplyImageStateRequest<I>::update_image_meta() {
std::set<std::string> keys_to_remove;
for (const auto& [key, value] : m_metadata) {
if (m_image_state.metadata.count(key) == 0) {
dout(15) << "removing image-meta key '" << key << "'" << dendl;
keys_to_remove.insert(key);
}
}
std::map<std::string, bufferlist> metadata_to_update;
for (const auto& [key, value] : m_image_state.metadata) {
auto it = m_metadata.find(key);
if (it == m_metadata.end() || !it->second.contents_equal(value)) {
dout(15) << "updating image-meta key '" << key << "'" << dendl;
metadata_to_update.insert({key, value});
}
}
if (keys_to_remove.empty() && metadata_to_update.empty()) {
unprotect_snapshot();
return;
}
dout(15) << dendl;
librados::ObjectWriteOperation op;
for (const auto& key : keys_to_remove) {
librbd::cls_client::metadata_remove(&op, key);
}
if (!metadata_to_update.empty()) {
librbd::cls_client::metadata_set(&op, metadata_to_update);
}
auto aio_comp = create_rados_callback<
ApplyImageStateRequest<I>,
&ApplyImageStateRequest<I>::handle_update_image_meta>(this);
int r = m_local_image_ctx->md_ctx.aio_operate(m_local_image_ctx->header_oid, aio_comp,
&op);
ceph_assert(r == 0);
aio_comp->release();
}
template <typename I>
void ApplyImageStateRequest<I>::handle_update_image_meta(int r) {
dout(15) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to update image metadata: " << cpp_strerror(r) << dendl;
finish(r);
return;
}
m_metadata.clear();
m_prev_snap_id = CEPH_NOSNAP;
unprotect_snapshot();
}
template <typename I>
void ApplyImageStateRequest<I>::unprotect_snapshot() {
std::shared_lock image_locker{m_local_image_ctx->image_lock};
auto snap_it = m_local_image_ctx->snap_info.begin();
if (m_prev_snap_id != CEPH_NOSNAP) {
snap_it = m_local_image_ctx->snap_info.upper_bound(m_prev_snap_id);
}
for (; snap_it != m_local_image_ctx->snap_info.end(); ++snap_it) {
auto snap_id = snap_it->first;
const auto& snap_info = snap_it->second;
auto user_ns = std::get_if<cls::rbd::UserSnapshotNamespace>(
&snap_info.snap_namespace);
if (user_ns == nullptr) {
dout(20) << "snapshot " << snap_id << " is not a user snapshot" << dendl;
continue;
}
if (snap_info.protection_status == RBD_PROTECTION_STATUS_UNPROTECTED) {
dout(20) << "snapshot " << snap_id << " is already unprotected" << dendl;
continue;
}
auto snap_id_map_it = m_local_to_remote_snap_ids.find(snap_id);
if (snap_id_map_it == m_local_to_remote_snap_ids.end()) {
dout(15) << "snapshot " << snap_id << " does not exist in remote image"
<< dendl;
break;
}
auto remote_snap_id = snap_id_map_it->second;
auto snap_state_it = m_image_state.snapshots.find(remote_snap_id);
if (snap_state_it == m_image_state.snapshots.end()) {
dout(15) << "snapshot " << snap_id << " does not exist in remote image "
<< "state" << dendl;
break;
}
const auto& snap_state = snap_state_it->second;
if (snap_state.protection_status == RBD_PROTECTION_STATUS_UNPROTECTED) {
dout(15) << "snapshot " << snap_id << " is unprotected in remote image"
<< dendl;
break;
}
}
if (snap_it == m_local_image_ctx->snap_info.end()) {
image_locker.unlock();
// no local snapshots to unprotect
m_prev_snap_id = CEPH_NOSNAP;
remove_snapshot();
return;
}
m_prev_snap_id = snap_it->first;
m_snap_name = snap_it->second.name;
image_locker.unlock();
dout(15) << "snap_name=" << m_snap_name << ", "
<< "snap_id=" << m_prev_snap_id << dendl;
std::shared_lock owner_locker{m_local_image_ctx->owner_lock};
auto ctx = create_context_callback<
ApplyImageStateRequest<I>,
&ApplyImageStateRequest<I>::handle_unprotect_snapshot>(this);
m_local_image_ctx->operations->execute_snap_unprotect(
cls::rbd::UserSnapshotNamespace{}, m_snap_name.c_str(), ctx);
}
template <typename I>
void ApplyImageStateRequest<I>::handle_unprotect_snapshot(int r) {
dout(15) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to unprotect snapshot " << m_snap_name << ": "
<< cpp_strerror(r) << dendl;
finish(r);
return;
}
unprotect_snapshot();
}
template <typename I>
void ApplyImageStateRequest<I>::remove_snapshot() {
std::shared_lock image_locker{m_local_image_ctx->image_lock};
auto snap_it = m_local_image_ctx->snap_info.begin();
if (m_prev_snap_id != CEPH_NOSNAP) {
snap_it = m_local_image_ctx->snap_info.upper_bound(m_prev_snap_id);
}
for (; snap_it != m_local_image_ctx->snap_info.end(); ++snap_it) {
auto snap_id = snap_it->first;
const auto& snap_info = snap_it->second;
auto user_ns = std::get_if<cls::rbd::UserSnapshotNamespace>(
&snap_info.snap_namespace);
if (user_ns == nullptr) {
dout(20) << "snapshot " << snap_id << " is not a user snapshot" << dendl;
continue;
}
auto snap_id_map_it = m_local_to_remote_snap_ids.find(snap_id);
if (snap_id_map_it == m_local_to_remote_snap_ids.end()) {
dout(15) << "snapshot " << snap_id << " does not exist in remote image"
<< dendl;
break;
}
auto remote_snap_id = snap_id_map_it->second;
auto snap_state_it = m_image_state.snapshots.find(remote_snap_id);
if (snap_state_it == m_image_state.snapshots.end()) {
dout(15) << "snapshot " << snap_id << " does not exist in remote image "
<< "state" << dendl;
break;
}
}
if (snap_it == m_local_image_ctx->snap_info.end()) {
image_locker.unlock();
// no local snapshots to remove
m_prev_snap_id = CEPH_NOSNAP;
protect_snapshot();
return;
}
m_prev_snap_id = snap_it->first;
m_snap_name = snap_it->second.name;
image_locker.unlock();
dout(15) << "snap_name=" << m_snap_name << ", "
<< "snap_id=" << m_prev_snap_id << dendl;
std::shared_lock owner_locker{m_local_image_ctx->owner_lock};
auto ctx = create_context_callback<
ApplyImageStateRequest<I>,
&ApplyImageStateRequest<I>::handle_remove_snapshot>(this);
m_local_image_ctx->operations->execute_snap_remove(
cls::rbd::UserSnapshotNamespace{}, m_snap_name.c_str(), ctx);
}
template <typename I>
void ApplyImageStateRequest<I>::handle_remove_snapshot(int r) {
dout(15) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to remove snapshot " << m_snap_name << ": "
<< cpp_strerror(r) << dendl;
finish(r);
return;
}
remove_snapshot();
}
template <typename I>
void ApplyImageStateRequest<I>::protect_snapshot() {
std::shared_lock image_locker{m_local_image_ctx->image_lock};
auto snap_it = m_local_image_ctx->snap_info.begin();
if (m_prev_snap_id != CEPH_NOSNAP) {
snap_it = m_local_image_ctx->snap_info.upper_bound(m_prev_snap_id);
}
for (; snap_it != m_local_image_ctx->snap_info.end(); ++snap_it) {
auto snap_id = snap_it->first;
const auto& snap_info = snap_it->second;
auto user_ns = std::get_if<cls::rbd::UserSnapshotNamespace>(
&snap_info.snap_namespace);
if (user_ns == nullptr) {
dout(20) << "snapshot " << snap_id << " is not a user snapshot" << dendl;
continue;
}
if (snap_info.protection_status == RBD_PROTECTION_STATUS_PROTECTED) {
dout(20) << "snapshot " << snap_id << " is already protected" << dendl;
continue;
}
auto snap_id_map_it = m_local_to_remote_snap_ids.find(snap_id);
if (snap_id_map_it == m_local_to_remote_snap_ids.end()) {
dout(15) << "snapshot " << snap_id << " does not exist in remote image"
<< dendl;
continue;
}
auto remote_snap_id = snap_id_map_it->second;
auto snap_state_it = m_image_state.snapshots.find(remote_snap_id);
if (snap_state_it == m_image_state.snapshots.end()) {
dout(15) << "snapshot " << snap_id << " does not exist in remote image "
<< "state" << dendl;
continue;
}
const auto& snap_state = snap_state_it->second;
if (snap_state.protection_status == RBD_PROTECTION_STATUS_PROTECTED) {
dout(15) << "snapshot " << snap_id << " is protected in remote image"
<< dendl;
break;
}
}
if (snap_it == m_local_image_ctx->snap_info.end()) {
image_locker.unlock();
// no local snapshots to protect
m_prev_snap_id = CEPH_NOSNAP;
rename_snapshot();
return;
}
m_prev_snap_id = snap_it->first;
m_snap_name = snap_it->second.name;
image_locker.unlock();
dout(15) << "snap_name=" << m_snap_name << ", "
<< "snap_id=" << m_prev_snap_id << dendl;
std::shared_lock owner_locker{m_local_image_ctx->owner_lock};
auto ctx = create_context_callback<
ApplyImageStateRequest<I>,
&ApplyImageStateRequest<I>::handle_protect_snapshot>(this);
m_local_image_ctx->operations->execute_snap_protect(
cls::rbd::UserSnapshotNamespace{}, m_snap_name.c_str(), ctx);
}
template <typename I>
void ApplyImageStateRequest<I>::handle_protect_snapshot(int r) {
dout(15) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to protect snapshot " << m_snap_name << ": "
<< cpp_strerror(r) << dendl;
finish(r);
return;
}
protect_snapshot();
}
template <typename I>
void ApplyImageStateRequest<I>::rename_snapshot() {
std::shared_lock image_locker{m_local_image_ctx->image_lock};
auto snap_it = m_local_image_ctx->snap_info.begin();
if (m_prev_snap_id != CEPH_NOSNAP) {
snap_it = m_local_image_ctx->snap_info.upper_bound(m_prev_snap_id);
}
for (; snap_it != m_local_image_ctx->snap_info.end(); ++snap_it) {
auto snap_id = snap_it->first;
const auto& snap_info = snap_it->second;
auto user_ns = std::get_if<cls::rbd::UserSnapshotNamespace>(
&snap_info.snap_namespace);
if (user_ns == nullptr) {
dout(20) << "snapshot " << snap_id << " is not a user snapshot" << dendl;
continue;
}
auto snap_id_map_it = m_local_to_remote_snap_ids.find(snap_id);
if (snap_id_map_it == m_local_to_remote_snap_ids.end()) {
dout(15) << "snapshot " << snap_id << " does not exist in remote image"
<< dendl;
continue;
}
auto remote_snap_id = snap_id_map_it->second;
auto snap_state_it = m_image_state.snapshots.find(remote_snap_id);
if (snap_state_it == m_image_state.snapshots.end()) {
dout(15) << "snapshot " << snap_id << " does not exist in remote image "
<< "state" << dendl;
continue;
}
const auto& snap_state = snap_state_it->second;
if (snap_info.name != snap_state.name) {
dout(15) << "snapshot " << snap_id << " has been renamed from '"
<< snap_info.name << "' to '" << snap_state.name << "'"
<< dendl;
m_snap_name = snap_state.name;
break;
}
}
if (snap_it == m_local_image_ctx->snap_info.end()) {
image_locker.unlock();
// no local snapshots to protect
m_prev_snap_id = CEPH_NOSNAP;
set_snapshot_limit();
return;
}
m_prev_snap_id = snap_it->first;
image_locker.unlock();
dout(15) << "snap_name=" << m_snap_name << ", "
<< "snap_id=" << m_prev_snap_id << dendl;
std::shared_lock owner_locker{m_local_image_ctx->owner_lock};
auto ctx = create_context_callback<
ApplyImageStateRequest<I>,
&ApplyImageStateRequest<I>::handle_rename_snapshot>(this);
m_local_image_ctx->operations->execute_snap_rename(
m_prev_snap_id, m_snap_name.c_str(), ctx);
}
template <typename I>
void ApplyImageStateRequest<I>::handle_rename_snapshot(int r) {
dout(15) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to protect snapshot " << m_snap_name << ": "
<< cpp_strerror(r) << dendl;
finish(r);
return;
}
rename_snapshot();
}
template <typename I>
void ApplyImageStateRequest<I>::set_snapshot_limit() {
dout(15) << "snap_limit=" << m_image_state.snap_limit << dendl;
// no need to even check the current limit -- just set it
std::shared_lock owner_locker{m_local_image_ctx->owner_lock};
auto ctx = create_context_callback<
ApplyImageStateRequest<I>,
&ApplyImageStateRequest<I>::handle_set_snapshot_limit>(this);
m_local_image_ctx->operations->execute_snap_set_limit(
m_image_state.snap_limit, ctx);
}
template <typename I>
void ApplyImageStateRequest<I>::handle_set_snapshot_limit(int r) {
dout(15) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to update snapshot limit: " << cpp_strerror(r)
<< dendl;
}
finish(r);
}
template <typename I>
void ApplyImageStateRequest<I>::finish(int r) {
dout(15) << "r=" << r << dendl;
m_on_finish->complete(r);
delete this;
}
template <typename I>
uint64_t ApplyImageStateRequest<I>::compute_remote_snap_id(
uint64_t local_snap_id) {
ceph_assert(ceph_mutex_is_locked(m_local_image_ctx->image_lock));
ceph_assert(ceph_mutex_is_locked(m_remote_image_ctx->image_lock));
// Search our local non-primary snapshots for a mapping to the remote
// snapshot. The non-primary mirror snapshot with the mappings will always
// come at or after the snapshot we are searching against
auto remote_snap_id = util::compute_remote_snap_id(
m_local_image_ctx->image_lock, m_local_image_ctx->snap_info,
local_snap_id, m_remote_mirror_uuid);
if (remote_snap_id != CEPH_NOSNAP) {
return remote_snap_id;
}
// if we failed to find a match to a remote snapshot in our local non-primary
// snapshots, check the remote image for non-primary snapshot mappings back
// to our snapshot
for (auto snap_it = m_remote_image_ctx->snap_info.begin();
snap_it != m_remote_image_ctx->snap_info.end(); ++snap_it) {
auto snap_id = snap_it->first;
auto mirror_ns = std::get_if<cls::rbd::MirrorSnapshotNamespace>(
&snap_it->second.snap_namespace);
if (mirror_ns == nullptr || !mirror_ns->is_non_primary()) {
continue;
}
if (mirror_ns->primary_mirror_uuid != m_local_mirror_uuid) {
dout(20) << "remote snapshot " << snap_id << " not tied to local"
<< dendl;
continue;
} else if (mirror_ns->primary_snap_id == local_snap_id) {
dout(15) << "local snapshot " << local_snap_id << " maps to "
<< "remote snapshot " << snap_id << dendl;
return snap_id;
}
const auto& snap_seqs = mirror_ns->snap_seqs;
for (auto [local_snap_id_seq, remote_snap_id_seq] : snap_seqs) {
if (local_snap_id_seq == local_snap_id) {
dout(15) << "local snapshot " << local_snap_id << " maps to "
<< "remote snapshot " << remote_snap_id_seq << dendl;
return remote_snap_id_seq;
}
}
}
return CEPH_NOSNAP;
}
template <typename I>
void ApplyImageStateRequest<I>::compute_local_to_remote_snap_ids() {
ceph_assert(ceph_mutex_is_locked(m_local_image_ctx->image_lock));
std::shared_lock remote_image_locker{m_remote_image_ctx->image_lock};
for (const auto& [snap_id, snap_info] : m_local_image_ctx->snap_info) {
m_local_to_remote_snap_ids[snap_id] = compute_remote_snap_id(snap_id);
}
dout(15) << "local_to_remote_snap_ids=" << m_local_to_remote_snap_ids
<< dendl;
}
} // namespace snapshot
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_replayer::snapshot::ApplyImageStateRequest<librbd::ImageCtx>;
| 20,398 | 29.954476 | 95 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/snapshot/ApplyImageStateRequest.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_REPLAYER_SNAPSHOT_APPLY_IMAGE_STATE_REQUEST_H
#define RBD_MIRROR_IMAGE_REPLAYER_SNAPSHOT_APPLY_IMAGE_STATE_REQUEST_H
#include "common/ceph_mutex.h"
#include "librbd/mirror/snapshot/Types.h"
#include <map>
#include <string>
struct Context;
namespace librbd {
struct ImageCtx;
} // namespace librbd
namespace rbd {
namespace mirror {
namespace image_replayer {
namespace snapshot {
template <typename> class EventPreprocessor;
template <typename> class ReplayStatusFormatter;
template <typename> class StateBuilder;
template <typename ImageCtxT>
class ApplyImageStateRequest {
public:
static ApplyImageStateRequest* create(
const std::string& local_mirror_uuid,
const std::string& remote_mirror_uuid,
ImageCtxT* local_image_ctx,
ImageCtxT* remote_image_ctx,
librbd::mirror::snapshot::ImageState image_state,
Context* on_finish) {
return new ApplyImageStateRequest(local_mirror_uuid, remote_mirror_uuid,
local_image_ctx, remote_image_ctx,
image_state, on_finish);
}
ApplyImageStateRequest(
const std::string& local_mirror_uuid,
const std::string& remote_mirror_uuid,
ImageCtxT* local_image_ctx,
ImageCtxT* remote_image_ctx,
librbd::mirror::snapshot::ImageState image_state,
Context* on_finish);
void send();
private:
/**
* @verbatim
*
* <start>
* |
* v
* RENAME_IMAGE
* |
* | /---------\
* | | |
* v v |
* UPDATE_FEATURES -----/
* |
* v
* GET_IMAGE_META
* |
* | /---------\
* | | |
* v v |
* UPDATE_IMAGE_META ---/
* |
* | /---------\
* | | |
* v v |
* UNPROTECT_SNAPSHOT |
* | |
* v |
* REMOVE_SNAPSHOT |
* | |
* v |
* PROTECT_SNAPSHOT |
* | |
* v |
* RENAME_SNAPSHOT -----/
* |
* v
* SET_SNAPSHOT_LIMIT
* |
* v
* <finish>
*
* @endverbatim
*/
std::string m_local_mirror_uuid;
std::string m_remote_mirror_uuid;
ImageCtxT* m_local_image_ctx;
ImageCtxT* m_remote_image_ctx;
librbd::mirror::snapshot::ImageState m_image_state;
Context* m_on_finish;
std::map<uint64_t, uint64_t> m_local_to_remote_snap_ids;
uint64_t m_features = 0;
std::map<std::string, bufferlist> m_metadata;
uint64_t m_prev_snap_id = 0;
std::string m_snap_name;
void rename_image();
void handle_rename_image(int r);
void update_features();
void handle_update_features(int r);
void get_image_meta();
void handle_get_image_meta(int r);
void update_image_meta();
void handle_update_image_meta(int r);
void unprotect_snapshot();
void handle_unprotect_snapshot(int r);
void remove_snapshot();
void handle_remove_snapshot(int r);
void protect_snapshot();
void handle_protect_snapshot(int r);
void rename_snapshot();
void handle_rename_snapshot(int r);
void set_snapshot_limit();
void handle_set_snapshot_limit(int r);
void finish(int r);
uint64_t compute_remote_snap_id(uint64_t snap_id);
void compute_local_to_remote_snap_ids();
};
} // namespace snapshot
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_replayer::snapshot::ApplyImageStateRequest<librbd::ImageCtx>;
#endif // RBD_MIRROR_IMAGE_REPLAYER_SNAPSHOT_APPLY_IMAGE_STATE_REQUEST_H
| 3,754 | 23.070513 | 102 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/snapshot/CreateLocalImageRequest.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "CreateLocalImageRequest.h"
#include "include/rados/librados.hpp"
#include "common/debug.h"
#include "common/dout.h"
#include "common/errno.h"
#include "cls/rbd/cls_rbd_client.h"
#include "cls/rbd/cls_rbd_types.h"
#include "librbd/ImageCtx.h"
#include "librbd/Utils.h"
#include "tools/rbd_mirror/ProgressContext.h"
#include "tools/rbd_mirror/image_replayer/CreateImageRequest.h"
#include "tools/rbd_mirror/image_replayer/snapshot/StateBuilder.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_replayer::snapshot::" \
<< "CreateLocalImageRequest: " << this << " " \
<< __func__ << ": "
namespace rbd {
namespace mirror {
namespace image_replayer {
namespace snapshot {
using librbd::util::create_context_callback;
using librbd::util::create_rados_callback;
template <typename I>
void CreateLocalImageRequest<I>::send() {
disable_mirror_image();
}
template <typename I>
void CreateLocalImageRequest<I>::disable_mirror_image() {
if (m_state_builder->local_image_id.empty()) {
add_mirror_image();
return;
}
dout(10) << dendl;
update_progress("DISABLE_MIRROR_IMAGE");
// need to send 'disabling' since the cls methods will fail if we aren't
// in that state
cls::rbd::MirrorImage mirror_image{
cls::rbd::MIRROR_IMAGE_MODE_SNAPSHOT, m_global_image_id,
cls::rbd::MIRROR_IMAGE_STATE_DISABLING};
librados::ObjectWriteOperation op;
librbd::cls_client::mirror_image_set(&op, m_state_builder->local_image_id,
mirror_image);
auto aio_comp = create_rados_callback<
CreateLocalImageRequest<I>,
&CreateLocalImageRequest<I>::handle_disable_mirror_image>(this);
int r = m_local_io_ctx.aio_operate(RBD_MIRRORING, aio_comp, &op);
ceph_assert(r == 0);
aio_comp->release();
}
template <typename I>
void CreateLocalImageRequest<I>::handle_disable_mirror_image(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to disable mirror image " << m_global_image_id << ": "
<< cpp_strerror(r) << dendl;
finish(r);
return;
}
remove_mirror_image();
}
template <typename I>
void CreateLocalImageRequest<I>::remove_mirror_image() {
dout(10) << dendl;
update_progress("REMOVE_MIRROR_IMAGE");
librados::ObjectWriteOperation op;
librbd::cls_client::mirror_image_remove(&op, m_state_builder->local_image_id);
auto aio_comp = create_rados_callback<
CreateLocalImageRequest<I>,
&CreateLocalImageRequest<I>::handle_remove_mirror_image>(this);
int r = m_local_io_ctx.aio_operate(RBD_MIRRORING, aio_comp, &op);
ceph_assert(r == 0);
aio_comp->release();
}
template <typename I>
void CreateLocalImageRequest<I>::handle_remove_mirror_image(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to remove mirror image " << m_global_image_id << ": "
<< cpp_strerror(r) << dendl;
finish(r);
return;
}
m_state_builder->local_image_id = "";
add_mirror_image();
}
template <typename I>
void CreateLocalImageRequest<I>::add_mirror_image() {
ceph_assert(m_state_builder->local_image_id.empty());
m_state_builder->local_image_id =
librbd::util::generate_image_id<I>(m_local_io_ctx);
dout(10) << "local_image_id=" << m_state_builder->local_image_id << dendl;
update_progress("ADD_MIRROR_IMAGE");
// use 'creating' to track a partially constructed image. it will
// be switched to 'enabled' once the image is fully created
cls::rbd::MirrorImage mirror_image{
cls::rbd::MIRROR_IMAGE_MODE_SNAPSHOT, m_global_image_id,
cls::rbd::MIRROR_IMAGE_STATE_CREATING};
librados::ObjectWriteOperation op;
librbd::cls_client::mirror_image_set(&op, m_state_builder->local_image_id,
mirror_image);
auto aio_comp = create_rados_callback<
CreateLocalImageRequest<I>,
&CreateLocalImageRequest<I>::handle_add_mirror_image>(this);
int r = m_local_io_ctx.aio_operate(RBD_MIRRORING, aio_comp, &op);
ceph_assert(r == 0);
aio_comp->release();
}
template <typename I>
void CreateLocalImageRequest<I>::handle_add_mirror_image(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to register mirror image " << m_global_image_id << ": "
<< cpp_strerror(r) << dendl;
this->finish(r);
return;
}
create_local_image();
}
template <typename I>
void CreateLocalImageRequest<I>::create_local_image() {
dout(10) << "local_image_id=" << m_state_builder->local_image_id << dendl;
update_progress("CREATE_LOCAL_IMAGE");
m_remote_image_ctx->image_lock.lock_shared();
std::string image_name = m_remote_image_ctx->name;
m_remote_image_ctx->image_lock.unlock_shared();
auto ctx = create_context_callback<
CreateLocalImageRequest<I>,
&CreateLocalImageRequest<I>::handle_create_local_image>(this);
auto request = CreateImageRequest<I>::create(
m_threads, m_local_io_ctx, m_global_image_id,
m_state_builder->remote_mirror_uuid, image_name,
m_state_builder->local_image_id, m_remote_image_ctx,
m_pool_meta_cache, cls::rbd::MIRROR_IMAGE_MODE_SNAPSHOT, ctx);
request->send();
}
template <typename I>
void CreateLocalImageRequest<I>::handle_create_local_image(int r) {
dout(10) << "r=" << r << dendl;
if (r == -EBADF) {
dout(5) << "image id " << m_state_builder->local_image_id << " "
<< "already in-use" << dendl;
disable_mirror_image();
return;
} else if (r < 0) {
if (r == -ENOENT) {
dout(10) << "parent image does not exist" << dendl;
} else {
derr << "failed to create local image: " << cpp_strerror(r) << dendl;
}
finish(r);
return;
}
finish(0);
}
template <typename I>
void CreateLocalImageRequest<I>::update_progress(
const std::string& description) {
dout(15) << description << dendl;
if (m_progress_ctx != nullptr) {
m_progress_ctx->update_progress(description);
}
}
} // namespace snapshot
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_replayer::snapshot::CreateLocalImageRequest<librbd::ImageCtx>;
| 6,315 | 29.809756 | 96 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/snapshot/CreateLocalImageRequest.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_REPLAYER_SNAPSHOT_CREATE_LOCAL_IMAGE_REQUEST_H
#define RBD_MIRROR_IMAGE_REPLAYER_SNAPSHOT_CREATE_LOCAL_IMAGE_REQUEST_H
#include "include/rados/librados_fwd.hpp"
#include "tools/rbd_mirror/BaseRequest.h"
#include <string>
struct Context;
namespace librbd { class ImageCtx; }
namespace rbd {
namespace mirror {
class PoolMetaCache;
class ProgressContext;
template <typename> struct Threads;
namespace image_replayer {
namespace snapshot {
template <typename> class StateBuilder;
template <typename ImageCtxT>
class CreateLocalImageRequest : public BaseRequest {
public:
typedef rbd::mirror::ProgressContext ProgressContext;
static CreateLocalImageRequest* create(
Threads<ImageCtxT>* threads,
librados::IoCtx& local_io_ctx,
ImageCtxT* remote_image_ctx,
const std::string& global_image_id,
PoolMetaCache* pool_meta_cache,
ProgressContext* progress_ctx,
StateBuilder<ImageCtxT>* state_builder,
Context* on_finish) {
return new CreateLocalImageRequest(threads, local_io_ctx, remote_image_ctx,
global_image_id, pool_meta_cache,
progress_ctx, state_builder, on_finish);
}
CreateLocalImageRequest(
Threads<ImageCtxT>* threads,
librados::IoCtx& local_io_ctx,
ImageCtxT* remote_image_ctx,
const std::string& global_image_id,
PoolMetaCache* pool_meta_cache,
ProgressContext* progress_ctx,
StateBuilder<ImageCtxT>* state_builder,
Context* on_finish)
: BaseRequest(on_finish),
m_threads(threads),
m_local_io_ctx(local_io_ctx),
m_remote_image_ctx(remote_image_ctx),
m_global_image_id(global_image_id),
m_pool_meta_cache(pool_meta_cache),
m_progress_ctx(progress_ctx),
m_state_builder(state_builder) {
}
void send();
private:
/**
* @verbatim
*
* <start>
* |
* v
* DISABLE_MIRROR_IMAGE < * * * * * *
* | *
* v *
* REMOVE_MIRROR_IMAGE *
* | *
* v *
* ADD_MIRROR_IMAGE *
* | *
* v (id exists) *
* CREATE_LOCAL_IMAGE * * * * * * * *
* |
* v
* <finish>
*
* @endverbatim
*/
Threads<ImageCtxT>* m_threads;
librados::IoCtx& m_local_io_ctx;
ImageCtxT* m_remote_image_ctx;
std::string m_global_image_id;
PoolMetaCache* m_pool_meta_cache;
ProgressContext* m_progress_ctx;
StateBuilder<ImageCtxT>* m_state_builder;
void disable_mirror_image();
void handle_disable_mirror_image(int r);
void remove_mirror_image();
void handle_remove_mirror_image(int r);
void add_mirror_image();
void handle_add_mirror_image(int r);
void create_local_image();
void handle_create_local_image(int r);
void update_progress(const std::string& description);
};
} // namespace snapshot
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_replayer::snapshot::CreateLocalImageRequest<librbd::ImageCtx>;
#endif // RBD_MIRROR_IMAGE_REPLAYER_SNAPSHOT_CREATE_LOCAL_IMAGE_REQUEST_H
| 3,377 | 26.688525 | 103 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/snapshot/PrepareReplayRequest.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "PrepareReplayRequest.h"
#include "common/debug.h"
#include "common/dout.h"
#include "common/errno.h"
#include "librbd/ImageCtx.h"
#include "librbd/Utils.h"
#include "librbd/mirror/snapshot/ImageMeta.h"
#include "tools/rbd_mirror/ProgressContext.h"
#include "tools/rbd_mirror/image_replayer/snapshot/StateBuilder.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_replayer::snapshot::" \
<< "PrepareReplayRequest: " << this << " " \
<< __func__ << ": "
namespace rbd {
namespace mirror {
namespace image_replayer {
namespace snapshot {
using librbd::util::create_context_callback;
template <typename I>
void PrepareReplayRequest<I>::send() {
*m_resync_requested = false;
*m_syncing = false;
load_local_image_meta();
}
template <typename I>
void PrepareReplayRequest<I>::load_local_image_meta() {
dout(15) << dendl;
ceph_assert(m_state_builder->local_image_meta == nullptr);
m_state_builder->local_image_meta =
librbd::mirror::snapshot::ImageMeta<I>::create(
m_state_builder->local_image_ctx, m_local_mirror_uuid);
auto ctx = create_context_callback<
PrepareReplayRequest<I>,
&PrepareReplayRequest<I>::handle_load_local_image_meta>(this);
m_state_builder->local_image_meta->load(ctx);
}
template <typename I>
void PrepareReplayRequest<I>::handle_load_local_image_meta(int r) {
dout(15) << "r=" << r << dendl;
if (r < 0 && r != -ENOENT) {
derr << "failed to load local image-meta: " << cpp_strerror(r) << dendl;
finish(r);
return;
}
*m_resync_requested = m_state_builder->local_image_meta->resync_requested;
finish(0);
}
} // namespace snapshot
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_replayer::snapshot::PrepareReplayRequest<librbd::ImageCtx>;
| 2,039 | 27.732394 | 93 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/snapshot/PrepareReplayRequest.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_REPLAYER_JOURNAL_PREPARE_REPLAY_REQUEST_H
#define RBD_MIRROR_IMAGE_REPLAYER_JOURNAL_PREPARE_REPLAY_REQUEST_H
#include "include/int_types.h"
#include "librbd/mirror/Types.h"
#include "tools/rbd_mirror/BaseRequest.h"
#include <list>
#include <string>
struct Context;
namespace librbd { struct ImageCtx; }
namespace rbd {
namespace mirror {
class ProgressContext;
namespace image_replayer {
namespace snapshot {
template <typename> class StateBuilder;
template <typename ImageCtxT>
class PrepareReplayRequest : public BaseRequest {
public:
static PrepareReplayRequest* create(
const std::string& local_mirror_uuid,
ProgressContext* progress_ctx,
StateBuilder<ImageCtxT>* state_builder,
bool* resync_requested,
bool* syncing,
Context* on_finish) {
return new PrepareReplayRequest(
local_mirror_uuid, progress_ctx, state_builder, resync_requested,
syncing, on_finish);
}
PrepareReplayRequest(
const std::string& local_mirror_uuid,
ProgressContext* progress_ctx,
StateBuilder<ImageCtxT>* state_builder,
bool* resync_requested,
bool* syncing,
Context* on_finish)
: BaseRequest(on_finish),
m_local_mirror_uuid(local_mirror_uuid),
m_progress_ctx(progress_ctx),
m_state_builder(state_builder),
m_resync_requested(resync_requested),
m_syncing(syncing) {
}
void send() override;
private:
// TODO
/**
* @verbatim
*
* <start>
* |
* v
* LOAD_LOCAL_IMAGE_META
* |
* v
* <finish>
*
* @endverbatim
*/
std::string m_local_mirror_uuid;
ProgressContext* m_progress_ctx;
StateBuilder<ImageCtxT>* m_state_builder;
bool* m_resync_requested;
bool* m_syncing;
void load_local_image_meta();
void handle_load_local_image_meta(int r);
};
} // namespace snapshot
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_replayer::snapshot::PrepareReplayRequest<librbd::ImageCtx>;
#endif // RBD_MIRROR_IMAGE_REPLAYER_JOURNAL_PREPARE_REPLAY_REQUEST_H
| 2,212 | 22.795699 | 100 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/snapshot/Replayer.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "Replayer.h"
#include "common/debug.h"
#include "common/errno.h"
#include "common/perf_counters.h"
#include "common/perf_counters_key.h"
#include "include/stringify.h"
#include "common/Timer.h"
#include "cls/rbd/cls_rbd_client.h"
#include "json_spirit/json_spirit.h"
#include "librbd/ImageCtx.h"
#include "librbd/ImageState.h"
#include "librbd/Operations.h"
#include "librbd/Utils.h"
#include "librbd/asio/ContextWQ.h"
#include "librbd/deep_copy/Handler.h"
#include "librbd/deep_copy/ImageCopyRequest.h"
#include "librbd/deep_copy/SnapshotCopyRequest.h"
#include "librbd/mirror/ImageStateUpdateRequest.h"
#include "librbd/mirror/snapshot/CreateNonPrimaryRequest.h"
#include "librbd/mirror/snapshot/GetImageStateRequest.h"
#include "librbd/mirror/snapshot/ImageMeta.h"
#include "librbd/mirror/snapshot/UnlinkPeerRequest.h"
#include "tools/rbd_mirror/InstanceWatcher.h"
#include "tools/rbd_mirror/PoolMetaCache.h"
#include "tools/rbd_mirror/Threads.h"
#include "tools/rbd_mirror/Types.h"
#include "tools/rbd_mirror/image_replayer/CloseImageRequest.h"
#include "tools/rbd_mirror/image_replayer/ReplayerListener.h"
#include "tools/rbd_mirror/image_replayer/Utils.h"
#include "tools/rbd_mirror/image_replayer/snapshot/ApplyImageStateRequest.h"
#include "tools/rbd_mirror/image_replayer/snapshot/StateBuilder.h"
#include "tools/rbd_mirror/image_replayer/snapshot/Utils.h"
#include <set>
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_replayer::snapshot::" \
<< "Replayer: " << this << " " << __func__ << ": "
extern PerfCounters *g_snapshot_perf_counters;
namespace rbd {
namespace mirror {
namespace image_replayer {
namespace snapshot {
namespace {
double round_to_two_places(double value) {
return abs(round(value * 100) / 100);
}
template<typename I>
std::pair<uint64_t, librbd::SnapInfo*> get_newest_mirror_snapshot(
I* image_ctx) {
for (auto snap_info_it = image_ctx->snap_info.rbegin();
snap_info_it != image_ctx->snap_info.rend(); ++snap_info_it) {
const auto& snap_ns = snap_info_it->second.snap_namespace;
auto mirror_ns = std::get_if<
cls::rbd::MirrorSnapshotNamespace>(&snap_ns);
if (mirror_ns == nullptr || !mirror_ns->complete) {
continue;
}
return {snap_info_it->first, &snap_info_it->second};
}
return {CEPH_NOSNAP, nullptr};
}
} // anonymous namespace
using librbd::util::create_async_context_callback;
using librbd::util::create_context_callback;
using librbd::util::create_rados_callback;
template <typename I>
struct Replayer<I>::C_UpdateWatchCtx : public librbd::UpdateWatchCtx {
Replayer<I>* replayer;
C_UpdateWatchCtx(Replayer<I>* replayer) : replayer(replayer) {
}
void handle_notify() override {
replayer->handle_image_update_notify();
}
};
template <typename I>
struct Replayer<I>::DeepCopyHandler : public librbd::deep_copy::Handler {
Replayer *replayer;
DeepCopyHandler(Replayer* replayer) : replayer(replayer) {
}
void handle_read(uint64_t bytes_read) override {
replayer->handle_copy_image_read(bytes_read);
}
int update_progress(uint64_t object_number, uint64_t object_count) override {
replayer->handle_copy_image_progress(object_number, object_count);
return 0;
}
};
template <typename I>
Replayer<I>::Replayer(
Threads<I>* threads,
InstanceWatcher<I>* instance_watcher,
const std::string& local_mirror_uuid,
PoolMetaCache* pool_meta_cache,
StateBuilder<I>* state_builder,
ReplayerListener* replayer_listener)
: m_threads(threads),
m_instance_watcher(instance_watcher),
m_local_mirror_uuid(local_mirror_uuid),
m_pool_meta_cache(pool_meta_cache),
m_state_builder(state_builder),
m_replayer_listener(replayer_listener),
m_lock(ceph::make_mutex(librbd::util::unique_lock_name(
"rbd::mirror::image_replayer::snapshot::Replayer", this))) {
dout(10) << dendl;
}
template <typename I>
Replayer<I>::~Replayer() {
dout(10) << dendl;
{
std::unique_lock locker{m_lock};
unregister_perf_counters();
}
ceph_assert(m_state == STATE_COMPLETE);
ceph_assert(m_update_watch_ctx == nullptr);
ceph_assert(m_deep_copy_handler == nullptr);
}
template <typename I>
void Replayer<I>::init(Context* on_finish) {
dout(10) << dendl;
ceph_assert(m_state == STATE_INIT);
RemotePoolMeta remote_pool_meta;
int r = m_pool_meta_cache->get_remote_pool_meta(
m_state_builder->remote_image_ctx->md_ctx.get_id(), &remote_pool_meta);
if (r < 0 || remote_pool_meta.mirror_peer_uuid.empty()) {
derr << "failed to retrieve mirror peer uuid from remote pool" << dendl;
m_state = STATE_COMPLETE;
m_threads->work_queue->queue(on_finish, r);
return;
}
m_remote_mirror_peer_uuid = remote_pool_meta.mirror_peer_uuid;
dout(10) << "remote_mirror_peer_uuid=" << m_remote_mirror_peer_uuid << dendl;
{
auto local_image_ctx = m_state_builder->local_image_ctx;
std::shared_lock image_locker{local_image_ctx->image_lock};
m_image_spec = image_replayer::util::compute_image_spec(
local_image_ctx->md_ctx, local_image_ctx->name);
}
{
std::unique_lock locker{m_lock};
register_perf_counters();
}
ceph_assert(m_on_init_shutdown == nullptr);
m_on_init_shutdown = on_finish;
register_local_update_watcher();
}
template <typename I>
void Replayer<I>::shut_down(Context* on_finish) {
dout(10) << dendl;
std::unique_lock locker{m_lock};
ceph_assert(m_on_init_shutdown == nullptr);
m_on_init_shutdown = on_finish;
m_error_code = 0;
m_error_description = "";
ceph_assert(m_state != STATE_INIT);
auto state = STATE_COMPLETE;
std::swap(m_state, state);
if (state == STATE_REPLAYING) {
// if a sync request was pending, request a cancelation
m_instance_watcher->cancel_sync_request(
m_state_builder->local_image_ctx->id);
// TODO interrupt snapshot copy and image copy state machines even if remote
// cluster is unreachable
dout(10) << "shut down pending on completion of snapshot replay" << dendl;
return;
}
locker.unlock();
unregister_remote_update_watcher();
}
template <typename I>
void Replayer<I>::flush(Context* on_finish) {
dout(10) << dendl;
// TODO
m_threads->work_queue->queue(on_finish, 0);
}
template <typename I>
bool Replayer<I>::get_replay_status(std::string* description,
Context* on_finish) {
dout(10) << dendl;
std::unique_lock locker{m_lock};
if (m_state != STATE_REPLAYING && m_state != STATE_IDLE) {
locker.unlock();
derr << "replay not running" << dendl;
on_finish->complete(-EAGAIN);
return false;
}
std::shared_lock local_image_locker{
m_state_builder->local_image_ctx->image_lock};
auto [local_snap_id, local_snap_info] = get_newest_mirror_snapshot(
m_state_builder->local_image_ctx);
std::shared_lock remote_image_locker{
m_state_builder->remote_image_ctx->image_lock};
auto [remote_snap_id, remote_snap_info] = get_newest_mirror_snapshot(
m_state_builder->remote_image_ctx);
if (remote_snap_info == nullptr) {
remote_image_locker.unlock();
local_image_locker.unlock();
locker.unlock();
derr << "remote image does not contain mirror snapshots" << dendl;
on_finish->complete(-EAGAIN);
return false;
}
std::string replay_state = "idle";
if (m_remote_snap_id_end != CEPH_NOSNAP) {
replay_state = "syncing";
}
json_spirit::mObject root_obj;
root_obj["replay_state"] = replay_state;
root_obj["remote_snapshot_timestamp"] = remote_snap_info->timestamp.sec();
if (m_perf_counters) {
m_perf_counters->tset(l_rbd_mirror_snapshot_remote_timestamp,
remote_snap_info->timestamp);
}
auto matching_remote_snap_id = util::compute_remote_snap_id(
m_state_builder->local_image_ctx->image_lock,
m_state_builder->local_image_ctx->snap_info,
local_snap_id, m_state_builder->remote_mirror_uuid);
auto matching_remote_snap_it =
m_state_builder->remote_image_ctx->snap_info.find(matching_remote_snap_id);
if (matching_remote_snap_id != CEPH_NOSNAP &&
matching_remote_snap_it !=
m_state_builder->remote_image_ctx->snap_info.end()) {
// use the timestamp from the matching remote image since
// the local snapshot would just be the time the snapshot was
// synced and not the consistency point in time.
root_obj["local_snapshot_timestamp"] =
matching_remote_snap_it->second.timestamp.sec();
if (m_perf_counters) {
m_perf_counters->tset(l_rbd_mirror_snapshot_local_timestamp,
matching_remote_snap_it->second.timestamp);
}
}
matching_remote_snap_it = m_state_builder->remote_image_ctx->snap_info.find(
m_remote_snap_id_end);
if (m_remote_snap_id_end != CEPH_NOSNAP &&
matching_remote_snap_it !=
m_state_builder->remote_image_ctx->snap_info.end()) {
root_obj["syncing_snapshot_timestamp"] = remote_snap_info->timestamp.sec();
if (m_local_object_count > 0) {
root_obj["syncing_percent"] =
100 * m_local_mirror_snap_ns.last_copied_object_number /
m_local_object_count;
} else {
// Set syncing_percent to 0 if m_local_object_count has
// not yet been set (last_copied_object_number may be > 0
// if the sync is being resumed).
root_obj["syncing_percent"] = 0;
}
}
m_bytes_per_second(0);
auto bytes_per_second = m_bytes_per_second.get_average();
root_obj["bytes_per_second"] = round_to_two_places(bytes_per_second);
auto bytes_per_snapshot = boost::accumulators::rolling_mean(
m_bytes_per_snapshot);
root_obj["bytes_per_snapshot"] = round_to_two_places(bytes_per_snapshot);
root_obj["last_snapshot_sync_seconds"] = m_last_snapshot_sync_seconds;
root_obj["last_snapshot_bytes"] = m_last_snapshot_bytes;
auto pending_bytes = bytes_per_snapshot * m_pending_snapshots;
if (bytes_per_second > 0 && m_pending_snapshots > 0) {
std::uint64_t seconds_until_synced = round_to_two_places(
pending_bytes / bytes_per_second);
if (seconds_until_synced >= std::numeric_limits<uint64_t>::max()) {
seconds_until_synced = std::numeric_limits<uint64_t>::max();
}
root_obj["seconds_until_synced"] = seconds_until_synced;
}
*description = json_spirit::write(
root_obj, json_spirit::remove_trailing_zeros);
local_image_locker.unlock();
remote_image_locker.unlock();
locker.unlock();
on_finish->complete(-EEXIST);
return true;
}
template <typename I>
void Replayer<I>::load_local_image_meta() {
dout(10) << dendl;
{
// reset state in case new snapshot is added while we are scanning
std::unique_lock locker{m_lock};
m_image_updated = false;
}
bool update_status = false;
{
auto local_image_ctx = m_state_builder->local_image_ctx;
std::shared_lock image_locker{local_image_ctx->image_lock};
auto image_spec = image_replayer::util::compute_image_spec(
local_image_ctx->md_ctx, local_image_ctx->name);
if (m_image_spec != image_spec) {
m_image_spec = image_spec;
update_status = true;
}
}
if (update_status) {
std::unique_lock locker{m_lock};
unregister_perf_counters();
register_perf_counters();
notify_status_updated();
}
ceph_assert(m_state_builder->local_image_meta != nullptr);
auto ctx = create_context_callback<
Replayer<I>, &Replayer<I>::handle_load_local_image_meta>(this);
m_state_builder->local_image_meta->load(ctx);
}
template <typename I>
void Replayer<I>::handle_load_local_image_meta(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0 && r != -ENOENT) {
derr << "failed to load local image-meta: " << cpp_strerror(r) << dendl;
handle_replay_complete(r, "failed to load local image-meta");
return;
}
if (r >= 0 && m_state_builder->local_image_meta->resync_requested) {
m_resync_requested = true;
dout(10) << "local image resync requested" << dendl;
handle_replay_complete(0, "resync requested");
return;
}
refresh_local_image();
}
template <typename I>
void Replayer<I>::refresh_local_image() {
if (!m_state_builder->local_image_ctx->state->is_refresh_required()) {
refresh_remote_image();
return;
}
dout(10) << dendl;
auto ctx = create_context_callback<
Replayer<I>, &Replayer<I>::handle_refresh_local_image>(this);
m_state_builder->local_image_ctx->state->refresh(ctx);
}
template <typename I>
void Replayer<I>::handle_refresh_local_image(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to refresh local image: " << cpp_strerror(r) << dendl;
handle_replay_complete(r, "failed to refresh local image");
return;
}
refresh_remote_image();
}
template <typename I>
void Replayer<I>::refresh_remote_image() {
if (!m_state_builder->remote_image_ctx->state->is_refresh_required()) {
std::unique_lock locker{m_lock};
scan_local_mirror_snapshots(&locker);
return;
}
dout(10) << dendl;
auto ctx = create_context_callback<
Replayer<I>, &Replayer<I>::handle_refresh_remote_image>(this);
m_state_builder->remote_image_ctx->state->refresh(ctx);
}
template <typename I>
void Replayer<I>::handle_refresh_remote_image(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to refresh remote image: " << cpp_strerror(r) << dendl;
handle_replay_complete(r, "failed to refresh remote image");
return;
}
std::unique_lock locker{m_lock};
scan_local_mirror_snapshots(&locker);
}
template <typename I>
void Replayer<I>::scan_local_mirror_snapshots(
std::unique_lock<ceph::mutex>* locker) {
if (is_replay_interrupted(locker)) {
return;
}
dout(10) << dendl;
m_local_snap_id_start = 0;
m_local_snap_id_end = CEPH_NOSNAP;
m_local_mirror_snap_ns = {};
m_local_object_count = 0;
m_remote_snap_id_start = 0;
m_remote_snap_id_end = CEPH_NOSNAP;
m_remote_mirror_snap_ns = {};
std::set<uint64_t> prune_snap_ids;
auto local_image_ctx = m_state_builder->local_image_ctx;
std::shared_lock image_locker{local_image_ctx->image_lock};
for (auto snap_info_it = local_image_ctx->snap_info.begin();
snap_info_it != local_image_ctx->snap_info.end(); ++snap_info_it) {
const auto& snap_ns = snap_info_it->second.snap_namespace;
auto mirror_ns = std::get_if<
cls::rbd::MirrorSnapshotNamespace>(&snap_ns);
if (mirror_ns == nullptr) {
continue;
}
dout(15) << "local mirror snapshot: id=" << snap_info_it->first << ", "
<< "mirror_ns=" << *mirror_ns << dendl;
m_local_mirror_snap_ns = *mirror_ns;
auto local_snap_id = snap_info_it->first;
if (mirror_ns->is_non_primary()) {
if (mirror_ns->complete) {
// if remote has new snapshots, we would sync from here
m_local_snap_id_start = local_snap_id;
ceph_assert(m_local_snap_id_end == CEPH_NOSNAP);
if (mirror_ns->mirror_peer_uuids.empty()) {
// no other peer will attempt to sync to this snapshot so store as
// a candidate for removal
prune_snap_ids.insert(local_snap_id);
}
} else if (mirror_ns->last_copied_object_number == 0 &&
m_local_snap_id_start > 0) {
// snapshot might be missing image state, object-map, etc, so just
// delete and re-create it if we haven't started copying data
// objects. Also only prune this snapshot since we will need the
// previous mirror snapshot for syncing. Special case exception for
// the first non-primary snapshot since we know its snapshot is
// well-formed because otherwise the mirror-image-state would have
// forced an image deletion.
prune_snap_ids.clear();
prune_snap_ids.insert(local_snap_id);
break;
} else {
// start snap will be last complete mirror snapshot or initial
// image revision
m_local_snap_id_end = local_snap_id;
break;
}
} else if (mirror_ns->is_primary()) {
if (mirror_ns->complete) {
m_local_snap_id_start = local_snap_id;
ceph_assert(m_local_snap_id_end == CEPH_NOSNAP);
} else {
derr << "incomplete local primary snapshot" << dendl;
handle_replay_complete(locker, -EINVAL,
"incomplete local primary snapshot");
return;
}
} else {
derr << "unknown local mirror snapshot state" << dendl;
handle_replay_complete(locker, -EINVAL,
"invalid local mirror snapshot state");
return;
}
}
image_locker.unlock();
if (m_local_snap_id_start > 0) {
// remove candidate that is required for delta snapshot sync
prune_snap_ids.erase(m_local_snap_id_start);
}
if (!prune_snap_ids.empty()) {
locker->unlock();
auto prune_snap_id = *prune_snap_ids.begin();
dout(5) << "pruning unused non-primary snapshot " << prune_snap_id << dendl;
prune_non_primary_snapshot(prune_snap_id);
return;
}
if (m_local_snap_id_start > 0 || m_local_snap_id_end != CEPH_NOSNAP) {
if (m_local_mirror_snap_ns.is_non_primary() &&
m_local_mirror_snap_ns.primary_mirror_uuid !=
m_state_builder->remote_mirror_uuid) {
if (m_local_mirror_snap_ns.is_orphan()) {
dout(5) << "local image being force promoted" << dendl;
handle_replay_complete(locker, 0, "orphan (force promoting)");
return;
}
// TODO support multiple peers
derr << "local image linked to unknown peer: "
<< m_local_mirror_snap_ns.primary_mirror_uuid << dendl;
handle_replay_complete(locker, -EEXIST,
"local image linked to unknown peer");
return;
} else if (m_local_mirror_snap_ns.state ==
cls::rbd::MIRROR_SNAPSHOT_STATE_PRIMARY) {
dout(5) << "local image promoted" << dendl;
handle_replay_complete(locker, 0, "force promoted");
return;
}
dout(10) << "found local mirror snapshot: "
<< "local_snap_id_start=" << m_local_snap_id_start << ", "
<< "local_snap_id_end=" << m_local_snap_id_end << ", "
<< "local_snap_ns=" << m_local_mirror_snap_ns << dendl;
if (!m_local_mirror_snap_ns.is_primary() &&
m_local_mirror_snap_ns.complete) {
// our remote sync should start after this completed snapshot
m_remote_snap_id_start = m_local_mirror_snap_ns.primary_snap_id;
}
}
// we don't have any mirror snapshots or only completed non-primary
// mirror snapshots
scan_remote_mirror_snapshots(locker);
}
template <typename I>
void Replayer<I>::scan_remote_mirror_snapshots(
std::unique_lock<ceph::mutex>* locker) {
dout(10) << dendl;
m_pending_snapshots = 0;
std::set<uint64_t> unlink_snap_ids;
bool split_brain = false;
bool remote_demoted = false;
auto remote_image_ctx = m_state_builder->remote_image_ctx;
std::shared_lock image_locker{remote_image_ctx->image_lock};
for (auto snap_info_it = remote_image_ctx->snap_info.begin();
snap_info_it != remote_image_ctx->snap_info.end(); ++snap_info_it) {
const auto& snap_ns = snap_info_it->second.snap_namespace;
auto mirror_ns = std::get_if<
cls::rbd::MirrorSnapshotNamespace>(&snap_ns);
if (mirror_ns == nullptr) {
continue;
}
dout(15) << "remote mirror snapshot: id=" << snap_info_it->first << ", "
<< "mirror_ns=" << *mirror_ns << dendl;
remote_demoted = mirror_ns->is_demoted();
if (!mirror_ns->is_primary() && !mirror_ns->is_non_primary()) {
derr << "unknown remote mirror snapshot state" << dendl;
handle_replay_complete(locker, -EINVAL,
"invalid remote mirror snapshot state");
return;
} else if (mirror_ns->mirror_peer_uuids.count(m_remote_mirror_peer_uuid) ==
0) {
dout(15) << "skipping remote snapshot due to missing mirror peer"
<< dendl;
continue;
}
auto remote_snap_id = snap_info_it->first;
if (m_local_snap_id_start > 0 || m_local_snap_id_end != CEPH_NOSNAP) {
// we have a local mirror snapshot
if (m_local_mirror_snap_ns.is_non_primary()) {
// previously validated that it was linked to remote
ceph_assert(m_local_mirror_snap_ns.primary_mirror_uuid ==
m_state_builder->remote_mirror_uuid);
if (m_remote_snap_id_end == CEPH_NOSNAP) {
// haven't found the end snap so treat this as a candidate for unlink
unlink_snap_ids.insert(remote_snap_id);
}
if (m_local_mirror_snap_ns.complete &&
m_local_mirror_snap_ns.primary_snap_id >= remote_snap_id) {
// skip past completed remote snapshot
m_remote_snap_id_start = remote_snap_id;
m_remote_mirror_snap_ns = *mirror_ns;
dout(15) << "skipping synced remote snapshot " << remote_snap_id
<< dendl;
continue;
} else if (!m_local_mirror_snap_ns.complete &&
m_local_mirror_snap_ns.primary_snap_id > remote_snap_id) {
// skip until we get to the in-progress remote snapshot
dout(15) << "skipping synced remote snapshot " << remote_snap_id
<< " while search for in-progress sync" << dendl;
m_remote_snap_id_start = remote_snap_id;
m_remote_mirror_snap_ns = *mirror_ns;
continue;
}
} else if (m_local_mirror_snap_ns.state ==
cls::rbd::MIRROR_SNAPSHOT_STATE_PRIMARY_DEMOTED) {
// find the matching demotion snapshot in remote image
ceph_assert(m_local_snap_id_start > 0);
if (mirror_ns->state ==
cls::rbd::MIRROR_SNAPSHOT_STATE_NON_PRIMARY_DEMOTED &&
mirror_ns->primary_mirror_uuid == m_local_mirror_uuid &&
mirror_ns->primary_snap_id == m_local_snap_id_start) {
dout(10) << "located matching demotion snapshot: "
<< "remote_snap_id=" << remote_snap_id << ", "
<< "local_snap_id=" << m_local_snap_id_start << dendl;
m_remote_snap_id_start = remote_snap_id;
split_brain = false;
continue;
} else if (m_remote_snap_id_start == 0) {
// still looking for our matching demotion snapshot
dout(15) << "skipping remote snapshot " << remote_snap_id << " "
<< "while searching for demotion" << dendl;
split_brain = true;
continue;
}
} else {
// should not have been able to reach this
ceph_assert(false);
}
} else if (!mirror_ns->is_primary()) {
dout(15) << "skipping non-primary remote snapshot" << dendl;
continue;
}
// found candidate snapshot to sync
++m_pending_snapshots;
if (m_remote_snap_id_end != CEPH_NOSNAP) {
continue;
}
// first primary snapshot where were are listed as a peer
m_remote_snap_id_end = remote_snap_id;
m_remote_mirror_snap_ns = *mirror_ns;
}
if (m_remote_snap_id_start != 0 &&
remote_image_ctx->snap_info.count(m_remote_snap_id_start) == 0) {
// the remote start snapshot was deleted out from under us
derr << "failed to locate remote start snapshot: "
<< "snap_id=" << m_remote_snap_id_start << dendl;
split_brain = true;
}
image_locker.unlock();
if (!split_brain) {
unlink_snap_ids.erase(m_remote_snap_id_start);
unlink_snap_ids.erase(m_remote_snap_id_end);
if (!unlink_snap_ids.empty()) {
locker->unlock();
// retry the unlinking process for a remote snapshot that we do not
// need anymore
auto remote_snap_id = *unlink_snap_ids.begin();
dout(10) << "unlinking from remote snapshot " << remote_snap_id << dendl;
unlink_peer(remote_snap_id);
return;
}
if (m_remote_snap_id_end != CEPH_NOSNAP) {
dout(10) << "found remote mirror snapshot: "
<< "remote_snap_id_start=" << m_remote_snap_id_start << ", "
<< "remote_snap_id_end=" << m_remote_snap_id_end << ", "
<< "remote_snap_ns=" << m_remote_mirror_snap_ns << dendl;
if (m_remote_mirror_snap_ns.complete) {
locker->unlock();
if (m_local_snap_id_end != CEPH_NOSNAP &&
!m_local_mirror_snap_ns.complete) {
// attempt to resume image-sync
dout(10) << "local image contains in-progress mirror snapshot"
<< dendl;
get_local_image_state();
} else {
copy_snapshots();
}
return;
} else {
// might have raced with the creation of a remote mirror snapshot
// so we will need to refresh and rescan once it completes
dout(15) << "remote mirror snapshot not complete" << dendl;
}
}
}
if (m_image_updated) {
// received update notification while scanning image, restart ...
m_image_updated = false;
locker->unlock();
dout(10) << "restarting snapshot scan due to remote update notification"
<< dendl;
load_local_image_meta();
return;
}
if (is_replay_interrupted(locker)) {
return;
} else if (split_brain) {
derr << "split-brain detected: failed to find matching non-primary "
<< "snapshot in remote image: "
<< "local_snap_id_start=" << m_local_snap_id_start << ", "
<< "local_snap_ns=" << m_local_mirror_snap_ns << dendl;
handle_replay_complete(locker, -EEXIST, "split-brain");
return;
} else if (remote_demoted) {
dout(10) << "remote image demoted" << dendl;
handle_replay_complete(locker, -EREMOTEIO, "remote image demoted");
return;
}
dout(10) << "all remote snapshots synced: idling waiting for new snapshot"
<< dendl;
ceph_assert(m_state == STATE_REPLAYING);
m_state = STATE_IDLE;
notify_status_updated();
}
template <typename I>
void Replayer<I>::prune_non_primary_snapshot(uint64_t snap_id) {
dout(10) << "snap_id=" << snap_id << dendl;
auto local_image_ctx = m_state_builder->local_image_ctx;
bool snap_valid = false;
cls::rbd::SnapshotNamespace snap_namespace;
std::string snap_name;
{
std::shared_lock image_locker{local_image_ctx->image_lock};
auto snap_info = local_image_ctx->get_snap_info(snap_id);
if (snap_info != nullptr) {
snap_valid = true;
snap_namespace = snap_info->snap_namespace;
snap_name = snap_info->name;
ceph_assert(std::holds_alternative<cls::rbd::MirrorSnapshotNamespace>(
snap_namespace));
}
}
if (!snap_valid) {
load_local_image_meta();
return;
}
auto ctx = create_context_callback<
Replayer<I>, &Replayer<I>::handle_prune_non_primary_snapshot>(this);
local_image_ctx->operations->snap_remove(snap_namespace, snap_name, ctx);
}
template <typename I>
void Replayer<I>::handle_prune_non_primary_snapshot(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0 && r != -ENOENT) {
derr << "failed to prune non-primary snapshot: " << cpp_strerror(r)
<< dendl;
handle_replay_complete(r, "failed to prune non-primary snapshot");
return;
}
if (is_replay_interrupted()) {
return;
}
load_local_image_meta();
}
template <typename I>
void Replayer<I>::copy_snapshots() {
dout(10) << "remote_snap_id_start=" << m_remote_snap_id_start << ", "
<< "remote_snap_id_end=" << m_remote_snap_id_end << ", "
<< "local_snap_id_start=" << m_local_snap_id_start << dendl;
ceph_assert(m_remote_snap_id_start != CEPH_NOSNAP);
ceph_assert(m_remote_snap_id_end > 0 &&
m_remote_snap_id_end != CEPH_NOSNAP);
ceph_assert(m_local_snap_id_start != CEPH_NOSNAP);
m_local_mirror_snap_ns = {};
auto ctx = create_context_callback<
Replayer<I>, &Replayer<I>::handle_copy_snapshots>(this);
auto req = librbd::deep_copy::SnapshotCopyRequest<I>::create(
m_state_builder->remote_image_ctx, m_state_builder->local_image_ctx,
m_remote_snap_id_start, m_remote_snap_id_end, m_local_snap_id_start,
false, m_threads->work_queue, &m_local_mirror_snap_ns.snap_seqs,
ctx);
req->send();
}
template <typename I>
void Replayer<I>::handle_copy_snapshots(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to copy snapshots from remote to local image: "
<< cpp_strerror(r) << dendl;
handle_replay_complete(
r, "failed to copy snapshots from remote to local image");
return;
}
dout(10) << "remote_snap_id_start=" << m_remote_snap_id_start << ", "
<< "remote_snap_id_end=" << m_remote_snap_id_end << ", "
<< "local_snap_id_start=" << m_local_snap_id_start << ", "
<< "snap_seqs=" << m_local_mirror_snap_ns.snap_seqs << dendl;
get_remote_image_state();
}
template <typename I>
void Replayer<I>::get_remote_image_state() {
dout(10) << dendl;
auto ctx = create_context_callback<
Replayer<I>, &Replayer<I>::handle_get_remote_image_state>(this);
auto req = librbd::mirror::snapshot::GetImageStateRequest<I>::create(
m_state_builder->remote_image_ctx, m_remote_snap_id_end,
&m_image_state, ctx);
req->send();
}
template <typename I>
void Replayer<I>::handle_get_remote_image_state(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to retrieve remote snapshot image state: "
<< cpp_strerror(r) << dendl;
handle_replay_complete(r, "failed to retrieve remote snapshot image state");
return;
}
create_non_primary_snapshot();
}
template <typename I>
void Replayer<I>::get_local_image_state() {
dout(10) << dendl;
ceph_assert(m_local_snap_id_end != CEPH_NOSNAP);
auto ctx = create_context_callback<
Replayer<I>, &Replayer<I>::handle_get_local_image_state>(this);
auto req = librbd::mirror::snapshot::GetImageStateRequest<I>::create(
m_state_builder->local_image_ctx, m_local_snap_id_end,
&m_image_state, ctx);
req->send();
}
template <typename I>
void Replayer<I>::handle_get_local_image_state(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to retrieve local snapshot image state: "
<< cpp_strerror(r) << dendl;
handle_replay_complete(r, "failed to retrieve local snapshot image state");
return;
}
request_sync();
}
template <typename I>
void Replayer<I>::create_non_primary_snapshot() {
auto local_image_ctx = m_state_builder->local_image_ctx;
if (m_local_snap_id_start > 0) {
std::shared_lock local_image_locker{local_image_ctx->image_lock};
auto local_snap_info_it = local_image_ctx->snap_info.find(
m_local_snap_id_start);
if (local_snap_info_it == local_image_ctx->snap_info.end()) {
local_image_locker.unlock();
derr << "failed to locate local snapshot " << m_local_snap_id_start
<< dendl;
handle_replay_complete(-ENOENT, "failed to locate local start snapshot");
return;
}
auto mirror_ns = std::get_if<cls::rbd::MirrorSnapshotNamespace>(
&local_snap_info_it->second.snap_namespace);
ceph_assert(mirror_ns != nullptr);
auto remote_image_ctx = m_state_builder->remote_image_ctx;
std::shared_lock remote_image_locker{remote_image_ctx->image_lock};
// (re)build a full mapping from remote to local snap ids for all user
// snapshots to support applying image state in the future
for (auto& [remote_snap_id, remote_snap_info] :
remote_image_ctx->snap_info) {
if (remote_snap_id >= m_remote_snap_id_end) {
break;
}
// we can ignore all non-user snapshots since image state only includes
// user snapshots
if (!std::holds_alternative<cls::rbd::UserSnapshotNamespace>(
remote_snap_info.snap_namespace)) {
continue;
}
uint64_t local_snap_id = CEPH_NOSNAP;
if (mirror_ns->is_demoted() && !m_remote_mirror_snap_ns.is_demoted()) {
// if we are creating a non-primary snapshot following a demotion,
// re-build the full snapshot sequence since we don't have a valid
// snapshot mapping
auto local_snap_id_it = local_image_ctx->snap_ids.find(
{remote_snap_info.snap_namespace, remote_snap_info.name});
if (local_snap_id_it != local_image_ctx->snap_ids.end()) {
local_snap_id = local_snap_id_it->second;
}
} else {
auto snap_seq_it = mirror_ns->snap_seqs.find(remote_snap_id);
if (snap_seq_it != mirror_ns->snap_seqs.end()) {
local_snap_id = snap_seq_it->second;
}
}
if (m_local_mirror_snap_ns.snap_seqs.count(remote_snap_id) == 0 &&
local_snap_id != CEPH_NOSNAP) {
dout(15) << "mapping remote snapshot " << remote_snap_id << " to "
<< "local snapshot " << local_snap_id << dendl;
m_local_mirror_snap_ns.snap_seqs[remote_snap_id] = local_snap_id;
}
}
}
dout(10) << "demoted=" << m_remote_mirror_snap_ns.is_demoted() << ", "
<< "primary_mirror_uuid="
<< m_state_builder->remote_mirror_uuid << ", "
<< "primary_snap_id=" << m_remote_snap_id_end << ", "
<< "snap_seqs=" << m_local_mirror_snap_ns.snap_seqs << dendl;
auto ctx = create_context_callback<
Replayer<I>, &Replayer<I>::handle_create_non_primary_snapshot>(this);
auto req = librbd::mirror::snapshot::CreateNonPrimaryRequest<I>::create(
local_image_ctx, m_remote_mirror_snap_ns.is_demoted(),
m_state_builder->remote_mirror_uuid, m_remote_snap_id_end,
m_local_mirror_snap_ns.snap_seqs, m_image_state, &m_local_snap_id_end, ctx);
req->send();
}
template <typename I>
void Replayer<I>::handle_create_non_primary_snapshot(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to create local mirror snapshot: " << cpp_strerror(r)
<< dendl;
handle_replay_complete(r, "failed to create local mirror snapshot");
return;
}
dout(15) << "local_snap_id_end=" << m_local_snap_id_end << dendl;
update_mirror_image_state();
}
template <typename I>
void Replayer<I>::update_mirror_image_state() {
if (m_local_snap_id_start > 0) {
request_sync();
return;
}
// a newly created non-primary image has a local mirror state of CREATING
// until this point so that we could avoid preserving the image until
// the first non-primary snapshot linked the two images together.
dout(10) << dendl;
auto ctx = create_context_callback<
Replayer<I>, &Replayer<I>::handle_update_mirror_image_state>(this);
auto req = librbd::mirror::ImageStateUpdateRequest<I>::create(
m_state_builder->local_image_ctx->md_ctx,
m_state_builder->local_image_ctx->id,
cls::rbd::MIRROR_IMAGE_STATE_ENABLED, {}, ctx);
req->send();
}
template <typename I>
void Replayer<I>::handle_update_mirror_image_state(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to update local mirror image state: " << cpp_strerror(r)
<< dendl;
handle_replay_complete(r, "failed to update local mirror image state");
return;
}
request_sync();
}
template <typename I>
void Replayer<I>::request_sync() {
if (m_remote_mirror_snap_ns.clean_since_snap_id == m_remote_snap_id_start) {
dout(10) << "skipping unnecessary image copy: "
<< "remote_snap_id_start=" << m_remote_snap_id_start << ", "
<< "remote_mirror_snap_ns=" << m_remote_mirror_snap_ns << dendl;
apply_image_state();
return;
}
dout(10) << dendl;
std::unique_lock locker{m_lock};
if (is_replay_interrupted(&locker)) {
return;
}
auto ctx = create_async_context_callback(
m_threads->work_queue, create_context_callback<
Replayer<I>, &Replayer<I>::handle_request_sync>(this));
m_instance_watcher->notify_sync_request(m_state_builder->local_image_ctx->id,
ctx);
}
template <typename I>
void Replayer<I>::handle_request_sync(int r) {
dout(10) << "r=" << r << dendl;
std::unique_lock locker{m_lock};
if (is_replay_interrupted(&locker)) {
return;
} else if (r == -ECANCELED) {
dout(5) << "image-sync canceled" << dendl;
handle_replay_complete(&locker, r, "image-sync canceled");
return;
} else if (r < 0) {
derr << "failed to request image-sync: " << cpp_strerror(r) << dendl;
handle_replay_complete(&locker, r, "failed to request image-sync");
return;
}
m_sync_in_progress = true;
locker.unlock();
copy_image();
}
template <typename I>
void Replayer<I>::copy_image() {
dout(10) << "remote_snap_id_start=" << m_remote_snap_id_start << ", "
<< "remote_snap_id_end=" << m_remote_snap_id_end << ", "
<< "local_snap_id_start=" << m_local_snap_id_start << ", "
<< "last_copied_object_number="
<< m_local_mirror_snap_ns.last_copied_object_number << ", "
<< "snap_seqs=" << m_local_mirror_snap_ns.snap_seqs << dendl;
m_snapshot_bytes = 0;
m_snapshot_replay_start = ceph_clock_now();
m_deep_copy_handler = new DeepCopyHandler(this);
auto ctx = create_context_callback<
Replayer<I>, &Replayer<I>::handle_copy_image>(this);
auto req = librbd::deep_copy::ImageCopyRequest<I>::create(
m_state_builder->remote_image_ctx, m_state_builder->local_image_ctx,
m_remote_snap_id_start, m_remote_snap_id_end, m_local_snap_id_start, false,
(m_local_mirror_snap_ns.last_copied_object_number > 0 ?
librbd::deep_copy::ObjectNumber{
m_local_mirror_snap_ns.last_copied_object_number} :
librbd::deep_copy::ObjectNumber{}),
m_local_mirror_snap_ns.snap_seqs, m_deep_copy_handler, ctx);
req->send();
}
template <typename I>
void Replayer<I>::handle_copy_image(int r) {
dout(10) << "r=" << r << dendl;
delete m_deep_copy_handler;
m_deep_copy_handler = nullptr;
if (r < 0) {
derr << "failed to copy remote image to local image: " << cpp_strerror(r)
<< dendl;
handle_replay_complete(r, "failed to copy remote image");
return;
}
{
std::unique_lock locker{m_lock};
m_last_snapshot_bytes = m_snapshot_bytes;
m_bytes_per_snapshot(m_snapshot_bytes);
utime_t duration = ceph_clock_now() - m_snapshot_replay_start;
m_last_snapshot_sync_seconds = duration.sec();
if (g_snapshot_perf_counters) {
g_snapshot_perf_counters->inc(l_rbd_mirror_snapshot_sync_bytes,
m_snapshot_bytes);
g_snapshot_perf_counters->inc(l_rbd_mirror_snapshot_snapshots);
g_snapshot_perf_counters->tinc(l_rbd_mirror_snapshot_sync_time,
duration);
}
if (m_perf_counters) {
m_perf_counters->inc(l_rbd_mirror_snapshot_sync_bytes, m_snapshot_bytes);
m_perf_counters->inc(l_rbd_mirror_snapshot_snapshots);
m_perf_counters->tinc(l_rbd_mirror_snapshot_sync_time, duration);
m_perf_counters->tset(l_rbd_mirror_snapshot_last_sync_time, duration);
m_perf_counters->set(l_rbd_mirror_snapshot_last_sync_bytes,
m_snapshot_bytes);
}
}
apply_image_state();
}
template <typename I>
void Replayer<I>::handle_copy_image_progress(uint64_t object_number,
uint64_t object_count) {
dout(10) << "object_number=" << object_number << ", "
<< "object_count=" << object_count << dendl;
std::unique_lock locker{m_lock};
m_local_mirror_snap_ns.last_copied_object_number = std::min(
object_number, object_count);
m_local_object_count = object_count;
update_non_primary_snapshot(false);
}
template <typename I>
void Replayer<I>::handle_copy_image_read(uint64_t bytes_read) {
dout(20) << "bytes_read=" << bytes_read << dendl;
std::unique_lock locker{m_lock};
m_bytes_per_second(bytes_read);
m_snapshot_bytes += bytes_read;
}
template <typename I>
void Replayer<I>::apply_image_state() {
dout(10) << dendl;
auto ctx = create_context_callback<
Replayer<I>, &Replayer<I>::handle_apply_image_state>(this);
auto req = ApplyImageStateRequest<I>::create(
m_local_mirror_uuid,
m_state_builder->remote_mirror_uuid,
m_state_builder->local_image_ctx,
m_state_builder->remote_image_ctx,
m_image_state, ctx);
req->send();
}
template <typename I>
void Replayer<I>::handle_apply_image_state(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0 && r != -ENOENT) {
derr << "failed to apply remote image state to local image: "
<< cpp_strerror(r) << dendl;
handle_replay_complete(r, "failed to apply remote image state");
return;
}
std::unique_lock locker{m_lock};
update_non_primary_snapshot(true);
}
template <typename I>
void Replayer<I>::update_non_primary_snapshot(bool complete) {
ceph_assert(ceph_mutex_is_locked_by_me(m_lock));
if (!complete) {
// disallow two in-flight updates if this isn't the completion of the sync
if (m_updating_sync_point) {
return;
}
m_updating_sync_point = true;
} else {
m_local_mirror_snap_ns.complete = true;
}
dout(10) << dendl;
librados::ObjectWriteOperation op;
librbd::cls_client::mirror_image_snapshot_set_copy_progress(
&op, m_local_snap_id_end, m_local_mirror_snap_ns.complete,
m_local_mirror_snap_ns.last_copied_object_number);
auto ctx = new C_TrackedOp(
m_in_flight_op_tracker, new LambdaContext([this, complete](int r) {
handle_update_non_primary_snapshot(complete, r);
}));
auto aio_comp = create_rados_callback(ctx);
int r = m_state_builder->local_image_ctx->md_ctx.aio_operate(
m_state_builder->local_image_ctx->header_oid, aio_comp, &op);
ceph_assert(r == 0);
aio_comp->release();
}
template <typename I>
void Replayer<I>::handle_update_non_primary_snapshot(bool complete, int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to update local snapshot progress: " << cpp_strerror(r)
<< dendl;
if (complete) {
// only fail if this was the final update
handle_replay_complete(r, "failed to update local snapshot progress");
return;
}
}
if (!complete) {
// periodic sync-point update -- do not advance state machine
std::unique_lock locker{m_lock};
ceph_assert(m_updating_sync_point);
m_updating_sync_point = false;
return;
}
notify_image_update();
}
template <typename I>
void Replayer<I>::notify_image_update() {
dout(10) << dendl;
auto ctx = create_context_callback<
Replayer<I>, &Replayer<I>::handle_notify_image_update>(this);
m_state_builder->local_image_ctx->notify_update(ctx);
}
template <typename I>
void Replayer<I>::handle_notify_image_update(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to notify local image update: " << cpp_strerror(r) << dendl;
}
unlink_peer(m_remote_snap_id_start);
}
template <typename I>
void Replayer<I>::unlink_peer(uint64_t remote_snap_id) {
if (remote_snap_id == 0) {
finish_sync();
return;
}
// local snapshot fully synced -- we no longer depend on the sync
// start snapshot in the remote image
dout(10) << "remote_snap_id=" << remote_snap_id << dendl;
auto ctx = create_context_callback<
Replayer<I>, &Replayer<I>::handle_unlink_peer>(this);
auto req = librbd::mirror::snapshot::UnlinkPeerRequest<I>::create(
m_state_builder->remote_image_ctx, remote_snap_id,
m_remote_mirror_peer_uuid, false, ctx);
req->send();
}
template <typename I>
void Replayer<I>::handle_unlink_peer(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0 && r != -ENOENT) {
derr << "failed to unlink local peer from remote image: " << cpp_strerror(r)
<< dendl;
handle_replay_complete(r, "failed to unlink local peer from remote image");
return;
}
finish_sync();
}
template <typename I>
void Replayer<I>::finish_sync() {
dout(10) << dendl;
{
std::unique_lock locker{m_lock};
notify_status_updated();
if (m_sync_in_progress) {
m_sync_in_progress = false;
m_instance_watcher->notify_sync_complete(
m_state_builder->local_image_ctx->id);
}
}
if (is_replay_interrupted()) {
return;
}
load_local_image_meta();
}
template <typename I>
void Replayer<I>::register_local_update_watcher() {
dout(10) << dendl;
m_update_watch_ctx = new C_UpdateWatchCtx(this);
int r = m_state_builder->local_image_ctx->state->register_update_watcher(
m_update_watch_ctx, &m_local_update_watcher_handle);
auto ctx = create_context_callback<
Replayer<I>, &Replayer<I>::handle_register_local_update_watcher>(this);
m_threads->work_queue->queue(ctx, r);
}
template <typename I>
void Replayer<I>::handle_register_local_update_watcher(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to register local update watcher: " << cpp_strerror(r)
<< dendl;
handle_replay_complete(r, "failed to register local image update watcher");
m_state = STATE_COMPLETE;
delete m_update_watch_ctx;
m_update_watch_ctx = nullptr;
Context* on_init = nullptr;
std::swap(on_init, m_on_init_shutdown);
on_init->complete(r);
return;
}
register_remote_update_watcher();
}
template <typename I>
void Replayer<I>::register_remote_update_watcher() {
dout(10) << dendl;
int r = m_state_builder->remote_image_ctx->state->register_update_watcher(
m_update_watch_ctx, &m_remote_update_watcher_handle);
auto ctx = create_context_callback<
Replayer<I>, &Replayer<I>::handle_register_remote_update_watcher>(this);
m_threads->work_queue->queue(ctx, r);
}
template <typename I>
void Replayer<I>::handle_register_remote_update_watcher(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to register remote update watcher: " << cpp_strerror(r)
<< dendl;
handle_replay_complete(r, "failed to register remote image update watcher");
m_state = STATE_COMPLETE;
unregister_local_update_watcher();
return;
}
m_state = STATE_REPLAYING;
Context* on_init = nullptr;
std::swap(on_init, m_on_init_shutdown);
on_init->complete(0);
// delay initial snapshot scan until after we have alerted
// image replayer that we have initialized in case an error
// occurs
{
std::unique_lock locker{m_lock};
notify_status_updated();
}
load_local_image_meta();
}
template <typename I>
void Replayer<I>::unregister_remote_update_watcher() {
dout(10) << dendl;
auto ctx = create_context_callback<
Replayer<I>,
&Replayer<I>::handle_unregister_remote_update_watcher>(this);
m_state_builder->remote_image_ctx->state->unregister_update_watcher(
m_remote_update_watcher_handle, ctx);
}
template <typename I>
void Replayer<I>::handle_unregister_remote_update_watcher(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to unregister remote update watcher: " << cpp_strerror(r)
<< dendl;
}
unregister_local_update_watcher();
}
template <typename I>
void Replayer<I>::unregister_local_update_watcher() {
dout(10) << dendl;
auto ctx = create_context_callback<
Replayer<I>,
&Replayer<I>::handle_unregister_local_update_watcher>(this);
m_state_builder->local_image_ctx->state->unregister_update_watcher(
m_local_update_watcher_handle, ctx);
}
template <typename I>
void Replayer<I>::handle_unregister_local_update_watcher(int r) {
dout(10) << "r=" << r << dendl;
if (r < 0) {
derr << "failed to unregister local update watcher: " << cpp_strerror(r)
<< dendl;
}
delete m_update_watch_ctx;
m_update_watch_ctx = nullptr;
wait_for_in_flight_ops();
}
template <typename I>
void Replayer<I>::wait_for_in_flight_ops() {
dout(10) << dendl;
auto ctx = create_async_context_callback(
m_threads->work_queue, create_context_callback<
Replayer<I>, &Replayer<I>::handle_wait_for_in_flight_ops>(this));
m_in_flight_op_tracker.wait_for_ops(ctx);
}
template <typename I>
void Replayer<I>::handle_wait_for_in_flight_ops(int r) {
dout(10) << "r=" << r << dendl;
Context* on_shutdown = nullptr;
{
std::unique_lock locker{m_lock};
ceph_assert(m_on_init_shutdown != nullptr);
std::swap(on_shutdown, m_on_init_shutdown);
}
on_shutdown->complete(m_error_code);
}
template <typename I>
void Replayer<I>::handle_image_update_notify() {
dout(10) << dendl;
std::unique_lock locker{m_lock};
if (m_state == STATE_REPLAYING) {
dout(15) << "flagging snapshot rescan required" << dendl;
m_image_updated = true;
} else if (m_state == STATE_IDLE) {
m_state = STATE_REPLAYING;
locker.unlock();
dout(15) << "restarting idle replayer" << dendl;
load_local_image_meta();
}
}
template <typename I>
void Replayer<I>::handle_replay_complete(int r,
const std::string& description) {
std::unique_lock locker{m_lock};
handle_replay_complete(&locker, r, description);
}
template <typename I>
void Replayer<I>::handle_replay_complete(std::unique_lock<ceph::mutex>* locker,
int r,
const std::string& description) {
ceph_assert(ceph_mutex_is_locked_by_me(m_lock));
if (m_sync_in_progress) {
m_sync_in_progress = false;
m_instance_watcher->notify_sync_complete(
m_state_builder->local_image_ctx->id);
}
// don't set error code and description if resuming a pending
// shutdown
if (is_replay_interrupted(locker)) {
return;
}
if (m_error_code == 0) {
m_error_code = r;
m_error_description = description;
}
if (m_state != STATE_REPLAYING && m_state != STATE_IDLE) {
return;
}
m_state = STATE_COMPLETE;
notify_status_updated();
}
template <typename I>
void Replayer<I>::notify_status_updated() {
ceph_assert(ceph_mutex_is_locked_by_me(m_lock));
dout(10) << dendl;
auto ctx = new C_TrackedOp(m_in_flight_op_tracker, new LambdaContext(
[this](int) {
m_replayer_listener->handle_notification();
}));
m_threads->work_queue->queue(ctx, 0);
}
template <typename I>
bool Replayer<I>::is_replay_interrupted() {
std::unique_lock locker{m_lock};
return is_replay_interrupted(&locker);
}
template <typename I>
bool Replayer<I>::is_replay_interrupted(std::unique_lock<ceph::mutex>* locker) {
if (m_state == STATE_COMPLETE) {
locker->unlock();
dout(10) << "resuming pending shut down" << dendl;
unregister_remote_update_watcher();
return true;
}
return false;
}
template <typename I>
void Replayer<I>::register_perf_counters() {
dout(5) << dendl;
ceph_assert(ceph_mutex_is_locked_by_me(m_lock));
ceph_assert(m_perf_counters == nullptr);
auto cct = static_cast<CephContext *>(m_state_builder->local_image_ctx->cct);
auto prio = cct->_conf.get_val<int64_t>("rbd_mirror_image_perf_stats_prio");
auto local_image_ctx = m_state_builder->local_image_ctx;
std::string labels = ceph::perf_counters::key_create(
"rbd_mirror_snapshot_image",
{{"pool", local_image_ctx->md_ctx.get_pool_name()},
{"namespace", local_image_ctx->md_ctx.get_namespace()},
{"image", local_image_ctx->name}});
PerfCountersBuilder plb(g_ceph_context, labels, l_rbd_mirror_snapshot_first,
l_rbd_mirror_snapshot_last);
plb.add_u64_counter(l_rbd_mirror_snapshot_snapshots, "snapshots",
"Number of snapshots synced", nullptr, prio);
plb.add_time_avg(l_rbd_mirror_snapshot_sync_time, "sync_time",
"Average sync time", nullptr, prio);
plb.add_u64_counter(l_rbd_mirror_snapshot_sync_bytes, "sync_bytes",
"Total bytes synced", nullptr, prio, unit_t(UNIT_BYTES));
plb.add_time(l_rbd_mirror_snapshot_remote_timestamp, "remote_timestamp",
"Timestamp of the remote snapshot", nullptr, prio);
plb.add_time(l_rbd_mirror_snapshot_local_timestamp, "local_timestamp",
"Timestamp of the local snapshot", nullptr, prio);
plb.add_time(l_rbd_mirror_snapshot_last_sync_time, "last_sync_time",
"Time taken to sync the last snapshot", nullptr, prio);
plb.add_u64(l_rbd_mirror_snapshot_last_sync_bytes, "last_sync_bytes",
"Bytes synced for the last snapshot", nullptr, prio,
unit_t(UNIT_BYTES));
m_perf_counters = plb.create_perf_counters();
g_ceph_context->get_perfcounters_collection()->add(m_perf_counters);
}
template <typename I>
void Replayer<I>::unregister_perf_counters() {
dout(5) << dendl;
ceph_assert(ceph_mutex_is_locked_by_me(m_lock));
PerfCounters *perf_counters = nullptr;
std::swap(perf_counters, m_perf_counters);
if (perf_counters != nullptr) {
g_ceph_context->get_perfcounters_collection()->remove(perf_counters);
delete perf_counters;
}
}
} // namespace snapshot
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_replayer::snapshot::Replayer<librbd::ImageCtx>;
| 53,099 | 31.49694 | 81 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/snapshot/Replayer.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_REPLAYER_SNAPSHOT_REPLAYER_H
#define RBD_MIRROR_IMAGE_REPLAYER_SNAPSHOT_REPLAYER_H
#include "tools/rbd_mirror/image_replayer/Replayer.h"
#include "common/ceph_mutex.h"
#include "common/AsyncOpTracker.h"
#include "cls/rbd/cls_rbd_types.h"
#include "librbd/mirror/snapshot/Types.h"
#include "tools/rbd_mirror/image_replayer/TimeRollingMean.h"
#include <boost/accumulators/accumulators.hpp>
#include <boost/accumulators/statistics/stats.hpp>
#include <boost/accumulators/statistics/rolling_mean.hpp>
#include <string>
#include <type_traits>
namespace librbd {
struct ImageCtx;
namespace snapshot { template <typename I> class Replay; }
} // namespace librbd
namespace rbd {
namespace mirror {
template <typename> struct InstanceWatcher;
class PoolMetaCache;
template <typename> struct Threads;
namespace image_replayer {
struct ReplayerListener;
namespace snapshot {
template <typename> class EventPreprocessor;
template <typename> class ReplayStatusFormatter;
template <typename> class StateBuilder;
template <typename ImageCtxT>
class Replayer : public image_replayer::Replayer {
public:
static Replayer* create(
Threads<ImageCtxT>* threads,
InstanceWatcher<ImageCtxT>* instance_watcher,
const std::string& local_mirror_uuid,
PoolMetaCache* pool_meta_cache,
StateBuilder<ImageCtxT>* state_builder,
ReplayerListener* replayer_listener) {
return new Replayer(threads, instance_watcher, local_mirror_uuid,
pool_meta_cache, state_builder, replayer_listener);
}
Replayer(
Threads<ImageCtxT>* threads,
InstanceWatcher<ImageCtxT>* instance_watcher,
const std::string& local_mirror_uuid,
PoolMetaCache* pool_meta_cache,
StateBuilder<ImageCtxT>* state_builder,
ReplayerListener* replayer_listener);
~Replayer();
void destroy() override {
delete this;
}
void init(Context* on_finish) override;
void shut_down(Context* on_finish) override;
void flush(Context* on_finish) override;
bool get_replay_status(std::string* description, Context* on_finish) override;
bool is_replaying() const override {
std::unique_lock locker{m_lock};
return (m_state == STATE_REPLAYING || m_state == STATE_IDLE);
}
bool is_resync_requested() const override {
std::unique_lock locker{m_lock};
return m_resync_requested;
}
int get_error_code() const override {
std::unique_lock locker(m_lock);
return m_error_code;
}
std::string get_error_description() const override {
std::unique_lock locker(m_lock);
return m_error_description;
}
std::string get_image_spec() const {
std::unique_lock locker(m_lock);
return m_image_spec;
}
private:
/**
* @verbatim
*
* <init>
* |
* v
* REGISTER_LOCAL_UPDATE_WATCHER
* |
* v
* REGISTER_REMOTE_UPDATE_WATCHER
* |
* v
* LOAD_LOCAL_IMAGE_META <----------------------------\
* | |
* v (skip if not needed) |
* REFRESH_LOCAL_IMAGE |
* | |
* v (skip if not needed) |
* REFRESH_REMOTE_IMAGE |
* | |
* | (unused non-primary snapshot) |
* |\--------------> PRUNE_NON_PRIMARY_SNAPSHOT---/|
* | |
* | (interrupted sync) |
* |\--------------> GET_LOCAL_IMAGE_STATE ------\ |
* | | |
* | (new snapshot) | |
* |\--------------> COPY_SNAPSHOTS | |
* | | | |
* | v | |
* | GET_REMOTE_IMAGE_STATE | |
* | | | |
* | v | |
* | CREATE_NON_PRIMARY_SNAPSHOT | |
* | | | |
* | v (skip if not needed)| |
* | UPDATE_MIRROR_IMAGE_STATE | |
* | | | |
* | |/--------------------/ |
* | | |
* | v |
* | REQUEST_SYNC |
* | | |
* | v |
* | COPY_IMAGE |
* | | |
* | v |
* | APPLY_IMAGE_STATE |
* | | |
* | v |
* | UPDATE_NON_PRIMARY_SNAPSHOT |
* | | |
* | v |
* | NOTIFY_IMAGE_UPDATE |
* | | |
* | (interrupted unlink) v |
* |\--------------> UNLINK_PEER |
* | | |
* | v |
* | NOTIFY_LISTENER |
* | | |
* | \----------------------/|
* | |
* | (remote demoted) |
* \---------------> NOTIFY_LISTENER |
* | | |
* |/--------------------/ |
* | |
* | (update notification) |
* <idle> --------------------------------------------/
* |
* v
* <shut down>
* |
* v
* UNREGISTER_REMOTE_UPDATE_WATCHER
* |
* v
* UNREGISTER_LOCAL_UPDATE_WATCHER
* |
* v
* WAIT_FOR_IN_FLIGHT_OPS
* |
* v
* <finish>
*
* @endverbatim
*/
enum State {
STATE_INIT,
STATE_REPLAYING,
STATE_IDLE,
STATE_COMPLETE
};
struct C_UpdateWatchCtx;
struct DeepCopyHandler;
Threads<ImageCtxT>* m_threads;
InstanceWatcher<ImageCtxT>* m_instance_watcher;
std::string m_local_mirror_uuid;
PoolMetaCache* m_pool_meta_cache;
StateBuilder<ImageCtxT>* m_state_builder;
ReplayerListener* m_replayer_listener;
mutable ceph::mutex m_lock;
State m_state = STATE_INIT;
std::string m_image_spec;
Context* m_on_init_shutdown = nullptr;
bool m_resync_requested = false;
int m_error_code = 0;
std::string m_error_description;
C_UpdateWatchCtx* m_update_watch_ctx = nullptr;
uint64_t m_local_update_watcher_handle = 0;
uint64_t m_remote_update_watcher_handle = 0;
bool m_image_updated = false;
AsyncOpTracker m_in_flight_op_tracker;
uint64_t m_local_snap_id_start = 0;
uint64_t m_local_snap_id_end = CEPH_NOSNAP;
cls::rbd::MirrorSnapshotNamespace m_local_mirror_snap_ns;
uint64_t m_local_object_count = 0;
std::string m_remote_mirror_peer_uuid;
uint64_t m_remote_snap_id_start = 0;
uint64_t m_remote_snap_id_end = CEPH_NOSNAP;
cls::rbd::MirrorSnapshotNamespace m_remote_mirror_snap_ns;
librbd::mirror::snapshot::ImageState m_image_state;
DeepCopyHandler* m_deep_copy_handler = nullptr;
TimeRollingMean m_bytes_per_second;
uint64_t m_last_snapshot_sync_seconds = 0;
uint64_t m_snapshot_bytes = 0;
uint64_t m_last_snapshot_bytes = 0;
boost::accumulators::accumulator_set<
uint64_t, boost::accumulators::stats<
boost::accumulators::tag::rolling_mean>> m_bytes_per_snapshot{
boost::accumulators::tag::rolling_window::window_size = 2};
utime_t m_snapshot_replay_start;
uint32_t m_pending_snapshots = 0;
bool m_remote_image_updated = false;
bool m_updating_sync_point = false;
bool m_sync_in_progress = false;
PerfCounters *m_perf_counters = nullptr;
void load_local_image_meta();
void handle_load_local_image_meta(int r);
void refresh_local_image();
void handle_refresh_local_image(int r);
void refresh_remote_image();
void handle_refresh_remote_image(int r);
void scan_local_mirror_snapshots(std::unique_lock<ceph::mutex>* locker);
void scan_remote_mirror_snapshots(std::unique_lock<ceph::mutex>* locker);
void prune_non_primary_snapshot(uint64_t snap_id);
void handle_prune_non_primary_snapshot(int r);
void copy_snapshots();
void handle_copy_snapshots(int r);
void get_remote_image_state();
void handle_get_remote_image_state(int r);
void get_local_image_state();
void handle_get_local_image_state(int r);
void create_non_primary_snapshot();
void handle_create_non_primary_snapshot(int r);
void update_mirror_image_state();
void handle_update_mirror_image_state(int r);
void request_sync();
void handle_request_sync(int r);
void copy_image();
void handle_copy_image(int r);
void handle_copy_image_progress(uint64_t object_number,
uint64_t object_count);
void handle_copy_image_read(uint64_t bytes_read);
void apply_image_state();
void handle_apply_image_state(int r);
void update_non_primary_snapshot(bool complete);
void handle_update_non_primary_snapshot(bool complete, int r);
void notify_image_update();
void handle_notify_image_update(int r);
void unlink_peer(uint64_t remote_snap_id);
void handle_unlink_peer(int r);
void finish_sync();
void register_local_update_watcher();
void handle_register_local_update_watcher(int r);
void register_remote_update_watcher();
void handle_register_remote_update_watcher(int r);
void unregister_remote_update_watcher();
void handle_unregister_remote_update_watcher(int r);
void unregister_local_update_watcher();
void handle_unregister_local_update_watcher(int r);
void wait_for_in_flight_ops();
void handle_wait_for_in_flight_ops(int r);
void handle_image_update_notify();
void handle_replay_complete(int r, const std::string& description);
void handle_replay_complete(std::unique_lock<ceph::mutex>* locker,
int r, const std::string& description);
void notify_status_updated();
bool is_replay_interrupted();
bool is_replay_interrupted(std::unique_lock<ceph::mutex>* lock);
void register_perf_counters();
void unregister_perf_counters();
};
} // namespace snapshot
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_replayer::snapshot::Replayer<librbd::ImageCtx>;
#endif // RBD_MIRROR_IMAGE_REPLAYER_SNAPSHOT_REPLAYER_H
| 11,210 | 31.031429 | 88 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/snapshot/StateBuilder.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "StateBuilder.h"
#include "include/ceph_assert.h"
#include "include/Context.h"
#include "common/debug.h"
#include "common/errno.h"
#include "librbd/ImageCtx.h"
#include "librbd/mirror/snapshot/ImageMeta.h"
#include "tools/rbd_mirror/image_replayer/snapshot/CreateLocalImageRequest.h"
#include "tools/rbd_mirror/image_replayer/snapshot/PrepareReplayRequest.h"
#include "tools/rbd_mirror/image_replayer/snapshot/Replayer.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_replayer::snapshot::" \
<< "StateBuilder: " << this << " " \
<< __func__ << ": "
namespace rbd {
namespace mirror {
namespace image_replayer {
namespace snapshot {
template <typename I>
StateBuilder<I>::StateBuilder(const std::string& global_image_id)
: image_replayer::StateBuilder<I>(global_image_id) {
}
template <typename I>
StateBuilder<I>::~StateBuilder() {
ceph_assert(local_image_meta == nullptr);
}
template <typename I>
void StateBuilder<I>::close(Context* on_finish) {
dout(10) << dendl;
delete local_image_meta;
local_image_meta = nullptr;
// close the remote image after closing the local
// image in case the remote cluster is unreachable and
// we cannot close it.
on_finish = new LambdaContext([this, on_finish](int) {
this->close_remote_image(on_finish);
});
this->close_local_image(on_finish);
}
template <typename I>
bool StateBuilder<I>::is_disconnected() const {
return false;
}
template <typename I>
bool StateBuilder<I>::is_linked_impl() const {
// the remote has to have us registered as a peer
return !remote_mirror_peer_uuid.empty();
}
template <typename I>
cls::rbd::MirrorImageMode StateBuilder<I>::get_mirror_image_mode() const {
return cls::rbd::MIRROR_IMAGE_MODE_SNAPSHOT;
}
template <typename I>
image_sync::SyncPointHandler* StateBuilder<I>::create_sync_point_handler() {
dout(10) << dendl;
// TODO
ceph_assert(false);
return nullptr;
}
template <typename I>
BaseRequest* StateBuilder<I>::create_local_image_request(
Threads<I>* threads,
librados::IoCtx& local_io_ctx,
const std::string& global_image_id,
PoolMetaCache* pool_meta_cache,
ProgressContext* progress_ctx,
Context* on_finish) {
return CreateLocalImageRequest<I>::create(
threads, local_io_ctx, this->remote_image_ctx, global_image_id,
pool_meta_cache, progress_ctx, this, on_finish);
}
template <typename I>
BaseRequest* StateBuilder<I>::create_prepare_replay_request(
const std::string& local_mirror_uuid,
ProgressContext* progress_ctx,
bool* resync_requested,
bool* syncing,
Context* on_finish) {
return PrepareReplayRequest<I>::create(
local_mirror_uuid, progress_ctx, this, resync_requested, syncing,
on_finish);
}
template <typename I>
image_replayer::Replayer* StateBuilder<I>::create_replayer(
Threads<I>* threads,
InstanceWatcher<I>* instance_watcher,
const std::string& local_mirror_uuid,
PoolMetaCache* pool_meta_cache,
ReplayerListener* replayer_listener) {
return Replayer<I>::create(
threads, instance_watcher, local_mirror_uuid, pool_meta_cache, this,
replayer_listener);
}
} // namespace snapshot
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_replayer::snapshot::StateBuilder<librbd::ImageCtx>;
| 3,553 | 28.371901 | 85 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/snapshot/StateBuilder.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef CEPH_RBD_MIRROR_IMAGE_REPLAYER_SNAPSHOT_STATE_BUILDER_H
#define CEPH_RBD_MIRROR_IMAGE_REPLAYER_SNAPSHOT_STATE_BUILDER_H
#include "tools/rbd_mirror/image_replayer/StateBuilder.h"
#include <string>
struct Context;
namespace librbd {
struct ImageCtx;
namespace mirror {
namespace snapshot {
template <typename> class ImageMeta;
} // namespace snapshot
} // namespace mirror
} // namespace librbd
namespace rbd {
namespace mirror {
namespace image_replayer {
namespace snapshot {
template <typename> class SyncPointHandler;
template <typename ImageCtxT>
class StateBuilder : public image_replayer::StateBuilder<ImageCtxT> {
public:
static StateBuilder* create(const std::string& global_image_id) {
return new StateBuilder(global_image_id);
}
StateBuilder(const std::string& global_image_id);
~StateBuilder() override;
void close(Context* on_finish) override;
bool is_disconnected() const override;
cls::rbd::MirrorImageMode get_mirror_image_mode() const override;
image_sync::SyncPointHandler* create_sync_point_handler() override;
bool replay_requires_remote_image() const override {
return true;
}
BaseRequest* create_local_image_request(
Threads<ImageCtxT>* threads,
librados::IoCtx& local_io_ctx,
const std::string& global_image_id,
PoolMetaCache* pool_meta_cache,
ProgressContext* progress_ctx,
Context* on_finish) override;
BaseRequest* create_prepare_replay_request(
const std::string& local_mirror_uuid,
ProgressContext* progress_ctx,
bool* resync_requested,
bool* syncing,
Context* on_finish) override;
image_replayer::Replayer* create_replayer(
Threads<ImageCtxT>* threads,
InstanceWatcher<ImageCtxT>* instance_watcher,
const std::string& local_mirror_uuid,
PoolMetaCache* pool_meta_cache,
ReplayerListener* replayer_listener) override;
SyncPointHandler<ImageCtxT>* sync_point_handler = nullptr;
std::string remote_mirror_peer_uuid;
librbd::mirror::snapshot::ImageMeta<ImageCtxT>* local_image_meta = nullptr;
private:
bool is_linked_impl() const override;
};
} // namespace snapshot
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_replayer::snapshot::StateBuilder<librbd::ImageCtx>;
#endif // CEPH_RBD_MIRROR_IMAGE_REPLAYER_SNAPSHOT_STATE_BUILDER_H
| 2,490 | 25.5 | 92 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/snapshot/Utils.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "Utils.h"
#include "common/debug.h"
#include "common/errno.h"
#include "cls/rbd/cls_rbd_types.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_replayer::snapshot::util::" \
<< __func__ << ": "
namespace rbd {
namespace mirror {
namespace image_replayer {
namespace snapshot {
namespace util {
uint64_t compute_remote_snap_id(
const ceph::shared_mutex& local_image_lock,
const std::map<librados::snap_t, librbd::SnapInfo>& local_snap_infos,
uint64_t local_snap_id, const std::string& remote_mirror_uuid) {
ceph_assert(ceph_mutex_is_locked(local_image_lock));
// Search our local non-primary snapshots for a mapping to the remote
// snapshot. The non-primary mirror snapshot with the mappings will always
// come at or after the snapshot we are searching against
for (auto snap_it = local_snap_infos.lower_bound(local_snap_id);
snap_it != local_snap_infos.end(); ++snap_it) {
auto mirror_ns = std::get_if<cls::rbd::MirrorSnapshotNamespace>(
&snap_it->second.snap_namespace);
if (mirror_ns == nullptr || !mirror_ns->is_non_primary()) {
continue;
}
if (mirror_ns->primary_mirror_uuid != remote_mirror_uuid) {
dout(20) << "local snapshot " << snap_it->first << " not tied to remote"
<< dendl;
continue;
} else if (local_snap_id == snap_it->first) {
dout(15) << "local snapshot " << local_snap_id << " maps to "
<< "remote snapshot " << mirror_ns->primary_snap_id << dendl;
return mirror_ns->primary_snap_id;
}
const auto& snap_seqs = mirror_ns->snap_seqs;
for (auto [remote_snap_id_seq, local_snap_id_seq] : snap_seqs) {
if (local_snap_id_seq == local_snap_id) {
dout(15) << "local snapshot " << local_snap_id << " maps to "
<< "remote snapshot " << remote_snap_id_seq << dendl;
return remote_snap_id_seq;
}
}
}
return CEPH_NOSNAP;
}
} // namespace util
} // namespace snapshot
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
| 2,269 | 33.393939 | 79 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_replayer/snapshot/Utils.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_REPLAYER_SNAPSHOT_UTILS_H
#define RBD_MIRROR_IMAGE_REPLAYER_SNAPSHOT_UTILS_H
#include "include/int_types.h"
#include "include/rados/librados.hpp"
#include "common/ceph_mutex.h"
#include "librbd/Types.h"
#include <map>
namespace rbd {
namespace mirror {
namespace image_replayer {
namespace snapshot {
namespace util {
uint64_t compute_remote_snap_id(
const ceph::shared_mutex& local_image_lock,
const std::map<librados::snap_t, librbd::SnapInfo>& local_snap_infos,
uint64_t local_snap_id, const std::string& remote_mirror_uuid);
} // namespace util
} // namespace snapshot
} // namespace image_replayer
} // namespace mirror
} // namespace rbd
#endif // RBD_MIRROR_IMAGE_REPLAYER_SNAPSHOT_UTILS_H
| 838 | 26.064516 | 73 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_sync/SyncPointCreateRequest.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "SyncPointCreateRequest.h"
#include "include/uuid.h"
#include "common/debug.h"
#include "common/errno.h"
#include "librbd/ImageCtx.h"
#include "librbd/ImageState.h"
#include "librbd/Operations.h"
#include "librbd/Utils.h"
#include "tools/rbd_mirror/image_sync/Types.h"
#include "tools/rbd_mirror/image_sync/Utils.h"
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_sync::SyncPointCreateRequest: " \
<< this << " " << __func__
namespace rbd {
namespace mirror {
namespace image_sync {
using librbd::util::create_context_callback;
template <typename I>
SyncPointCreateRequest<I>::SyncPointCreateRequest(
I *remote_image_ctx,
const std::string &local_mirror_uuid,
SyncPointHandler* sync_point_handler,
Context *on_finish)
: m_remote_image_ctx(remote_image_ctx),
m_local_mirror_uuid(local_mirror_uuid),
m_sync_point_handler(sync_point_handler),
m_on_finish(on_finish) {
m_sync_points_copy = m_sync_point_handler->get_sync_points();
ceph_assert(m_sync_points_copy.size() < 2);
// initialize the updated client meta with the new sync point
m_sync_points_copy.emplace_back();
if (m_sync_points_copy.size() > 1) {
m_sync_points_copy.back().from_snap_name =
m_sync_points_copy.front().snap_name;
}
}
template <typename I>
void SyncPointCreateRequest<I>::send() {
send_update_sync_points();
}
template <typename I>
void SyncPointCreateRequest<I>::send_update_sync_points() {
uuid_d uuid_gen;
uuid_gen.generate_random();
auto& sync_point = m_sync_points_copy.back();
sync_point.snap_name = util::get_snapshot_name_prefix(m_local_mirror_uuid) +
uuid_gen.to_string();
auto ctx = create_context_callback<
SyncPointCreateRequest<I>,
&SyncPointCreateRequest<I>::handle_update_sync_points>(this);
m_sync_point_handler->update_sync_points(
m_sync_point_handler->get_snap_seqs(), m_sync_points_copy, false, ctx);
}
template <typename I>
void SyncPointCreateRequest<I>::handle_update_sync_points(int r) {
dout(20) << ": r=" << r << dendl;
if (r < 0) {
derr << ": failed to update client data: " << cpp_strerror(r)
<< dendl;
finish(r);
return;
}
send_refresh_image();
}
template <typename I>
void SyncPointCreateRequest<I>::send_refresh_image() {
dout(20) << dendl;
Context *ctx = create_context_callback<
SyncPointCreateRequest<I>, &SyncPointCreateRequest<I>::handle_refresh_image>(
this);
m_remote_image_ctx->state->refresh(ctx);
}
template <typename I>
void SyncPointCreateRequest<I>::handle_refresh_image(int r) {
dout(20) << ": r=" << r << dendl;
if (r < 0) {
derr << ": remote image refresh failed: " << cpp_strerror(r) << dendl;
finish(r);
return;
}
send_create_snap();
}
template <typename I>
void SyncPointCreateRequest<I>::send_create_snap() {
dout(20) << dendl;
auto& sync_point = m_sync_points_copy.back();
Context *ctx = create_context_callback<
SyncPointCreateRequest<I>, &SyncPointCreateRequest<I>::handle_create_snap>(
this);
m_remote_image_ctx->operations->snap_create(
cls::rbd::UserSnapshotNamespace(), sync_point.snap_name.c_str(),
librbd::SNAP_CREATE_FLAG_SKIP_NOTIFY_QUIESCE, m_prog_ctx, ctx);
}
template <typename I>
void SyncPointCreateRequest<I>::handle_create_snap(int r) {
dout(20) << ": r=" << r << dendl;
if (r == -EEXIST) {
send_update_sync_points();
return;
} else if (r < 0) {
derr << ": failed to create snapshot: " << cpp_strerror(r) << dendl;
finish(r);
return;
}
send_final_refresh_image();
}
template <typename I>
void SyncPointCreateRequest<I>::send_final_refresh_image() {
dout(20) << dendl;
Context *ctx = create_context_callback<
SyncPointCreateRequest<I>,
&SyncPointCreateRequest<I>::handle_final_refresh_image>(this);
m_remote_image_ctx->state->refresh(ctx);
}
template <typename I>
void SyncPointCreateRequest<I>::handle_final_refresh_image(int r) {
dout(20) << ": r=" << r << dendl;
if (r < 0) {
derr << ": failed to refresh image for snapshot: " << cpp_strerror(r)
<< dendl;
finish(r);
return;
}
finish(0);
}
template <typename I>
void SyncPointCreateRequest<I>::finish(int r) {
dout(20) << ": r=" << r << dendl;
m_on_finish->complete(r);
delete this;
}
} // namespace image_sync
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_sync::SyncPointCreateRequest<librbd::ImageCtx>;
| 4,667 | 25.982659 | 83 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_sync/SyncPointCreateRequest.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_SYNC_SYNC_POINT_CREATE_REQUEST_H
#define RBD_MIRROR_IMAGE_SYNC_SYNC_POINT_CREATE_REQUEST_H
#include "librbd/internal.h"
#include "Types.h"
#include <string>
class Context;
namespace journal { class Journaler; }
namespace librbd { class ImageCtx; }
namespace librbd { namespace journal { struct MirrorPeerClientMeta; } }
namespace rbd {
namespace mirror {
namespace image_sync {
template <typename ImageCtxT = librbd::ImageCtx>
class SyncPointCreateRequest {
public:
static SyncPointCreateRequest* create(
ImageCtxT *remote_image_ctx,
const std::string &local_mirror_uuid,
SyncPointHandler* sync_point_handler,
Context *on_finish) {
return new SyncPointCreateRequest(remote_image_ctx, local_mirror_uuid,
sync_point_handler, on_finish);
}
SyncPointCreateRequest(
ImageCtxT *remote_image_ctx,
const std::string &local_mirror_uuid,
SyncPointHandler* sync_point_handler,
Context *on_finish);
void send();
private:
/**
* @verbatim
*
* <start>
* |
* v
* UPDATE_SYNC_POINTS < . .
* | .
* v .
* REFRESH_IMAGE .
* | . (repeat on EEXIST)
* v .
* CREATE_SNAP . . . . . .
* |
* v
* REFRESH_IMAGE
* |
* v
* <finish>
*
* @endverbatim
*/
ImageCtxT *m_remote_image_ctx;
std::string m_local_mirror_uuid;
SyncPointHandler* m_sync_point_handler;
Context *m_on_finish;
SyncPoints m_sync_points_copy;
librbd::NoOpProgressContext m_prog_ctx;
void send_update_sync_points();
void handle_update_sync_points(int r);
void send_refresh_image();
void handle_refresh_image(int r);
void send_create_snap();
void handle_create_snap(int r);
void send_final_refresh_image();
void handle_final_refresh_image(int r);
void finish(int r);
};
} // namespace image_sync
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_sync::SyncPointCreateRequest<librbd::ImageCtx>;
#endif // RBD_MIRROR_IMAGE_SYNC_SYNC_POINT_CREATE_REQUEST_H
| 2,266 | 23.117021 | 88 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_sync/SyncPointPruneRequest.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "SyncPointPruneRequest.h"
#include "common/debug.h"
#include "common/errno.h"
#include "librbd/ImageCtx.h"
#include "librbd/ImageState.h"
#include "librbd/Operations.h"
#include "librbd/Utils.h"
#include <set>
#define dout_context g_ceph_context
#define dout_subsys ceph_subsys_rbd_mirror
#undef dout_prefix
#define dout_prefix *_dout << "rbd::mirror::image_sync::SyncPointPruneRequest: " \
<< this << " " << __func__
namespace rbd {
namespace mirror {
namespace image_sync {
using librbd::util::create_context_callback;
template <typename I>
SyncPointPruneRequest<I>::SyncPointPruneRequest(
I *remote_image_ctx,
bool sync_complete,
SyncPointHandler* sync_point_handler,
Context *on_finish)
: m_remote_image_ctx(remote_image_ctx),
m_sync_complete(sync_complete),
m_sync_point_handler(sync_point_handler),
m_on_finish(on_finish) {
m_sync_points_copy = m_sync_point_handler->get_sync_points();
}
template <typename I>
void SyncPointPruneRequest<I>::send() {
if (m_sync_points_copy.empty()) {
send_remove_snap();
return;
}
if (m_sync_complete) {
// if sync is complete, we can remove the master sync point
auto it = m_sync_points_copy.begin();
auto& sync_point = *it;
++it;
if (it == m_sync_points_copy.end() ||
it->from_snap_name != sync_point.snap_name) {
m_snap_names.push_back(sync_point.snap_name);
}
if (!sync_point.from_snap_name.empty()) {
m_snap_names.push_back(sync_point.from_snap_name);
}
} else {
// if we have more than one sync point or invalid sync points,
// trim them off
std::shared_lock image_locker{m_remote_image_ctx->image_lock};
std::set<std::string> snap_names;
for (auto it = m_sync_points_copy.rbegin();
it != m_sync_points_copy.rend(); ++it) {
auto& sync_point = *it;
if (&sync_point == &m_sync_points_copy.front()) {
if (m_remote_image_ctx->get_snap_id(
cls::rbd::UserSnapshotNamespace(), sync_point.snap_name) ==
CEPH_NOSNAP) {
derr << ": failed to locate sync point snapshot: "
<< sync_point.snap_name << dendl;
} else if (!sync_point.from_snap_name.empty()) {
derr << ": unexpected from_snap_name in primary sync point: "
<< sync_point.from_snap_name << dendl;
} else {
// first sync point is OK -- keep it
break;
}
m_invalid_master_sync_point = true;
}
if (snap_names.count(sync_point.snap_name) == 0) {
snap_names.insert(sync_point.snap_name);
m_snap_names.push_back(sync_point.snap_name);
}
auto& front_sync_point = m_sync_points_copy.front();
if (!sync_point.from_snap_name.empty() &&
snap_names.count(sync_point.from_snap_name) == 0 &&
sync_point.from_snap_name != front_sync_point.snap_name) {
snap_names.insert(sync_point.from_snap_name);
m_snap_names.push_back(sync_point.from_snap_name);
}
}
}
send_remove_snap();
}
template <typename I>
void SyncPointPruneRequest<I>::send_remove_snap() {
if (m_snap_names.empty()) {
send_refresh_image();
return;
}
const std::string &snap_name = m_snap_names.front();
dout(20) << ": snap_name=" << snap_name << dendl;
Context *ctx = create_context_callback<
SyncPointPruneRequest<I>, &SyncPointPruneRequest<I>::handle_remove_snap>(
this);
m_remote_image_ctx->operations->snap_remove(cls::rbd::UserSnapshotNamespace(),
snap_name.c_str(),
ctx);
}
template <typename I>
void SyncPointPruneRequest<I>::handle_remove_snap(int r) {
dout(20) << ": r=" << r << dendl;
ceph_assert(!m_snap_names.empty());
std::string snap_name = m_snap_names.front();
m_snap_names.pop_front();
if (r == -ENOENT) {
r = 0;
}
if (r < 0) {
derr << ": failed to remove snapshot '" << snap_name << "': "
<< cpp_strerror(r) << dendl;
finish(r);
return;
}
send_remove_snap();
}
template <typename I>
void SyncPointPruneRequest<I>::send_refresh_image() {
dout(20) << dendl;
Context *ctx = create_context_callback<
SyncPointPruneRequest<I>, &SyncPointPruneRequest<I>::handle_refresh_image>(
this);
m_remote_image_ctx->state->refresh(ctx);
}
template <typename I>
void SyncPointPruneRequest<I>::handle_refresh_image(int r) {
dout(20) << ": r=" << r << dendl;
if (r < 0) {
derr << ": remote image refresh failed: " << cpp_strerror(r) << dendl;
finish(r);
return;
}
send_update_sync_points();
}
template <typename I>
void SyncPointPruneRequest<I>::send_update_sync_points() {
dout(20) << dendl;
if (m_sync_complete) {
m_sync_points_copy.pop_front();
} else {
while (m_sync_points_copy.size() > 1) {
m_sync_points_copy.pop_back();
}
if (m_invalid_master_sync_point) {
// all subsequent sync points would have been pruned
m_sync_points_copy.clear();
}
}
auto ctx = create_context_callback<
SyncPointPruneRequest<I>,
&SyncPointPruneRequest<I>::handle_update_sync_points>(this);
m_sync_point_handler->update_sync_points(
m_sync_point_handler->get_snap_seqs(), m_sync_points_copy,
m_sync_complete, ctx);
}
template <typename I>
void SyncPointPruneRequest<I>::handle_update_sync_points(int r) {
dout(20) << ": r=" << r << dendl;
if (r < 0) {
derr << ": failed to update client data: " << cpp_strerror(r)
<< dendl;
finish(r);
return;
}
finish(0);
}
template <typename I>
void SyncPointPruneRequest<I>::finish(int r) {
dout(20) << ": r=" << r << dendl;
m_on_finish->complete(r);
delete this;
}
} // namespace image_sync
} // namespace mirror
} // namespace rbd
template class rbd::mirror::image_sync::SyncPointPruneRequest<librbd::ImageCtx>;
| 5,936 | 26.742991 | 82 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_sync/SyncPointPruneRequest.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_SYNC_SYNC_POINT_PRUNE_REQUEST_H
#define RBD_MIRROR_IMAGE_SYNC_SYNC_POINT_PRUNE_REQUEST_H
#include "tools/rbd_mirror/image_sync/Types.h"
#include <list>
#include <string>
class Context;
namespace journal { class Journaler; }
namespace librbd { class ImageCtx; }
namespace librbd { namespace journal { struct MirrorPeerClientMeta; } }
namespace rbd {
namespace mirror {
namespace image_sync {
template <typename ImageCtxT = librbd::ImageCtx>
class SyncPointPruneRequest {
public:
static SyncPointPruneRequest* create(
ImageCtxT *remote_image_ctx,
bool sync_complete,
SyncPointHandler* sync_point_handler,
Context *on_finish) {
return new SyncPointPruneRequest(remote_image_ctx, sync_complete,
sync_point_handler, on_finish);
}
SyncPointPruneRequest(
ImageCtxT *remote_image_ctx,
bool sync_complete,
SyncPointHandler* sync_point_handler,
Context *on_finish);
void send();
private:
/**
* @verbatim
*
* <start>
* |
* | . . . . .
* | . .
* v v . (repeat if from snap
* REMOVE_SNAP . . . unused by other sync)
* |
* v
* REFRESH_IMAGE
* |
* v
* UPDATE_CLIENT
* |
* v
* <finish>
*
* @endverbatim
*/
ImageCtxT *m_remote_image_ctx;
bool m_sync_complete;
SyncPointHandler* m_sync_point_handler;
Context *m_on_finish;
SyncPoints m_sync_points_copy;
std::list<std::string> m_snap_names;
bool m_invalid_master_sync_point = false;
void send_remove_snap();
void handle_remove_snap(int r);
void send_refresh_image();
void handle_refresh_image(int r);
void send_update_sync_points();
void handle_update_sync_points(int r);
void finish(int r);
};
} // namespace image_sync
} // namespace mirror
} // namespace rbd
extern template class rbd::mirror::image_sync::SyncPointPruneRequest<librbd::ImageCtx>;
#endif // RBD_MIRROR_IMAGE_SYNC_SYNC_POINT_PRUNE_REQUEST_H
| 2,121 | 22.065217 | 87 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_sync/Types.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_IMAGE_SYNC_TYPES_H
#define RBD_MIRROR_IMAGE_SYNC_TYPES_H
#include "cls/rbd/cls_rbd_types.h"
#include "librbd/Types.h"
#include <list>
#include <string>
#include <boost/optional.hpp>
struct Context;
namespace rbd {
namespace mirror {
namespace image_sync {
struct SyncPoint {
typedef boost::optional<uint64_t> ObjectNumber;
SyncPoint() {
}
SyncPoint(const cls::rbd::SnapshotNamespace& snap_namespace,
const std::string& snap_name,
const std::string& from_snap_name,
const ObjectNumber& object_number)
: snap_namespace(snap_namespace), snap_name(snap_name),
from_snap_name(from_snap_name), object_number(object_number) {
}
cls::rbd::SnapshotNamespace snap_namespace =
{cls::rbd::UserSnapshotNamespace{}};
std::string snap_name;
std::string from_snap_name;
ObjectNumber object_number = boost::none;
bool operator==(const SyncPoint& rhs) const {
return (snap_namespace == rhs.snap_namespace &&
snap_name == rhs.snap_name &&
from_snap_name == rhs.from_snap_name &&
object_number == rhs.object_number);
}
};
typedef std::list<SyncPoint> SyncPoints;
struct SyncPointHandler {
public:
SyncPointHandler(const SyncPointHandler&) = delete;
SyncPointHandler& operator=(const SyncPointHandler&) = delete;
virtual ~SyncPointHandler() {}
virtual void destroy() {
delete this;
}
virtual SyncPoints get_sync_points() const = 0;
virtual librbd::SnapSeqs get_snap_seqs() const = 0;
virtual void update_sync_points(const librbd::SnapSeqs& snap_seq,
const SyncPoints& sync_points,
bool sync_complete,
Context* on_finish) = 0;
protected:
SyncPointHandler() {}
};
} // namespace image_sync
} // namespace mirror
} // namespace rbd
#endif // RBD_MIRROR_IMAGE_SYNC_TYPES_H
| 2,017 | 25.906667 | 70 |
h
|
null |
ceph-main/src/tools/rbd_mirror/image_sync/Utils.cc
|
// -*- mode:c++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "Utils.h"
namespace rbd {
namespace mirror {
namespace image_sync {
namespace util {
namespace {
static const std::string SNAP_NAME_PREFIX(".rbd-mirror");
} // anonymous namespace
std::string get_snapshot_name_prefix(const std::string& local_mirror_uuid) {
return SNAP_NAME_PREFIX + "." + local_mirror_uuid + ".";
}
} // namespace util
} // namespace image_sync
} // namespace mirror
} // namespace rbd
| 519 | 19.8 | 76 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/image_sync/Utils.h
|
// -*- mode:c++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include <string>
namespace rbd {
namespace mirror {
namespace image_sync {
namespace util {
std::string get_snapshot_name_prefix(const std::string& local_mirror_uuid);
} // namespace util
} // namespace image_sync
} // namespace mirror
} // namespace rbd
| 358 | 20.117647 | 75 |
h
|
null |
ceph-main/src/tools/rbd_mirror/instance_watcher/Types.cc
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#include "Types.h"
#include "include/ceph_assert.h"
#include "include/stringify.h"
#include "common/Formatter.h"
namespace rbd {
namespace mirror {
namespace instance_watcher {
namespace {
class EncodePayloadVisitor : public boost::static_visitor<void> {
public:
explicit EncodePayloadVisitor(bufferlist &bl) : m_bl(bl) {}
template <typename Payload>
inline void operator()(const Payload &payload) const {
using ceph::encode;
encode(static_cast<uint32_t>(Payload::NOTIFY_OP), m_bl);
payload.encode(m_bl);
}
private:
bufferlist &m_bl;
};
class DecodePayloadVisitor : public boost::static_visitor<void> {
public:
DecodePayloadVisitor(__u8 version, bufferlist::const_iterator &iter)
: m_version(version), m_iter(iter) {}
template <typename Payload>
inline void operator()(Payload &payload) const {
payload.decode(m_version, m_iter);
}
private:
__u8 m_version;
bufferlist::const_iterator &m_iter;
};
class DumpPayloadVisitor : public boost::static_visitor<void> {
public:
explicit DumpPayloadVisitor(Formatter *formatter) : m_formatter(formatter) {}
template <typename Payload>
inline void operator()(const Payload &payload) const {
NotifyOp notify_op = Payload::NOTIFY_OP;
m_formatter->dump_string("notify_op", stringify(notify_op));
payload.dump(m_formatter);
}
private:
ceph::Formatter *m_formatter;
};
} // anonymous namespace
void PayloadBase::encode(bufferlist &bl) const {
using ceph::encode;
encode(request_id, bl);
}
void PayloadBase::decode(__u8 version, bufferlist::const_iterator &iter) {
using ceph::decode;
decode(request_id, iter);
}
void PayloadBase::dump(Formatter *f) const {
f->dump_unsigned("request_id", request_id);
}
void ImagePayloadBase::encode(bufferlist &bl) const {
using ceph::encode;
PayloadBase::encode(bl);
encode(global_image_id, bl);
}
void ImagePayloadBase::decode(__u8 version, bufferlist::const_iterator &iter) {
using ceph::decode;
PayloadBase::decode(version, iter);
decode(global_image_id, iter);
}
void ImagePayloadBase::dump(Formatter *f) const {
PayloadBase::dump(f);
f->dump_string("global_image_id", global_image_id);
}
void PeerImageRemovedPayload::encode(bufferlist &bl) const {
using ceph::encode;
PayloadBase::encode(bl);
encode(global_image_id, bl);
encode(peer_mirror_uuid, bl);
}
void PeerImageRemovedPayload::decode(__u8 version, bufferlist::const_iterator &iter) {
using ceph::decode;
PayloadBase::decode(version, iter);
decode(global_image_id, iter);
decode(peer_mirror_uuid, iter);
}
void PeerImageRemovedPayload::dump(Formatter *f) const {
PayloadBase::dump(f);
f->dump_string("global_image_id", global_image_id);
f->dump_string("peer_mirror_uuid", peer_mirror_uuid);
}
void SyncPayloadBase::encode(bufferlist &bl) const {
using ceph::encode;
PayloadBase::encode(bl);
encode(sync_id, bl);
}
void SyncPayloadBase::decode(__u8 version, bufferlist::const_iterator &iter) {
using ceph::decode;
PayloadBase::decode(version, iter);
decode(sync_id, iter);
}
void SyncPayloadBase::dump(Formatter *f) const {
PayloadBase::dump(f);
f->dump_string("sync_id", sync_id);
}
void UnknownPayload::encode(bufferlist &bl) const {
ceph_abort();
}
void UnknownPayload::decode(__u8 version, bufferlist::const_iterator &iter) {
}
void UnknownPayload::dump(Formatter *f) const {
}
void NotifyMessage::encode(bufferlist& bl) const {
ENCODE_START(2, 2, bl);
boost::apply_visitor(EncodePayloadVisitor(bl), payload);
ENCODE_FINISH(bl);
}
void NotifyMessage::decode(bufferlist::const_iterator& iter) {
DECODE_START(2, iter);
uint32_t notify_op;
decode(notify_op, iter);
// select the correct payload variant based upon the encoded op
switch (notify_op) {
case NOTIFY_OP_IMAGE_ACQUIRE:
payload = ImageAcquirePayload();
break;
case NOTIFY_OP_IMAGE_RELEASE:
payload = ImageReleasePayload();
break;
case NOTIFY_OP_PEER_IMAGE_REMOVED:
payload = PeerImageRemovedPayload();
break;
case NOTIFY_OP_SYNC_REQUEST:
payload = SyncRequestPayload();
break;
case NOTIFY_OP_SYNC_START:
payload = SyncStartPayload();
break;
default:
payload = UnknownPayload();
break;
}
apply_visitor(DecodePayloadVisitor(struct_v, iter), payload);
DECODE_FINISH(iter);
}
void NotifyMessage::dump(Formatter *f) const {
apply_visitor(DumpPayloadVisitor(f), payload);
}
void NotifyMessage::generate_test_instances(std::list<NotifyMessage *> &o) {
o.push_back(new NotifyMessage(ImageAcquirePayload()));
o.push_back(new NotifyMessage(ImageAcquirePayload(1, "gid")));
o.push_back(new NotifyMessage(ImageReleasePayload()));
o.push_back(new NotifyMessage(ImageReleasePayload(1, "gid")));
o.push_back(new NotifyMessage(PeerImageRemovedPayload()));
o.push_back(new NotifyMessage(PeerImageRemovedPayload(1, "gid", "uuid")));
o.push_back(new NotifyMessage(SyncRequestPayload()));
o.push_back(new NotifyMessage(SyncRequestPayload(1, "sync_id")));
o.push_back(new NotifyMessage(SyncStartPayload()));
o.push_back(new NotifyMessage(SyncStartPayload(1, "sync_id")));
}
std::ostream &operator<<(std::ostream &out, const NotifyOp &op) {
switch (op) {
case NOTIFY_OP_IMAGE_ACQUIRE:
out << "ImageAcquire";
break;
case NOTIFY_OP_IMAGE_RELEASE:
out << "ImageRelease";
break;
case NOTIFY_OP_PEER_IMAGE_REMOVED:
out << "PeerImageRemoved";
break;
case NOTIFY_OP_SYNC_REQUEST:
out << "SyncRequest";
break;
case NOTIFY_OP_SYNC_START:
out << "SyncStart";
break;
default:
out << "Unknown (" << static_cast<uint32_t>(op) << ")";
break;
}
return out;
}
void NotifyAckPayload::encode(bufferlist &bl) const {
using ceph::encode;
encode(instance_id, bl);
encode(request_id, bl);
encode(ret_val, bl);
}
void NotifyAckPayload::decode(bufferlist::const_iterator &iter) {
using ceph::decode;
decode(instance_id, iter);
decode(request_id, iter);
decode(ret_val, iter);
}
void NotifyAckPayload::dump(Formatter *f) const {
f->dump_string("instance_id", instance_id);
f->dump_unsigned("request_id", request_id);
f->dump_int("request_id", ret_val);
}
} // namespace instance_watcher
} // namespace mirror
} // namespace rbd
| 6,335 | 24.756098 | 86 |
cc
|
null |
ceph-main/src/tools/rbd_mirror/instance_watcher/Types.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef RBD_MIRROR_INSTANCE_WATCHER_TYPES_H
#define RBD_MIRROR_INSTANCE_WATCHER_TYPES_H
#include <string>
#include <set>
#include <boost/variant.hpp>
#include "include/buffer_fwd.h"
#include "include/encoding.h"
#include "include/int_types.h"
namespace ceph { class Formatter; }
namespace rbd {
namespace mirror {
namespace instance_watcher {
enum NotifyOp {
NOTIFY_OP_IMAGE_ACQUIRE = 0,
NOTIFY_OP_IMAGE_RELEASE = 1,
NOTIFY_OP_PEER_IMAGE_REMOVED = 2,
NOTIFY_OP_SYNC_REQUEST = 3,
NOTIFY_OP_SYNC_START = 4
};
struct PayloadBase {
uint64_t request_id;
PayloadBase() : request_id(0) {
}
PayloadBase(uint64_t request_id) : request_id(request_id) {
}
void encode(bufferlist &bl) const;
void decode(__u8 version, bufferlist::const_iterator &iter);
void dump(Formatter *f) const;
};
struct ImagePayloadBase : public PayloadBase {
std::string global_image_id;
ImagePayloadBase() : PayloadBase() {
}
ImagePayloadBase(uint64_t request_id, const std::string &global_image_id)
: PayloadBase(request_id), global_image_id(global_image_id) {
}
void encode(bufferlist &bl) const;
void decode(__u8 version, bufferlist::const_iterator &iter);
void dump(Formatter *f) const;
};
struct ImageAcquirePayload : public ImagePayloadBase {
static const NotifyOp NOTIFY_OP = NOTIFY_OP_IMAGE_ACQUIRE;
ImageAcquirePayload() {
}
ImageAcquirePayload(uint64_t request_id, const std::string &global_image_id)
: ImagePayloadBase(request_id, global_image_id) {
}
};
struct ImageReleasePayload : public ImagePayloadBase {
static const NotifyOp NOTIFY_OP = NOTIFY_OP_IMAGE_RELEASE;
ImageReleasePayload() {
}
ImageReleasePayload(uint64_t request_id, const std::string &global_image_id)
: ImagePayloadBase(request_id, global_image_id) {
}
};
struct PeerImageRemovedPayload : public PayloadBase {
static const NotifyOp NOTIFY_OP = NOTIFY_OP_PEER_IMAGE_REMOVED;
std::string global_image_id;
std::string peer_mirror_uuid;
PeerImageRemovedPayload() {
}
PeerImageRemovedPayload(uint64_t request_id,
const std::string& global_image_id,
const std::string& peer_mirror_uuid)
: PayloadBase(request_id),
global_image_id(global_image_id), peer_mirror_uuid(peer_mirror_uuid) {
}
void encode(bufferlist &bl) const;
void decode(__u8 version, bufferlist::const_iterator &iter);
void dump(Formatter *f) const;
};
struct SyncPayloadBase : public PayloadBase {
std::string sync_id;
SyncPayloadBase() : PayloadBase() {
}
SyncPayloadBase(uint64_t request_id, const std::string &sync_id)
: PayloadBase(request_id), sync_id(sync_id) {
}
void encode(bufferlist &bl) const;
void decode(__u8 version, bufferlist::const_iterator &iter);
void dump(Formatter *f) const;
};
struct SyncRequestPayload : public SyncPayloadBase {
static const NotifyOp NOTIFY_OP = NOTIFY_OP_SYNC_REQUEST;
SyncRequestPayload() : SyncPayloadBase() {
}
SyncRequestPayload(uint64_t request_id, const std::string &sync_id)
: SyncPayloadBase(request_id, sync_id) {
}
};
struct SyncStartPayload : public SyncPayloadBase {
static const NotifyOp NOTIFY_OP = NOTIFY_OP_SYNC_START;
SyncStartPayload() : SyncPayloadBase() {
}
SyncStartPayload(uint64_t request_id, const std::string &sync_id)
: SyncPayloadBase(request_id, sync_id) {
}
};
struct UnknownPayload {
static const NotifyOp NOTIFY_OP = static_cast<NotifyOp>(-1);
UnknownPayload() {
}
void encode(bufferlist &bl) const;
void decode(__u8 version, bufferlist::const_iterator &iter);
void dump(Formatter *f) const;
};
typedef boost::variant<ImageAcquirePayload,
ImageReleasePayload,
PeerImageRemovedPayload,
SyncRequestPayload,
SyncStartPayload,
UnknownPayload> Payload;
struct NotifyMessage {
NotifyMessage(const Payload &payload = UnknownPayload()) : payload(payload) {
}
Payload payload;
void encode(bufferlist& bl) const;
void decode(bufferlist::const_iterator& it);
void dump(Formatter *f) const;
static void generate_test_instances(std::list<NotifyMessage *> &o);
};
WRITE_CLASS_ENCODER(NotifyMessage);
std::ostream &operator<<(std::ostream &out, const NotifyOp &op);
struct NotifyAckPayload {
std::string instance_id;
uint64_t request_id;
int ret_val;
NotifyAckPayload() : request_id(0), ret_val(0) {
}
NotifyAckPayload(const std::string &instance_id, uint64_t request_id,
int ret_val)
: instance_id(instance_id), request_id(request_id), ret_val(ret_val) {
}
void encode(bufferlist &bl) const;
void decode(bufferlist::const_iterator& it);
void dump(Formatter *f) const;
};
WRITE_CLASS_ENCODER(NotifyAckPayload);
} // namespace instance_watcher
} // namespace mirror
} // namespace librbd
using rbd::mirror::instance_watcher::encode;
using rbd::mirror::instance_watcher::decode;
#endif // RBD_MIRROR_INSTANCE_WATCHER_TYPES_H
| 5,146 | 24.994949 | 79 |
h
|
null |
ceph-main/src/tools/rbd_mirror/instances/Types.h
|
// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab
#ifndef CEPH_RBD_MIRROR_INSTANCES_TYPES_H
#define CEPH_RBD_MIRROR_INSTANCES_TYPES_H
#include <string>
#include <vector>
namespace rbd {
namespace mirror {
namespace instances {
struct Listener {
typedef std::vector<std::string> InstanceIds;
virtual ~Listener() {
}
virtual void handle_added(const InstanceIds& instance_ids) = 0;
virtual void handle_removed(const InstanceIds& instance_ids) = 0;
};
} // namespace instances
} // namespace mirror
} // namespace rbd
#endif // CEPH_RBD_MIRROR_INSTANCES_TYPES_H
| 624 | 20.551724 | 70 |
h
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.