diff options
author | Henning Baldersheim <balder@yahoo-inc.com> | 2022-03-31 13:00:32 +0000 |
---|---|---|
committer | Henning Baldersheim <balder@yahoo-inc.com> | 2022-03-31 13:00:32 +0000 |
commit | 94f53ab6a93aefa007200b97846ce47ea166bd70 (patch) | |
tree | 0421134fe3a839da8b36c18636bd901ae9939f60 /searchcore | |
parent | d0fa48f2559c297e7b9c15137f3c04bc32fc65fa (diff) |
GC unused Context parameter
Diffstat (limited to 'searchcore')
5 files changed, 46 insertions, 68 deletions
diff --git a/searchcore/src/tests/proton/persistenceengine/persistenceengine_test.cpp b/searchcore/src/tests/proton/persistenceengine/persistenceengine_test.cpp index 6fb7a35fffe..cf58056cae8 100644 --- a/searchcore/src/tests/proton/persistenceengine/persistenceengine_test.cpp +++ b/searchcore/src/tests/proton/persistenceengine/persistenceengine_test.cpp @@ -446,17 +446,16 @@ TEST_F("require that getPartitionStates() prepares all handlers", SimpleFixture) TEST_F("require that puts are routed to handler", SimpleFixture) { - Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); - f.engine.put(bucket1, tstamp1, doc1, context); + f.engine.put(bucket1, tstamp1, doc1); TEST_DO(assertHandler(bucket1, tstamp1, docId1, f.hset.handler1)); TEST_DO(assertHandler(bucket0, tstamp0, docId0, f.hset.handler2)); - f.engine.put(bucket1, tstamp1, doc2, context); + f.engine.put(bucket1, tstamp1, doc2); TEST_DO(assertHandler(bucket1, tstamp1, docId1, f.hset.handler1)); TEST_DO(assertHandler(bucket1, tstamp1, docId2, f.hset.handler2)); EXPECT_EQUAL(Result(Result::ErrorType::PERMANENT_ERROR, "No handler for document type 'type3'"), - f.engine.put(bucket1, tstamp1, doc3, context)); + f.engine.put(bucket1, tstamp1, doc3)); } @@ -465,39 +464,35 @@ TEST_F("require that put is rejected if resource limit is reached", SimpleFixtur f._writeFilter._acceptWriteOperation = false; f._writeFilter._message = "Disk is full"; - Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); EXPECT_EQUAL( Result(Result::ErrorType::RESOURCE_EXHAUSTED, "Put operation rejected for document 'id:type3:type3::1': 'Disk is full'"), - f.engine.put(bucket1, tstamp1, doc3, context)); + f.engine.put(bucket1, tstamp1, doc3)); } TEST_F("require that updates are routed to handler", SimpleFixture) { - Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); f.hset.handler1.setExistingTimestamp(tstamp2); - UpdateResult ur = f.engine.update(bucket1, tstamp1, upd1, context); + UpdateResult ur = f.engine.update(bucket1, tstamp1, upd1); TEST_DO(assertHandler(bucket1, tstamp1, docId1, f.hset.handler1)); TEST_DO(assertHandler(bucket0, tstamp0, docId0, f.hset.handler2)); EXPECT_EQUAL(tstamp2, ur.getExistingTimestamp()); f.hset.handler2.setExistingTimestamp(tstamp3); - ur = f.engine.update(bucket1, tstamp1, upd2, context); + ur = f.engine.update(bucket1, tstamp1, upd2); TEST_DO(assertHandler(bucket1, tstamp1, docId1, f.hset.handler1)); TEST_DO(assertHandler(bucket1, tstamp1, docId2, f.hset.handler2)); EXPECT_EQUAL(tstamp3, ur.getExistingTimestamp()); EXPECT_EQUAL(Result(Result::ErrorType::PERMANENT_ERROR, "No handler for document type 'type3'"), - f.engine.update(bucket1, tstamp1, upd3, context)); + f.engine.update(bucket1, tstamp1, upd3)); } TEST_F("require that updates with bad ids are rejected", SimpleFixture) { - Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); - EXPECT_EQUAL(UpdateResult(Result::ErrorType::PERMANENT_ERROR, "Update operation rejected due to bad id (id:type2:type2::1, type1)"), - f.engine.update(bucket1, tstamp1, bad_id_upd, context)); + f.engine.update(bucket1, tstamp1, bad_id_upd)); } TEST_F("require that simple, cheap update is not rejected if resource limit is reached", SimpleFixture) @@ -505,10 +500,8 @@ TEST_F("require that simple, cheap update is not rejected if resource limit is r f._writeFilter._acceptWriteOperation = false; f._writeFilter._message = "Disk is full"; - Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); - EXPECT_EQUAL(Result(Result::ErrorType::NONE, ""), - f.engine.update(bucket1, tstamp1, upd1, context)); + f.engine.update(bucket1, tstamp1, upd1)); } TEST_F("require that update is rejected if resource limit is reached", SimpleFixture) @@ -516,7 +509,6 @@ TEST_F("require that update is rejected if resource limit is reached", SimpleFix f._writeFilter._acceptWriteOperation = false; f._writeFilter._message = "Disk is full"; - Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); DocumentType type(createDocType("type_with_one_string", 1)); document::Field field("string", 1, *document::DataType::STRING); type.addField(field); @@ -526,13 +518,12 @@ TEST_F("require that update is rejected if resource limit is reached", SimpleFix EXPECT_EQUAL( Result(Result::ErrorType::RESOURCE_EXHAUSTED, "Update operation rejected for document 'id:type1:type1::1': 'Disk is full'"), - f.engine.update(bucket1, tstamp1, upd, context)); + f.engine.update(bucket1, tstamp1, upd)); } TEST_F("require that removes are routed to handlers", SimpleFixture) { - Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); - RemoveResult rr = f.engine.remove(bucket1, tstamp1, docId3, context); + RemoveResult rr = f.engine.remove(bucket1, tstamp1, docId3); TEST_DO(assertHandler(bucket0, tstamp0, docId0, f.hset.handler1)); TEST_DO(assertHandler(bucket0, tstamp0, docId0, f.hset.handler2)); EXPECT_FALSE(rr.wasFound()); @@ -540,7 +531,7 @@ TEST_F("require that removes are routed to handlers", SimpleFixture) EXPECT_EQUAL(Result(Result::ErrorType::PERMANENT_ERROR, "No handler for document type 'type3'"), rr); f.hset.handler1.setExistingTimestamp(tstamp2); - rr = f.engine.remove(bucket1, tstamp1, docId1, context); + rr = f.engine.remove(bucket1, tstamp1, docId1); TEST_DO(assertHandler(bucket1, tstamp1, docId1, f.hset.handler1)); TEST_DO(assertHandler(bucket0, tstamp0, docId0, f.hset.handler2)); EXPECT_TRUE(rr.wasFound()); @@ -548,14 +539,14 @@ TEST_F("require that removes are routed to handlers", SimpleFixture) f.hset.handler1.setExistingTimestamp(tstamp0); f.hset.handler2.setExistingTimestamp(tstamp3); - rr = f.engine.remove(bucket1, tstamp1, docId2, context); + rr = f.engine.remove(bucket1, tstamp1, docId2); TEST_DO(assertHandler(bucket1, tstamp1, docId1, f.hset.handler1)); TEST_DO(assertHandler(bucket1, tstamp1, docId2, f.hset.handler2)); EXPECT_TRUE(rr.wasFound()); EXPECT_FALSE(rr.hasError()); f.hset.handler2.setExistingTimestamp(tstamp0); - rr = f.engine.remove(bucket1, tstamp1, docId2, context); + rr = f.engine.remove(bucket1, tstamp1, docId2); TEST_DO(assertHandler(bucket1, tstamp1, docId1, f.hset.handler1)); TEST_DO(assertHandler(bucket1, tstamp1, docId2, f.hset.handler2)); EXPECT_FALSE(rr.wasFound()); @@ -567,9 +558,7 @@ TEST_F("require that remove is NOT rejected if resource limit is reached", Simpl f._writeFilter._acceptWriteOperation = false; f._writeFilter._message = "Disk is full"; - Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); - - EXPECT_EQUAL(RemoveResult(false), f.engine.remove(bucket1, tstamp1, docId1, context)); + EXPECT_EQUAL(RemoveResult(false), f.engine.remove(bucket1, tstamp1, docId1)); } @@ -619,11 +608,10 @@ TEST_F("require that getBucketInfo() is routed to handlers and merged", SimpleFi TEST_F("require that createBucket() is routed to handlers and merged", SimpleFixture) { - Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); f.hset.handler1._createBucketResult = Result(Result::ErrorType::TRANSIENT_ERROR, "err1a"); f.hset.handler2._createBucketResult = Result(Result::ErrorType::PERMANENT_ERROR, "err2a"); - Result result = f.engine.createBucket(bucket1, context); + Result result = f.engine.createBucket(bucket1); EXPECT_EQUAL(Result::ErrorType::PERMANENT_ERROR, result.getErrorCode()); EXPECT_EQUAL("err1a, err2a", result.getErrorMessage()); } @@ -631,11 +619,10 @@ TEST_F("require that createBucket() is routed to handlers and merged", SimpleFix TEST_F("require that deleteBucket() is routed to handlers and merged", SimpleFixture) { - Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); f.hset.handler1.deleteBucketResult = Result(Result::ErrorType::TRANSIENT_ERROR, "err1"); f.hset.handler2.deleteBucketResult = Result(Result::ErrorType::PERMANENT_ERROR, "err2"); - Result result = f.engine.deleteBucket(bucket1, context); + Result result = f.engine.deleteBucket(bucket1); EXPECT_EQUAL(Result::ErrorType::PERMANENT_ERROR, result.getErrorCode()); EXPECT_EQUAL("err1, err2", result.getErrorMessage()); } @@ -680,7 +667,7 @@ TEST_F("require that createIterator does", SimpleFixture) { EXPECT_TRUE(result.getIteratorId()); uint64_t max_size = 1024; - IterateResult it_result = f.engine.iterate(result.getIteratorId(), max_size, context); + IterateResult it_result = f.engine.iterate(result.getIteratorId(), max_size); EXPECT_FALSE(it_result.hasError()); } @@ -700,7 +687,7 @@ TEST_F("require that iterator ids are unique", SimpleFixture) { TEST_F("require that iterate requires valid iterator", SimpleFixture) { uint64_t max_size = 1024; Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); - IterateResult it_result = f.engine.iterate(IteratorId(1), max_size, context); + IterateResult it_result = f.engine.iterate(IteratorId(1), max_size); EXPECT_TRUE(it_result.hasError()); EXPECT_EQUAL(Result::ErrorType::PERMANENT_ERROR, it_result.getErrorCode()); EXPECT_EQUAL("Unknown iterator with id 1", it_result.getErrorMessage()); @@ -710,7 +697,7 @@ TEST_F("require that iterate requires valid iterator", SimpleFixture) { storage::spi::NEWEST_DOCUMENT_ONLY, context); EXPECT_TRUE(result.getIteratorId()); - it_result = f.engine.iterate(result.getIteratorId(), max_size, context); + it_result = f.engine.iterate(result.getIteratorId(), max_size); EXPECT_FALSE(it_result.hasError()); } @@ -725,7 +712,7 @@ TEST_F("require that iterate returns documents", SimpleFixture) { storage::spi::NEWEST_DOCUMENT_ONLY, context); EXPECT_TRUE(result.getIteratorId()); - IterateResult it_result = f.engine.iterate(result.getIteratorId(), max_size, context); + IterateResult it_result = f.engine.iterate(result.getIteratorId(), max_size); EXPECT_FALSE(it_result.hasError()); EXPECT_EQUAL(2u, it_result.getEntries().size()); } @@ -739,11 +726,11 @@ TEST_F("require that destroyIterator prevents iteration", SimpleFixture) { storage::spi::NEWEST_DOCUMENT_ONLY, context); EXPECT_TRUE(create_result.getIteratorId()); - Result result = f.engine.destroyIterator(create_result.getIteratorId(), context); + Result result = f.engine.destroyIterator(create_result.getIteratorId()); EXPECT_FALSE(result.hasError()); uint64_t max_size = 1024; - IterateResult it_result = f.engine.iterate(create_result.getIteratorId(), max_size, context); + IterateResult it_result = f.engine.iterate(create_result.getIteratorId(), max_size); EXPECT_TRUE(it_result.hasError()); EXPECT_EQUAL(Result::ErrorType::PERMANENT_ERROR, it_result.getErrorCode()); string msg_prefix = "Unknown iterator with id"; diff --git a/searchcore/src/vespa/searchcore/bmcluster/bm_node.cpp b/searchcore/src/vespa/searchcore/bmcluster/bm_node.cpp index 9cab70cd72b..4db46ead525 100644 --- a/searchcore/src/vespa/searchcore/bmcluster/bm_node.cpp +++ b/searchcore/src/vespa/searchcore/bmcluster/bm_node.cpp @@ -7,12 +7,7 @@ #include "bm_node_stats.h" #include "bm_storage_chain_builder.h" #include "bm_storage_link_context.h" -#include "document_api_message_bus_bm_feed_handler.h" #include "i_bm_distribution.h" -#include "i_bm_feed_handler.h" -#include "spi_bm_feed_handler.h" -#include "storage_api_chain_bm_feed_handler.h" -#include "storage_api_message_bus_bm_feed_handler.h" #include "storage_api_rpc_bm_feed_handler.h" #include <tests/proton/common/dummydbowner.h> #include <vespa/config-attributes.h> @@ -60,7 +55,6 @@ #include <vespa/searchsummary/config/config-juniperrc.h> #include <vespa/storage/bucketdb/config-stor-bucket-init.h> #include <vespa/storage/common/i_storage_chain_builder.h> -#include <vespa/storage/common/storagelink.h> #include <vespa/storage/config/config-stor-bouncer.h> #include <vespa/storage/config/config-stor-communicationmanager.h> #include <vespa/storage/config/config-stor-distributormanager.h> @@ -155,8 +149,6 @@ int port_number(int base_port, PortBias bias) return base_port + static_cast<int>(bias); } -storage::spi::Context context(storage::spi::Priority(0), 0); - template <class ChainLink, class Process> ChainLink* extract_chain_link(Process &process) { @@ -610,7 +602,7 @@ MyBmNode::initialize_persistence_provider() void MyBmNode::create_bucket(const document::Bucket& bucket) { - get_persistence_provider()->createBucket(storage::spi::Bucket(bucket), context); + get_persistence_provider()->createBucket(storage::spi::Bucket(bucket)); } void diff --git a/searchcore/src/vespa/searchcore/bmcluster/spi_bm_feed_handler.cpp b/searchcore/src/vespa/searchcore/bmcluster/spi_bm_feed_handler.cpp index d5421eaaeca..69013e8d7c5 100644 --- a/searchcore/src/vespa/searchcore/bmcluster/spi_bm_feed_handler.cpp +++ b/searchcore/src/vespa/searchcore/bmcluster/spi_bm_feed_handler.cpp @@ -108,7 +108,7 @@ SpiBmFeedHandler::put(const document::Bucket& bucket, std::unique_ptr<Document> auto provider = get_provider(bucket); if (provider) { Bucket spi_bucket(bucket); - provider->putAsync(spi_bucket, Timestamp(timestamp), std::move(document), context, std::make_unique<MyOperationComplete>(provider, _errors, spi_bucket, tracker)); + provider->putAsync(spi_bucket, Timestamp(timestamp), std::move(document), std::make_unique<MyOperationComplete>(provider, _errors, spi_bucket, tracker)); } else { ++_errors; } @@ -121,7 +121,7 @@ SpiBmFeedHandler::update(const document::Bucket& bucket, std::unique_ptr<Documen auto provider = get_provider(bucket); if (provider) { Bucket spi_bucket(bucket); - provider->updateAsync(spi_bucket, Timestamp(timestamp), std::move(document_update), context, std::make_unique<MyOperationComplete>(provider, _errors, spi_bucket, tracker)); + provider->updateAsync(spi_bucket, Timestamp(timestamp), std::move(document_update), std::make_unique<MyOperationComplete>(provider, _errors, spi_bucket, tracker)); } else { ++_errors; } @@ -136,7 +136,7 @@ SpiBmFeedHandler::remove(const document::Bucket& bucket, const DocumentId& docum Bucket spi_bucket(bucket); std::vector<storage::spi::PersistenceProvider::TimeStampAndDocumentId> ids; ids.emplace_back(Timestamp(timestamp), document_id); - provider->removeAsync(spi_bucket, std::move(ids), context, std::make_unique<MyOperationComplete>(provider, _errors, spi_bucket, tracker)); + provider->removeAsync(spi_bucket, std::move(ids), std::make_unique<MyOperationComplete>(provider, _errors, spi_bucket, tracker)); } else { ++_errors; } diff --git a/searchcore/src/vespa/searchcore/proton/persistenceengine/persistenceengine.cpp b/searchcore/src/vespa/searchcore/proton/persistenceengine/persistenceengine.cpp index 94882bfadf5..0621612aca7 100644 --- a/searchcore/src/vespa/searchcore/proton/persistenceengine/persistenceengine.cpp +++ b/searchcore/src/vespa/searchcore/proton/persistenceengine/persistenceengine.cpp @@ -343,7 +343,7 @@ PersistenceEngine::getBucketInfo(const Bucket& b) const void -PersistenceEngine::putAsync(const Bucket &bucket, Timestamp ts, storage::spi::DocumentSP doc, Context &, OperationComplete::UP onComplete) +PersistenceEngine::putAsync(const Bucket &bucket, Timestamp ts, storage::spi::DocumentSP doc, OperationComplete::UP onComplete) { if (!_writeFilter.acceptWriteOperation()) { IResourceWriteFilter::State state = _writeFilter.getAcceptState(); @@ -370,7 +370,7 @@ PersistenceEngine::putAsync(const Bucket &bucket, Timestamp ts, storage::spi::Do } void -PersistenceEngine::removeAsync(const Bucket& b, std::vector<TimeStampAndDocumentId> ids, Context&, OperationComplete::UP onComplete) +PersistenceEngine::removeAsync(const Bucket& b, std::vector<TimeStampAndDocumentId> ids, OperationComplete::UP onComplete) { if (ids.size() == 1) { removeAsyncSingle(b, ids[0].first, ids[0].second, std::move(onComplete)); @@ -429,7 +429,7 @@ PersistenceEngine::removeAsyncSingle(const Bucket& b, Timestamp t, const Documen void -PersistenceEngine::updateAsync(const Bucket& b, Timestamp t, DocumentUpdate::SP upd, Context&, OperationComplete::UP onComplete) +PersistenceEngine::updateAsync(const Bucket& b, Timestamp t, DocumentUpdate::SP upd, OperationComplete::UP onComplete) { if (!_writeFilter.acceptWriteOperation()) { IResourceWriteFilter::State state = _writeFilter.getAcceptState(); @@ -534,7 +534,7 @@ PersistenceEngine::createIterator(const Bucket &bucket, FieldSetSP fields, const PersistenceEngine::IterateResult -PersistenceEngine::iterate(IteratorId id, uint64_t maxByteSize, Context&) const +PersistenceEngine::iterate(IteratorId id, uint64_t maxByteSize) const { ReadGuard rguard(_rwMutex); IteratorEntry *iteratorEntry; @@ -568,7 +568,7 @@ PersistenceEngine::iterate(IteratorId id, uint64_t maxByteSize, Context&) const Result -PersistenceEngine::destroyIterator(IteratorId id, Context&) +PersistenceEngine::destroyIterator(IteratorId id) { ReadGuard rguard(_rwMutex); std::lock_guard<std::mutex> guard(_iterators_lock); @@ -586,7 +586,7 @@ PersistenceEngine::destroyIterator(IteratorId id, Context&) void -PersistenceEngine::createBucketAsync(const Bucket &b, Context &, OperationComplete::UP onComplete) noexcept +PersistenceEngine::createBucketAsync(const Bucket &b, OperationComplete::UP onComplete) noexcept { ReadGuard rguard(_rwMutex); LOG(spam, "createBucket(%s)", b.toString().c_str()); @@ -606,7 +606,7 @@ PersistenceEngine::createBucketAsync(const Bucket &b, Context &, OperationComple void -PersistenceEngine::deleteBucketAsync(const Bucket& b, Context&, OperationComplete::UP onComplete) noexcept +PersistenceEngine::deleteBucketAsync(const Bucket& b, OperationComplete::UP onComplete) noexcept { ReadGuard rguard(_rwMutex); LOG(spam, "deleteBucket(%s)", b.toString().c_str()); @@ -651,7 +651,7 @@ PersistenceEngine::getModifiedBuckets(BucketSpace bucketSpace) const Result -PersistenceEngine::split(const Bucket& source, const Bucket& target1, const Bucket& target2, Context&) +PersistenceEngine::split(const Bucket& source, const Bucket& target1, const Bucket& target2) { ReadGuard rguard(_rwMutex); LOG(spam, "split(%s, %s, %s)", source.toString().c_str(), target1.toString().c_str(), target2.toString().c_str()); @@ -669,7 +669,7 @@ PersistenceEngine::split(const Bucket& source, const Bucket& target1, const Buck Result -PersistenceEngine::join(const Bucket& source1, const Bucket& source2, const Bucket& target, Context&) +PersistenceEngine::join(const Bucket& source1, const Bucket& source2, const Bucket& target) { ReadGuard rguard(_rwMutex); LOG(spam, "join(%s, %s, %s)", source1.toString().c_str(), source2.toString().c_str(), target.toString().c_str()); @@ -694,7 +694,6 @@ PersistenceEngine::register_resource_usage_listener(IResourceUsageListener& list void PersistenceEngine::destroyIterators() { - Context context(storage::spi::Priority(0x80), 0); for (;;) { IteratorId id; { @@ -703,7 +702,7 @@ PersistenceEngine::destroyIterators() break; id = _iterators.begin()->first; } - Result res(destroyIterator(id, context)); + Result res(destroyIterator(id)); if (res.hasError()) { LOG(debug, "%zu iterator left. Can not destroy iterator '%" PRIu64 "'. Reason='%s'", _iterators.size(), id.getValue(), res.toString().c_str()); std::this_thread::sleep_for(100ms); diff --git a/searchcore/src/vespa/searchcore/proton/persistenceengine/persistenceengine.h b/searchcore/src/vespa/searchcore/proton/persistenceengine/persistenceengine.h index 7c8040fae9d..a8886e19def 100644 --- a/searchcore/src/vespa/searchcore/proton/persistenceengine/persistenceengine.h +++ b/searchcore/src/vespa/searchcore/proton/persistenceengine/persistenceengine.h @@ -106,20 +106,20 @@ public: Result setClusterState(BucketSpace bucketSpace, const ClusterState& calc) override; void setActiveStateAsync(const Bucket&, BucketInfo::ActiveState, OperationComplete::UP) override; BucketInfoResult getBucketInfo(const Bucket&) const override; - void putAsync(const Bucket &, Timestamp, storage::spi::DocumentSP, Context &context, OperationComplete::UP) override; - void removeAsync(const Bucket&, std::vector<TimeStampAndDocumentId> ids, Context&, OperationComplete::UP) override; - void updateAsync(const Bucket&, Timestamp, storage::spi::DocumentUpdateSP, Context&, OperationComplete::UP) override; + void putAsync(const Bucket &, Timestamp, storage::spi::DocumentSP, OperationComplete::UP) override; + void removeAsync(const Bucket&, std::vector<TimeStampAndDocumentId> ids, OperationComplete::UP) override; + void updateAsync(const Bucket&, Timestamp, storage::spi::DocumentUpdateSP, OperationComplete::UP) override; GetResult get(const Bucket&, const document::FieldSet&, const document::DocumentId&, Context&) const override; CreateIteratorResult createIterator(const Bucket &bucket, FieldSetSP, const Selection &, IncludedVersions, Context &context) override; - IterateResult iterate(IteratorId, uint64_t maxByteSize, Context&) const override; - Result destroyIterator(IteratorId, Context&) override; + IterateResult iterate(IteratorId, uint64_t maxByteSize) const override; + Result destroyIterator(IteratorId) override; - void createBucketAsync(const Bucket &bucketId, Context &, OperationComplete::UP) noexcept override; - void deleteBucketAsync(const Bucket&, Context&, OperationComplete::UP) noexcept override; + void createBucketAsync(const Bucket &bucketId, OperationComplete::UP) noexcept override; + void deleteBucketAsync(const Bucket&, OperationComplete::UP) noexcept override; BucketIdListResult getModifiedBuckets(BucketSpace bucketSpace) const override; - Result split(const Bucket& source, const Bucket& target1, const Bucket& target2, Context&) override; - Result join(const Bucket& source1, const Bucket& source2, const Bucket& target, Context&) override; + Result split(const Bucket& source, const Bucket& target1, const Bucket& target2) override; + Result join(const Bucket& source1, const Bucket& source2, const Bucket& target) override; std::unique_ptr<vespalib::IDestructorCallback> register_resource_usage_listener(IResourceUsageListener& listener) override; std::unique_ptr<vespalib::IDestructorCallback> register_executor(std::shared_ptr<BucketExecutor>) override; void destroyIterators(); |