summaryrefslogtreecommitdiffstats
path: root/searchcore
diff options
context:
space:
mode:
authorHenning Baldersheim <balder@yahoo-inc.com>2022-03-31 13:00:32 +0000
committerHenning Baldersheim <balder@yahoo-inc.com>2022-03-31 13:00:32 +0000
commit94f53ab6a93aefa007200b97846ce47ea166bd70 (patch)
tree0421134fe3a839da8b36c18636bd901ae9939f60 /searchcore
parentd0fa48f2559c297e7b9c15137f3c04bc32fc65fa (diff)
GC unused Context parameter
Diffstat (limited to 'searchcore')
-rw-r--r--searchcore/src/tests/proton/persistenceengine/persistenceengine_test.cpp59
-rw-r--r--searchcore/src/vespa/searchcore/bmcluster/bm_node.cpp10
-rw-r--r--searchcore/src/vespa/searchcore/bmcluster/spi_bm_feed_handler.cpp6
-rw-r--r--searchcore/src/vespa/searchcore/proton/persistenceengine/persistenceengine.cpp21
-rw-r--r--searchcore/src/vespa/searchcore/proton/persistenceengine/persistenceengine.h18
5 files changed, 46 insertions, 68 deletions
diff --git a/searchcore/src/tests/proton/persistenceengine/persistenceengine_test.cpp b/searchcore/src/tests/proton/persistenceengine/persistenceengine_test.cpp
index 6fb7a35fffe..cf58056cae8 100644
--- a/searchcore/src/tests/proton/persistenceengine/persistenceengine_test.cpp
+++ b/searchcore/src/tests/proton/persistenceengine/persistenceengine_test.cpp
@@ -446,17 +446,16 @@ TEST_F("require that getPartitionStates() prepares all handlers", SimpleFixture)
TEST_F("require that puts are routed to handler", SimpleFixture)
{
- Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0));
- f.engine.put(bucket1, tstamp1, doc1, context);
+ f.engine.put(bucket1, tstamp1, doc1);
TEST_DO(assertHandler(bucket1, tstamp1, docId1, f.hset.handler1));
TEST_DO(assertHandler(bucket0, tstamp0, docId0, f.hset.handler2));
- f.engine.put(bucket1, tstamp1, doc2, context);
+ f.engine.put(bucket1, tstamp1, doc2);
TEST_DO(assertHandler(bucket1, tstamp1, docId1, f.hset.handler1));
TEST_DO(assertHandler(bucket1, tstamp1, docId2, f.hset.handler2));
EXPECT_EQUAL(Result(Result::ErrorType::PERMANENT_ERROR, "No handler for document type 'type3'"),
- f.engine.put(bucket1, tstamp1, doc3, context));
+ f.engine.put(bucket1, tstamp1, doc3));
}
@@ -465,39 +464,35 @@ TEST_F("require that put is rejected if resource limit is reached", SimpleFixtur
f._writeFilter._acceptWriteOperation = false;
f._writeFilter._message = "Disk is full";
- Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0));
EXPECT_EQUAL(
Result(Result::ErrorType::RESOURCE_EXHAUSTED,
"Put operation rejected for document 'id:type3:type3::1': 'Disk is full'"),
- f.engine.put(bucket1, tstamp1, doc3, context));
+ f.engine.put(bucket1, tstamp1, doc3));
}
TEST_F("require that updates are routed to handler", SimpleFixture)
{
- Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0));
f.hset.handler1.setExistingTimestamp(tstamp2);
- UpdateResult ur = f.engine.update(bucket1, tstamp1, upd1, context);
+ UpdateResult ur = f.engine.update(bucket1, tstamp1, upd1);
TEST_DO(assertHandler(bucket1, tstamp1, docId1, f.hset.handler1));
TEST_DO(assertHandler(bucket0, tstamp0, docId0, f.hset.handler2));
EXPECT_EQUAL(tstamp2, ur.getExistingTimestamp());
f.hset.handler2.setExistingTimestamp(tstamp3);
- ur = f.engine.update(bucket1, tstamp1, upd2, context);
+ ur = f.engine.update(bucket1, tstamp1, upd2);
TEST_DO(assertHandler(bucket1, tstamp1, docId1, f.hset.handler1));
TEST_DO(assertHandler(bucket1, tstamp1, docId2, f.hset.handler2));
EXPECT_EQUAL(tstamp3, ur.getExistingTimestamp());
EXPECT_EQUAL(Result(Result::ErrorType::PERMANENT_ERROR, "No handler for document type 'type3'"),
- f.engine.update(bucket1, tstamp1, upd3, context));
+ f.engine.update(bucket1, tstamp1, upd3));
}
TEST_F("require that updates with bad ids are rejected", SimpleFixture)
{
- Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0));
-
EXPECT_EQUAL(UpdateResult(Result::ErrorType::PERMANENT_ERROR, "Update operation rejected due to bad id (id:type2:type2::1, type1)"),
- f.engine.update(bucket1, tstamp1, bad_id_upd, context));
+ f.engine.update(bucket1, tstamp1, bad_id_upd));
}
TEST_F("require that simple, cheap update is not rejected if resource limit is reached", SimpleFixture)
@@ -505,10 +500,8 @@ TEST_F("require that simple, cheap update is not rejected if resource limit is r
f._writeFilter._acceptWriteOperation = false;
f._writeFilter._message = "Disk is full";
- Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0));
-
EXPECT_EQUAL(Result(Result::ErrorType::NONE, ""),
- f.engine.update(bucket1, tstamp1, upd1, context));
+ f.engine.update(bucket1, tstamp1, upd1));
}
TEST_F("require that update is rejected if resource limit is reached", SimpleFixture)
@@ -516,7 +509,6 @@ TEST_F("require that update is rejected if resource limit is reached", SimpleFix
f._writeFilter._acceptWriteOperation = false;
f._writeFilter._message = "Disk is full";
- Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0));
DocumentType type(createDocType("type_with_one_string", 1));
document::Field field("string", 1, *document::DataType::STRING);
type.addField(field);
@@ -526,13 +518,12 @@ TEST_F("require that update is rejected if resource limit is reached", SimpleFix
EXPECT_EQUAL(
Result(Result::ErrorType::RESOURCE_EXHAUSTED,
"Update operation rejected for document 'id:type1:type1::1': 'Disk is full'"),
- f.engine.update(bucket1, tstamp1, upd, context));
+ f.engine.update(bucket1, tstamp1, upd));
}
TEST_F("require that removes are routed to handlers", SimpleFixture)
{
- Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0));
- RemoveResult rr = f.engine.remove(bucket1, tstamp1, docId3, context);
+ RemoveResult rr = f.engine.remove(bucket1, tstamp1, docId3);
TEST_DO(assertHandler(bucket0, tstamp0, docId0, f.hset.handler1));
TEST_DO(assertHandler(bucket0, tstamp0, docId0, f.hset.handler2));
EXPECT_FALSE(rr.wasFound());
@@ -540,7 +531,7 @@ TEST_F("require that removes are routed to handlers", SimpleFixture)
EXPECT_EQUAL(Result(Result::ErrorType::PERMANENT_ERROR, "No handler for document type 'type3'"), rr);
f.hset.handler1.setExistingTimestamp(tstamp2);
- rr = f.engine.remove(bucket1, tstamp1, docId1, context);
+ rr = f.engine.remove(bucket1, tstamp1, docId1);
TEST_DO(assertHandler(bucket1, tstamp1, docId1, f.hset.handler1));
TEST_DO(assertHandler(bucket0, tstamp0, docId0, f.hset.handler2));
EXPECT_TRUE(rr.wasFound());
@@ -548,14 +539,14 @@ TEST_F("require that removes are routed to handlers", SimpleFixture)
f.hset.handler1.setExistingTimestamp(tstamp0);
f.hset.handler2.setExistingTimestamp(tstamp3);
- rr = f.engine.remove(bucket1, tstamp1, docId2, context);
+ rr = f.engine.remove(bucket1, tstamp1, docId2);
TEST_DO(assertHandler(bucket1, tstamp1, docId1, f.hset.handler1));
TEST_DO(assertHandler(bucket1, tstamp1, docId2, f.hset.handler2));
EXPECT_TRUE(rr.wasFound());
EXPECT_FALSE(rr.hasError());
f.hset.handler2.setExistingTimestamp(tstamp0);
- rr = f.engine.remove(bucket1, tstamp1, docId2, context);
+ rr = f.engine.remove(bucket1, tstamp1, docId2);
TEST_DO(assertHandler(bucket1, tstamp1, docId1, f.hset.handler1));
TEST_DO(assertHandler(bucket1, tstamp1, docId2, f.hset.handler2));
EXPECT_FALSE(rr.wasFound());
@@ -567,9 +558,7 @@ TEST_F("require that remove is NOT rejected if resource limit is reached", Simpl
f._writeFilter._acceptWriteOperation = false;
f._writeFilter._message = "Disk is full";
- Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0));
-
- EXPECT_EQUAL(RemoveResult(false), f.engine.remove(bucket1, tstamp1, docId1, context));
+ EXPECT_EQUAL(RemoveResult(false), f.engine.remove(bucket1, tstamp1, docId1));
}
@@ -619,11 +608,10 @@ TEST_F("require that getBucketInfo() is routed to handlers and merged", SimpleFi
TEST_F("require that createBucket() is routed to handlers and merged", SimpleFixture)
{
- Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0));
f.hset.handler1._createBucketResult = Result(Result::ErrorType::TRANSIENT_ERROR, "err1a");
f.hset.handler2._createBucketResult = Result(Result::ErrorType::PERMANENT_ERROR, "err2a");
- Result result = f.engine.createBucket(bucket1, context);
+ Result result = f.engine.createBucket(bucket1);
EXPECT_EQUAL(Result::ErrorType::PERMANENT_ERROR, result.getErrorCode());
EXPECT_EQUAL("err1a, err2a", result.getErrorMessage());
}
@@ -631,11 +619,10 @@ TEST_F("require that createBucket() is routed to handlers and merged", SimpleFix
TEST_F("require that deleteBucket() is routed to handlers and merged", SimpleFixture)
{
- Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0));
f.hset.handler1.deleteBucketResult = Result(Result::ErrorType::TRANSIENT_ERROR, "err1");
f.hset.handler2.deleteBucketResult = Result(Result::ErrorType::PERMANENT_ERROR, "err2");
- Result result = f.engine.deleteBucket(bucket1, context);
+ Result result = f.engine.deleteBucket(bucket1);
EXPECT_EQUAL(Result::ErrorType::PERMANENT_ERROR, result.getErrorCode());
EXPECT_EQUAL("err1, err2", result.getErrorMessage());
}
@@ -680,7 +667,7 @@ TEST_F("require that createIterator does", SimpleFixture) {
EXPECT_TRUE(result.getIteratorId());
uint64_t max_size = 1024;
- IterateResult it_result = f.engine.iterate(result.getIteratorId(), max_size, context);
+ IterateResult it_result = f.engine.iterate(result.getIteratorId(), max_size);
EXPECT_FALSE(it_result.hasError());
}
@@ -700,7 +687,7 @@ TEST_F("require that iterator ids are unique", SimpleFixture) {
TEST_F("require that iterate requires valid iterator", SimpleFixture) {
uint64_t max_size = 1024;
Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0));
- IterateResult it_result = f.engine.iterate(IteratorId(1), max_size, context);
+ IterateResult it_result = f.engine.iterate(IteratorId(1), max_size);
EXPECT_TRUE(it_result.hasError());
EXPECT_EQUAL(Result::ErrorType::PERMANENT_ERROR, it_result.getErrorCode());
EXPECT_EQUAL("Unknown iterator with id 1", it_result.getErrorMessage());
@@ -710,7 +697,7 @@ TEST_F("require that iterate requires valid iterator", SimpleFixture) {
storage::spi::NEWEST_DOCUMENT_ONLY, context);
EXPECT_TRUE(result.getIteratorId());
- it_result = f.engine.iterate(result.getIteratorId(), max_size, context);
+ it_result = f.engine.iterate(result.getIteratorId(), max_size);
EXPECT_FALSE(it_result.hasError());
}
@@ -725,7 +712,7 @@ TEST_F("require that iterate returns documents", SimpleFixture) {
storage::spi::NEWEST_DOCUMENT_ONLY, context);
EXPECT_TRUE(result.getIteratorId());
- IterateResult it_result = f.engine.iterate(result.getIteratorId(), max_size, context);
+ IterateResult it_result = f.engine.iterate(result.getIteratorId(), max_size);
EXPECT_FALSE(it_result.hasError());
EXPECT_EQUAL(2u, it_result.getEntries().size());
}
@@ -739,11 +726,11 @@ TEST_F("require that destroyIterator prevents iteration", SimpleFixture) {
storage::spi::NEWEST_DOCUMENT_ONLY, context);
EXPECT_TRUE(create_result.getIteratorId());
- Result result = f.engine.destroyIterator(create_result.getIteratorId(), context);
+ Result result = f.engine.destroyIterator(create_result.getIteratorId());
EXPECT_FALSE(result.hasError());
uint64_t max_size = 1024;
- IterateResult it_result = f.engine.iterate(create_result.getIteratorId(), max_size, context);
+ IterateResult it_result = f.engine.iterate(create_result.getIteratorId(), max_size);
EXPECT_TRUE(it_result.hasError());
EXPECT_EQUAL(Result::ErrorType::PERMANENT_ERROR, it_result.getErrorCode());
string msg_prefix = "Unknown iterator with id";
diff --git a/searchcore/src/vespa/searchcore/bmcluster/bm_node.cpp b/searchcore/src/vespa/searchcore/bmcluster/bm_node.cpp
index 9cab70cd72b..4db46ead525 100644
--- a/searchcore/src/vespa/searchcore/bmcluster/bm_node.cpp
+++ b/searchcore/src/vespa/searchcore/bmcluster/bm_node.cpp
@@ -7,12 +7,7 @@
#include "bm_node_stats.h"
#include "bm_storage_chain_builder.h"
#include "bm_storage_link_context.h"
-#include "document_api_message_bus_bm_feed_handler.h"
#include "i_bm_distribution.h"
-#include "i_bm_feed_handler.h"
-#include "spi_bm_feed_handler.h"
-#include "storage_api_chain_bm_feed_handler.h"
-#include "storage_api_message_bus_bm_feed_handler.h"
#include "storage_api_rpc_bm_feed_handler.h"
#include <tests/proton/common/dummydbowner.h>
#include <vespa/config-attributes.h>
@@ -60,7 +55,6 @@
#include <vespa/searchsummary/config/config-juniperrc.h>
#include <vespa/storage/bucketdb/config-stor-bucket-init.h>
#include <vespa/storage/common/i_storage_chain_builder.h>
-#include <vespa/storage/common/storagelink.h>
#include <vespa/storage/config/config-stor-bouncer.h>
#include <vespa/storage/config/config-stor-communicationmanager.h>
#include <vespa/storage/config/config-stor-distributormanager.h>
@@ -155,8 +149,6 @@ int port_number(int base_port, PortBias bias)
return base_port + static_cast<int>(bias);
}
-storage::spi::Context context(storage::spi::Priority(0), 0);
-
template <class ChainLink, class Process>
ChainLink* extract_chain_link(Process &process)
{
@@ -610,7 +602,7 @@ MyBmNode::initialize_persistence_provider()
void
MyBmNode::create_bucket(const document::Bucket& bucket)
{
- get_persistence_provider()->createBucket(storage::spi::Bucket(bucket), context);
+ get_persistence_provider()->createBucket(storage::spi::Bucket(bucket));
}
void
diff --git a/searchcore/src/vespa/searchcore/bmcluster/spi_bm_feed_handler.cpp b/searchcore/src/vespa/searchcore/bmcluster/spi_bm_feed_handler.cpp
index d5421eaaeca..69013e8d7c5 100644
--- a/searchcore/src/vespa/searchcore/bmcluster/spi_bm_feed_handler.cpp
+++ b/searchcore/src/vespa/searchcore/bmcluster/spi_bm_feed_handler.cpp
@@ -108,7 +108,7 @@ SpiBmFeedHandler::put(const document::Bucket& bucket, std::unique_ptr<Document>
auto provider = get_provider(bucket);
if (provider) {
Bucket spi_bucket(bucket);
- provider->putAsync(spi_bucket, Timestamp(timestamp), std::move(document), context, std::make_unique<MyOperationComplete>(provider, _errors, spi_bucket, tracker));
+ provider->putAsync(spi_bucket, Timestamp(timestamp), std::move(document), std::make_unique<MyOperationComplete>(provider, _errors, spi_bucket, tracker));
} else {
++_errors;
}
@@ -121,7 +121,7 @@ SpiBmFeedHandler::update(const document::Bucket& bucket, std::unique_ptr<Documen
auto provider = get_provider(bucket);
if (provider) {
Bucket spi_bucket(bucket);
- provider->updateAsync(spi_bucket, Timestamp(timestamp), std::move(document_update), context, std::make_unique<MyOperationComplete>(provider, _errors, spi_bucket, tracker));
+ provider->updateAsync(spi_bucket, Timestamp(timestamp), std::move(document_update), std::make_unique<MyOperationComplete>(provider, _errors, spi_bucket, tracker));
} else {
++_errors;
}
@@ -136,7 +136,7 @@ SpiBmFeedHandler::remove(const document::Bucket& bucket, const DocumentId& docum
Bucket spi_bucket(bucket);
std::vector<storage::spi::PersistenceProvider::TimeStampAndDocumentId> ids;
ids.emplace_back(Timestamp(timestamp), document_id);
- provider->removeAsync(spi_bucket, std::move(ids), context, std::make_unique<MyOperationComplete>(provider, _errors, spi_bucket, tracker));
+ provider->removeAsync(spi_bucket, std::move(ids), std::make_unique<MyOperationComplete>(provider, _errors, spi_bucket, tracker));
} else {
++_errors;
}
diff --git a/searchcore/src/vespa/searchcore/proton/persistenceengine/persistenceengine.cpp b/searchcore/src/vespa/searchcore/proton/persistenceengine/persistenceengine.cpp
index 94882bfadf5..0621612aca7 100644
--- a/searchcore/src/vespa/searchcore/proton/persistenceengine/persistenceengine.cpp
+++ b/searchcore/src/vespa/searchcore/proton/persistenceengine/persistenceengine.cpp
@@ -343,7 +343,7 @@ PersistenceEngine::getBucketInfo(const Bucket& b) const
void
-PersistenceEngine::putAsync(const Bucket &bucket, Timestamp ts, storage::spi::DocumentSP doc, Context &, OperationComplete::UP onComplete)
+PersistenceEngine::putAsync(const Bucket &bucket, Timestamp ts, storage::spi::DocumentSP doc, OperationComplete::UP onComplete)
{
if (!_writeFilter.acceptWriteOperation()) {
IResourceWriteFilter::State state = _writeFilter.getAcceptState();
@@ -370,7 +370,7 @@ PersistenceEngine::putAsync(const Bucket &bucket, Timestamp ts, storage::spi::Do
}
void
-PersistenceEngine::removeAsync(const Bucket& b, std::vector<TimeStampAndDocumentId> ids, Context&, OperationComplete::UP onComplete)
+PersistenceEngine::removeAsync(const Bucket& b, std::vector<TimeStampAndDocumentId> ids, OperationComplete::UP onComplete)
{
if (ids.size() == 1) {
removeAsyncSingle(b, ids[0].first, ids[0].second, std::move(onComplete));
@@ -429,7 +429,7 @@ PersistenceEngine::removeAsyncSingle(const Bucket& b, Timestamp t, const Documen
void
-PersistenceEngine::updateAsync(const Bucket& b, Timestamp t, DocumentUpdate::SP upd, Context&, OperationComplete::UP onComplete)
+PersistenceEngine::updateAsync(const Bucket& b, Timestamp t, DocumentUpdate::SP upd, OperationComplete::UP onComplete)
{
if (!_writeFilter.acceptWriteOperation()) {
IResourceWriteFilter::State state = _writeFilter.getAcceptState();
@@ -534,7 +534,7 @@ PersistenceEngine::createIterator(const Bucket &bucket, FieldSetSP fields, const
PersistenceEngine::IterateResult
-PersistenceEngine::iterate(IteratorId id, uint64_t maxByteSize, Context&) const
+PersistenceEngine::iterate(IteratorId id, uint64_t maxByteSize) const
{
ReadGuard rguard(_rwMutex);
IteratorEntry *iteratorEntry;
@@ -568,7 +568,7 @@ PersistenceEngine::iterate(IteratorId id, uint64_t maxByteSize, Context&) const
Result
-PersistenceEngine::destroyIterator(IteratorId id, Context&)
+PersistenceEngine::destroyIterator(IteratorId id)
{
ReadGuard rguard(_rwMutex);
std::lock_guard<std::mutex> guard(_iterators_lock);
@@ -586,7 +586,7 @@ PersistenceEngine::destroyIterator(IteratorId id, Context&)
void
-PersistenceEngine::createBucketAsync(const Bucket &b, Context &, OperationComplete::UP onComplete) noexcept
+PersistenceEngine::createBucketAsync(const Bucket &b, OperationComplete::UP onComplete) noexcept
{
ReadGuard rguard(_rwMutex);
LOG(spam, "createBucket(%s)", b.toString().c_str());
@@ -606,7 +606,7 @@ PersistenceEngine::createBucketAsync(const Bucket &b, Context &, OperationComple
void
-PersistenceEngine::deleteBucketAsync(const Bucket& b, Context&, OperationComplete::UP onComplete) noexcept
+PersistenceEngine::deleteBucketAsync(const Bucket& b, OperationComplete::UP onComplete) noexcept
{
ReadGuard rguard(_rwMutex);
LOG(spam, "deleteBucket(%s)", b.toString().c_str());
@@ -651,7 +651,7 @@ PersistenceEngine::getModifiedBuckets(BucketSpace bucketSpace) const
Result
-PersistenceEngine::split(const Bucket& source, const Bucket& target1, const Bucket& target2, Context&)
+PersistenceEngine::split(const Bucket& source, const Bucket& target1, const Bucket& target2)
{
ReadGuard rguard(_rwMutex);
LOG(spam, "split(%s, %s, %s)", source.toString().c_str(), target1.toString().c_str(), target2.toString().c_str());
@@ -669,7 +669,7 @@ PersistenceEngine::split(const Bucket& source, const Bucket& target1, const Buck
Result
-PersistenceEngine::join(const Bucket& source1, const Bucket& source2, const Bucket& target, Context&)
+PersistenceEngine::join(const Bucket& source1, const Bucket& source2, const Bucket& target)
{
ReadGuard rguard(_rwMutex);
LOG(spam, "join(%s, %s, %s)", source1.toString().c_str(), source2.toString().c_str(), target.toString().c_str());
@@ -694,7 +694,6 @@ PersistenceEngine::register_resource_usage_listener(IResourceUsageListener& list
void
PersistenceEngine::destroyIterators()
{
- Context context(storage::spi::Priority(0x80), 0);
for (;;) {
IteratorId id;
{
@@ -703,7 +702,7 @@ PersistenceEngine::destroyIterators()
break;
id = _iterators.begin()->first;
}
- Result res(destroyIterator(id, context));
+ Result res(destroyIterator(id));
if (res.hasError()) {
LOG(debug, "%zu iterator left. Can not destroy iterator '%" PRIu64 "'. Reason='%s'", _iterators.size(), id.getValue(), res.toString().c_str());
std::this_thread::sleep_for(100ms);
diff --git a/searchcore/src/vespa/searchcore/proton/persistenceengine/persistenceengine.h b/searchcore/src/vespa/searchcore/proton/persistenceengine/persistenceengine.h
index 7c8040fae9d..a8886e19def 100644
--- a/searchcore/src/vespa/searchcore/proton/persistenceengine/persistenceengine.h
+++ b/searchcore/src/vespa/searchcore/proton/persistenceengine/persistenceengine.h
@@ -106,20 +106,20 @@ public:
Result setClusterState(BucketSpace bucketSpace, const ClusterState& calc) override;
void setActiveStateAsync(const Bucket&, BucketInfo::ActiveState, OperationComplete::UP) override;
BucketInfoResult getBucketInfo(const Bucket&) const override;
- void putAsync(const Bucket &, Timestamp, storage::spi::DocumentSP, Context &context, OperationComplete::UP) override;
- void removeAsync(const Bucket&, std::vector<TimeStampAndDocumentId> ids, Context&, OperationComplete::UP) override;
- void updateAsync(const Bucket&, Timestamp, storage::spi::DocumentUpdateSP, Context&, OperationComplete::UP) override;
+ void putAsync(const Bucket &, Timestamp, storage::spi::DocumentSP, OperationComplete::UP) override;
+ void removeAsync(const Bucket&, std::vector<TimeStampAndDocumentId> ids, OperationComplete::UP) override;
+ void updateAsync(const Bucket&, Timestamp, storage::spi::DocumentUpdateSP, OperationComplete::UP) override;
GetResult get(const Bucket&, const document::FieldSet&, const document::DocumentId&, Context&) const override;
CreateIteratorResult
createIterator(const Bucket &bucket, FieldSetSP, const Selection &, IncludedVersions, Context &context) override;
- IterateResult iterate(IteratorId, uint64_t maxByteSize, Context&) const override;
- Result destroyIterator(IteratorId, Context&) override;
+ IterateResult iterate(IteratorId, uint64_t maxByteSize) const override;
+ Result destroyIterator(IteratorId) override;
- void createBucketAsync(const Bucket &bucketId, Context &, OperationComplete::UP) noexcept override;
- void deleteBucketAsync(const Bucket&, Context&, OperationComplete::UP) noexcept override;
+ void createBucketAsync(const Bucket &bucketId, OperationComplete::UP) noexcept override;
+ void deleteBucketAsync(const Bucket&, OperationComplete::UP) noexcept override;
BucketIdListResult getModifiedBuckets(BucketSpace bucketSpace) const override;
- Result split(const Bucket& source, const Bucket& target1, const Bucket& target2, Context&) override;
- Result join(const Bucket& source1, const Bucket& source2, const Bucket& target, Context&) override;
+ Result split(const Bucket& source, const Bucket& target1, const Bucket& target2) override;
+ Result join(const Bucket& source1, const Bucket& source2, const Bucket& target) override;
std::unique_ptr<vespalib::IDestructorCallback> register_resource_usage_listener(IResourceUsageListener& listener) override;
std::unique_ptr<vespalib::IDestructorCallback> register_executor(std::shared_ptr<BucketExecutor>) override;
void destroyIterators();