diff options
author | Henning Baldersheim <balder@yahoo-inc.com> | 2020-11-24 05:45:54 +0000 |
---|---|---|
committer | Henning Baldersheim <balder@yahoo-inc.com> | 2020-11-24 13:05:21 +0000 |
commit | 1545c0208a32e96db1bf17237b465757e7eab617 (patch) | |
tree | 5104d2d02273eabf8e8a6d6615daa0f127b393db /searchcore | |
parent | 6534a31a780aec69a208c11ac50d43a3621773a0 (diff) |
Remove load type from Context
Diffstat (limited to 'searchcore')
3 files changed, 21 insertions, 43 deletions
diff --git a/searchcore/src/apps/vespa-feed-bm/spi_bm_feed_handler.cpp b/searchcore/src/apps/vespa-feed-bm/spi_bm_feed_handler.cpp index 24439a10925..daebc8a7a47 100644 --- a/searchcore/src/apps/vespa-feed-bm/spi_bm_feed_handler.cpp +++ b/searchcore/src/apps/vespa-feed-bm/spi_bm_feed_handler.cpp @@ -6,7 +6,6 @@ #include <vespa/document/fieldset/fieldsetrepo.h> #include <vespa/document/fieldvalue/document.h> #include <vespa/document/update/documentupdate.h> -#include <vespa/metrics/loadtype.h> #include <vespa/persistence/spi/persistenceprovider.h> using document::Document; @@ -20,8 +19,7 @@ namespace feedbm { namespace { -storage::spi::LoadType default_load_type(0, "default"); -storage::spi::Context context(default_load_type, storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); +storage::spi::Context context(storage::spi::Priority(0), 0); void get_bucket_info_loop(PendingTracker &tracker) { @@ -38,7 +36,7 @@ class MyOperationComplete : public storage::spi::OperationComplete PendingTracker& _tracker; public: MyOperationComplete(std::atomic<uint32_t> &errors, const Bucket& bucket, PendingTracker& tracker); - ~MyOperationComplete(); + ~MyOperationComplete() override; void onComplete(std::unique_ptr<storage::spi::Result> result) override; void addResultHandler(const storage::spi::ResultHandler* resultHandler) override; }; diff --git a/searchcore/src/tests/proton/persistenceengine/persistenceengine_test.cpp b/searchcore/src/tests/proton/persistenceengine/persistenceengine_test.cpp index 2b7bacfaec0..95a41a255ce 100644 --- a/searchcore/src/tests/proton/persistenceengine/persistenceengine_test.cpp +++ b/searchcore/src/tests/proton/persistenceengine/persistenceengine_test.cpp @@ -453,8 +453,7 @@ TEST_F("require that getPartitionStates() prepares all handlers", SimpleFixture) TEST_F("require that puts are routed to handler", SimpleFixture) { - storage::spi::LoadType loadType(0, "default"); - Context context(loadType, storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); + Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); f.engine.put(bucket1, tstamp1, doc1, context); TEST_DO(assertHandler(bucket1, tstamp1, docId1, f.hset.handler1)); TEST_DO(assertHandler(bucket0, tstamp0, docId0, f.hset.handler2)); @@ -473,8 +472,7 @@ TEST_F("require that put is rejected if resource limit is reached", SimpleFixtur f._writeFilter._acceptWriteOperation = false; f._writeFilter._message = "Disk is full"; - storage::spi::LoadType loadType(0, "default"); - Context context(loadType, storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); + Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); EXPECT_EQUAL( Result(Result::ErrorType::RESOURCE_EXHAUSTED, "Put operation rejected for document 'id:type3:type3::1': 'Disk is full'"), @@ -484,8 +482,7 @@ TEST_F("require that put is rejected if resource limit is reached", SimpleFixtur TEST_F("require that updates are routed to handler", SimpleFixture) { - storage::spi::LoadType loadType(0, "default"); - Context context(loadType, storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); + Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); f.hset.handler1.setExistingTimestamp(tstamp2); UpdateResult ur = f.engine.update(bucket1, tstamp1, upd1, context); TEST_DO(assertHandler(bucket1, tstamp1, docId1, f.hset.handler1)); @@ -504,8 +501,7 @@ TEST_F("require that updates are routed to handler", SimpleFixture) TEST_F("require that updates with bad ids are rejected", SimpleFixture) { - storage::spi::LoadType loadType(0, "default"); - Context context(loadType, storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); + Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); EXPECT_EQUAL(UpdateResult(Result::ErrorType::PERMANENT_ERROR, "Update operation rejected due to bad id (id:type2:type2::1, type1)"), f.engine.update(bucket1, tstamp1, bad_id_upd, context)); @@ -516,8 +512,7 @@ TEST_F("require that update is rejected if resource limit is reached", SimpleFix f._writeFilter._acceptWriteOperation = false; f._writeFilter._message = "Disk is full"; - storage::spi::LoadType loadType(0, "default"); - Context context(loadType, storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); + Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); EXPECT_EQUAL( Result(Result::ErrorType::RESOURCE_EXHAUSTED, @@ -528,8 +523,7 @@ TEST_F("require that update is rejected if resource limit is reached", SimpleFix TEST_F("require that removes are routed to handlers", SimpleFixture) { - storage::spi::LoadType loadType(0, "default"); - Context context(loadType, storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); + Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); RemoveResult rr = f.engine.remove(bucket1, tstamp1, docId3, context); TEST_DO(assertHandler(bucket0, tstamp0, docId0, f.hset.handler1)); TEST_DO(assertHandler(bucket0, tstamp0, docId0, f.hset.handler2)); @@ -565,8 +559,7 @@ TEST_F("require that remove is NOT rejected if resource limit is reached", Simpl f._writeFilter._acceptWriteOperation = false; f._writeFilter._message = "Disk is full"; - storage::spi::LoadType loadType(0, "default"); - Context context(loadType, storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); + Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); EXPECT_EQUAL(RemoveResult(false), f.engine.remove(bucket1, tstamp1, docId1, context)); } @@ -618,8 +611,7 @@ TEST_F("require that getBucketInfo() is routed to handlers and merged", SimpleFi TEST_F("require that createBucket() is routed to handlers and merged", SimpleFixture) { - storage::spi::LoadType loadType(0, "default"); - Context context(loadType, storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); + Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); f.hset.handler1._createBucketResult = Result(Result::ErrorType::TRANSIENT_ERROR, "err1a"); f.hset.handler2._createBucketResult = Result(Result::ErrorType::PERMANENT_ERROR, "err2a"); @@ -631,8 +623,7 @@ TEST_F("require that createBucket() is routed to handlers and merged", SimpleFix TEST_F("require that deleteBucket() is routed to handlers and merged", SimpleFixture) { - storage::spi::LoadType loadType(0, "default"); - Context context(loadType, storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); + Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); f.hset.handler1.deleteBucketResult = Result(Result::ErrorType::TRANSIENT_ERROR, "err1"); f.hset.handler2.deleteBucketResult = Result(Result::ErrorType::PERMANENT_ERROR, "err2"); @@ -650,8 +641,7 @@ TEST_F("require that getModifiedBuckets() is routed to handlers and merged", Sim TEST_F("require that get is sent to all handlers", SimpleFixture) { - storage::spi::LoadType loadType(0, "default"); - Context context(loadType, storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); + Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); GetResult result = f.engine.get(bucket1, document::AllFields(), docId1, context); EXPECT_EQUAL(docId1, f.hset.handler1.lastDocId); @@ -661,8 +651,7 @@ TEST_F("require that get is sent to all handlers", SimpleFixture) { TEST_F("require that get freezes the bucket", SimpleFixture) { EXPECT_FALSE(f.hset.handler1.wasFrozen(bucket1)); EXPECT_FALSE(f.hset.handler2.wasFrozen(bucket1)); - storage::spi::LoadType loadType(0, "default"); - Context context(loadType, storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); + Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); f.engine.get(bucket1, document::AllFields(), docId1, context); EXPECT_TRUE(f.hset.handler1.wasFrozen(bucket1)); EXPECT_TRUE(f.hset.handler2.wasFrozen(bucket1)); @@ -673,8 +662,7 @@ TEST_F("require that get freezes the bucket", SimpleFixture) { TEST_F("require that get returns the first document found", SimpleFixture) { f.hset.handler1.setDocument(*doc1, tstamp1); f.hset.handler2.setDocument(*doc2, tstamp2); - storage::spi::LoadType loadType(0, "default"); - Context context(loadType, storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); + Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); GetResult result = f.engine.get(bucket1, document::AllFields(), docId1, context); EXPECT_EQUAL(docId1, f.hset.handler1.lastDocId); @@ -687,8 +675,7 @@ TEST_F("require that get returns the first document found", SimpleFixture) { } TEST_F("require that createIterator does", SimpleFixture) { - storage::spi::LoadType loadType(0, "default"); - Context context(loadType, storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); + Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); CreateIteratorResult result = f.engine.createIterator(bucket1, std::make_shared<document::AllFields>(), selection, storage::spi::NEWEST_DOCUMENT_ONLY, context); @@ -701,8 +688,7 @@ TEST_F("require that createIterator does", SimpleFixture) { } TEST_F("require that iterator ids are unique", SimpleFixture) { - storage::spi::LoadType loadType(0, "default"); - Context context(loadType, storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); + Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); CreateIteratorResult result = f.engine.createIterator(bucket1, std::make_shared<document::AllFields>(), selection, storage::spi::NEWEST_DOCUMENT_ONLY, context); @@ -716,8 +702,7 @@ TEST_F("require that iterator ids are unique", SimpleFixture) { TEST_F("require that iterate requires valid iterator", SimpleFixture) { uint64_t max_size = 1024; - storage::spi::LoadType loadType(0, "default"); - Context context(loadType, storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); + Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); IterateResult it_result = f.engine.iterate(IteratorId(1), max_size, context); EXPECT_TRUE(it_result.hasError()); EXPECT_EQUAL(Result::ErrorType::PERMANENT_ERROR, it_result.getErrorCode()); @@ -736,8 +721,7 @@ TEST_F("require that iterate returns documents", SimpleFixture) { f.hset.handler1.setDocument(*doc1, tstamp1); f.hset.handler2.setDocument(*doc2, tstamp2); - storage::spi::LoadType loadType(0, "default"); - Context context(loadType, storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); + Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); uint64_t max_size = 1024; CreateIteratorResult result = f.engine.createIterator(bucket1, std::make_shared<document::AllFields>(), selection, @@ -752,8 +736,7 @@ TEST_F("require that iterate returns documents", SimpleFixture) { TEST_F("require that destroyIterator prevents iteration", SimpleFixture) { f.hset.handler1.setDocument(*doc1, tstamp1); - storage::spi::LoadType loadType(0, "default"); - Context context(loadType, storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); + Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); CreateIteratorResult create_result = f.engine.createIterator(bucket1, std::make_shared<document::AllFields>(), selection, storage::spi::NEWEST_DOCUMENT_ONLY, context); @@ -773,8 +756,7 @@ TEST_F("require that destroyIterator prevents iteration", SimpleFixture) { TEST_F("require that buckets are frozen during iterator life", SimpleFixture) { EXPECT_FALSE(f.hset.handler1.isFrozen(bucket1)); EXPECT_FALSE(f.hset.handler2.isFrozen(bucket1)); - storage::spi::LoadType loadType(0, "default"); - Context context(loadType, storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); + Context context(storage::spi::Priority(0), storage::spi::Trace::TraceLevel(0)); CreateIteratorResult create_result = f.engine.createIterator(bucket1, std::make_shared<document::AllFields>(), selection, storage::spi::NEWEST_DOCUMENT_ONLY, context); diff --git a/searchcore/src/vespa/searchcore/proton/persistenceengine/persistenceengine.cpp b/searchcore/src/vespa/searchcore/proton/persistenceengine/persistenceengine.cpp index 0eec650cfdb..02efd1ef11f 100644 --- a/searchcore/src/vespa/searchcore/proton/persistenceengine/persistenceengine.cpp +++ b/searchcore/src/vespa/searchcore/proton/persistenceengine/persistenceengine.cpp @@ -611,9 +611,7 @@ PersistenceEngine::join(const Bucket& source1, const Bucket& source2, const Buck void PersistenceEngine::destroyIterators() { - Context context(storage::spi::LoadType(0, "default"), - storage::spi::Priority(0x80), - storage::spi::Trace::TraceLevel(0)); + Context context(storage::spi::Priority(0x80), 0); for (;;) { IteratorId id; { |