mirror of
https://github.com/hoytech/strfry.git
synced 2025-06-20 09:50:28 +00:00
177 lines
5.5 KiB
C++
177 lines
5.5 KiB
C++
#pragma once
|
|
|
|
#include <hoytech/protected_queue.h>
|
|
|
|
#include "golpe.h"
|
|
|
|
#include "events.h"
|
|
|
|
|
|
struct WriterPipelineInput {
|
|
tao::json::value eventJson;
|
|
EventSourceType sourceType;
|
|
std::string sourceInfo;
|
|
};
|
|
|
|
|
|
struct WriterPipeline {
|
|
public:
|
|
uint64_t debounceDelayMilliseconds = 1'000;
|
|
uint64_t writeBatchSize = 1'000;
|
|
|
|
private:
|
|
hoytech::protected_queue<WriterPipelineInput> validatorInbox;
|
|
hoytech::protected_queue<EventToWrite> writerInbox;
|
|
hoytech::protected_queue<bool> flushInbox;
|
|
std::thread validatorThread;
|
|
std::thread writerThread;
|
|
|
|
std::condition_variable shutdownCv;
|
|
std::mutex shutdownMutex;
|
|
std::atomic<bool> shutdownRequested = false;
|
|
std::atomic<bool> shutdownComplete = false;
|
|
|
|
std::atomic<uint64_t> numLive = 0;
|
|
std::condition_variable backpressureCv;
|
|
std::mutex backpressureMutex;
|
|
|
|
public:
|
|
WriterPipeline() {
|
|
validatorThread = std::thread([&]() {
|
|
setThreadName("Validator");
|
|
|
|
secp256k1_context *secpCtx = secp256k1_context_create(SECP256K1_CONTEXT_VERIFY);
|
|
|
|
while (1) {
|
|
auto msgs = validatorInbox.pop_all();
|
|
|
|
for (auto &m : msgs) {
|
|
if (m.eventJson.is_null()) {
|
|
shutdownRequested = true;
|
|
writerInbox.push_move({});
|
|
shutdownCv.notify_all();
|
|
return;
|
|
}
|
|
|
|
std::string flatStr;
|
|
std::string jsonStr;
|
|
|
|
try {
|
|
parseAndVerifyEvent(m.eventJson, secpCtx, true, true, flatStr, jsonStr);
|
|
} catch (std::exception &e) {
|
|
LW << "Rejected event: " << m.eventJson << " reason: " << e.what();
|
|
numLive--;
|
|
continue;
|
|
}
|
|
|
|
writerInbox.push_move({ std::move(flatStr), std::move(jsonStr), hoytech::curr_time_us(), m.sourceType, std::move(m.sourceInfo) });
|
|
}
|
|
}
|
|
});
|
|
|
|
writerThread = std::thread([&]() {
|
|
setThreadName("Writer");
|
|
|
|
while (1) {
|
|
// Debounce
|
|
|
|
{
|
|
auto numPendingElems = writerInbox.wait();
|
|
|
|
if (!shutdownRequested && numPendingElems < writeBatchSize) {
|
|
std::unique_lock<std::mutex> lk(shutdownMutex);
|
|
shutdownCv.wait_for(lk, std::chrono::milliseconds(debounceDelayMilliseconds), [&]{return !!shutdownRequested;});
|
|
}
|
|
}
|
|
|
|
auto newEvents = writerInbox.pop_all();
|
|
|
|
uint64_t written = 0, dups = 0;
|
|
|
|
// Collect a certain amount of records in a batch, push the rest back into the writerInbox
|
|
// Pre-filter out dups in a read-only txn as an optimisation
|
|
|
|
std::vector<EventToWrite> newEventsToProc;
|
|
|
|
{
|
|
auto txn = env.txn_ro();
|
|
|
|
while (newEvents.size()) {
|
|
if (newEventsToProc.size() >= writeBatchSize) {
|
|
// Put the rest back in the writerInbox
|
|
writerInbox.unshift_move_all(newEvents);
|
|
newEvents.clear();
|
|
break;
|
|
}
|
|
|
|
auto event = std::move(newEvents.front());
|
|
newEvents.pop_front();
|
|
|
|
if (event.flatStr.size() == 0) {
|
|
shutdownComplete = true;
|
|
break;
|
|
}
|
|
|
|
numLive--;
|
|
|
|
auto *flat = flatStrToFlatEvent(event.flatStr);
|
|
if (lookupEventById(txn, sv(flat->id()))) {
|
|
dups++;
|
|
continue;
|
|
}
|
|
|
|
newEventsToProc.emplace_back(std::move(event));
|
|
}
|
|
}
|
|
|
|
if (newEventsToProc.size()) {
|
|
{
|
|
auto txn = env.txn_rw();
|
|
writeEvents(txn, newEventsToProc);
|
|
txn.commit();
|
|
}
|
|
|
|
for (auto &ev : newEventsToProc) {
|
|
if (ev.status == EventWriteStatus::Written) written++;
|
|
else dups++;
|
|
// FIXME: log rejected stats too
|
|
}
|
|
}
|
|
|
|
if (written || dups) LI << "Writer: added: " << written << " dups: " << dups;
|
|
|
|
if (shutdownComplete) {
|
|
flushInbox.push_move(true);
|
|
if (numLive != 0) LW << "numLive was not 0 after shutdown!";
|
|
return;
|
|
}
|
|
|
|
backpressureCv.notify_all();
|
|
}
|
|
});
|
|
}
|
|
|
|
~WriterPipeline() {
|
|
flush();
|
|
validatorThread.join();
|
|
writerThread.join();
|
|
}
|
|
|
|
void write(WriterPipelineInput &&inp) {
|
|
numLive++;
|
|
validatorInbox.push_move(std::move(inp));
|
|
}
|
|
|
|
void flush() {
|
|
validatorInbox.push_move({ tao::json::null, EventSourceType::None, "" });
|
|
flushInbox.wait();
|
|
}
|
|
|
|
void wait() {
|
|
uint64_t drainUntil = writeBatchSize * 2;
|
|
if (numLive < drainUntil) return;
|
|
std::unique_lock<std::mutex> lk(backpressureMutex);
|
|
backpressureCv.wait_for(lk, std::chrono::milliseconds(50), [&]{return numLive < drainUntil;});
|
|
}
|
|
};
|