diff --git a/src/StreamController.cpp b/src/StreamController.cpp index b1c73583deaa6ba660b74ad2e52768737c4f4f99..a4bb15fdb27394b60234edbb38b21643c7e5be4e 100644 --- a/src/StreamController.cpp +++ b/src/StreamController.cpp @@ -25,9 +25,7 @@ StreamController::StreamController( Registrar->getNewRegistrar("stream.writer")), StreamerOptions(Settings), MetaDataTracker(std::move(Tracker)), _metadata_enquirer(std::move(metadata_enquirer)), - _consumer_factory(std::move(consumer_factory)) { - Logger::Error("streamcontroller constructed"); -} + _consumer_factory(std::move(consumer_factory)) {} StreamController::~StreamController() { Logger::Error("streamcontroller destructor start"); @@ -40,14 +38,12 @@ StreamController::~StreamController() { } void StreamController::start() { - Logger::Error("streamcontroller start start"); MdatWriter->set_start_time(StreamerOptions.StartTimestamp); MdatWriter->set_stop_time(StreamerOptions.StopTimestamp); Executor.sendLowPriorityWork([=]() { CurrentMetadataTimeOut = StreamerOptions.BrokerSettings.MinMetadataTimeout; getTopicNames(); }); - Logger::Error("streamcontroller start end"); } void StreamController::setStopTime(time_point const &StopTime) { @@ -99,7 +95,6 @@ bool StreamController::isDoneWriting() { std::string StreamController::getJobId() const { return WriterTask->jobID(); } void StreamController::getTopicNames() { - Logger::Error("streamcontroller gettopicnames start"); try { auto TopicNames = _metadata_enquirer->getTopicList( StreamerOptions.BrokerSettings.Address, CurrentMetadataTimeOut, @@ -118,11 +113,9 @@ void StreamController::getTopicNames() { .count()); Executor.sendLowPriorityWork([=]() { getTopicNames(); }); } - Logger::Error("streamcontroller gettopicnames end"); } void StreamController::initStreams(std::set<std::string> known_topic_names) { - Logger::Error("streamcontroller initstreams start"); std::map<std::string, Stream::SrcToDst> topic_src_map; std::string errors_collector; for (auto &src : WriterTask->sources()) { @@ -164,7 +157,6 @@ void StreamController::initStreams(std::set<std::string> known_topic_names) { Streamers.emplace_back(std::move(topic)); } Executor.sendLowPriorityWork([=]() { performPeriodicChecks(); }); - Logger::Error("streamcontroller initstreams end"); } bool StreamController::hasErrorState() const { return HasError; } @@ -175,16 +167,13 @@ std::string StreamController::errorMessage() { } void StreamController::performPeriodicChecks() { - Logger::Error("streamcontroller performperiodicchecks start"); checkIfStreamsAreDone(); throttleIfWriteQueueIsFull(); std::this_thread::sleep_for(PeriodicChecksInterval); Executor.sendLowPriorityWork([=]() { performPeriodicChecks(); }); - Logger::Error("streamcontroller performperiodicchecks end"); } void StreamController::checkIfStreamsAreDone() { - Logger::Error("streamcontroller checkifstreamsaredone start"); try { Streamers.erase( std::remove_if(Streamers.begin(), Streamers.end(), @@ -201,11 +190,9 @@ void StreamController::checkIfStreamsAreDone() { stop(); StreamersRemaining.store(false); } - Logger::Error("streamcontroller checkifstreamsaredone end"); } void StreamController::throttleIfWriteQueueIsFull() { - Logger::Error("streamcontroller throttle start"); auto QueuedWrites = WriterThread.nrOfWritesQueued(); if (QueuedWrites > StreamerOptions.MaxQueuedWrites && !StreamersPaused.load()) { @@ -220,7 +207,6 @@ void StreamController::throttleIfWriteQueueIsFull() { QueuedWrites); resumeStreamers(); } - Logger::Error("streamcontroller throttle end"); } } // namespace FileWriter diff --git a/src/WriterModule/f144/f144_Writer.cpp b/src/WriterModule/f144/f144_Writer.cpp index f56d7fec6beabeb91d953e60cc3d78b67e0f4e57..27ac0592ab948f082e195d47af98e4f1d6bc141c 100644 --- a/src/WriterModule/f144/f144_Writer.cpp +++ b/src/WriterModule/f144/f144_Writer.cpp @@ -30,15 +30,12 @@ struct ValuesInformation { }; template <typename Type> void makeIt(hdf5::node::Group const &Parent) { - Logger::Error("f144_Writer makeIt start"); NeXusDataset::ExtensibleDataset<Type>( // NOLINT(bugprone-unused-raii) Parent, "value", NeXusDataset::Mode::Create); // NOLINT(bugprone-unused-raii) - Logger::Error("f144_Writer makeIt end"); } void initValueDataset(hdf5::node::Group const &Parent, Type ElementType) { - Logger::Error("f144_Writer initValueDataset start"); using OpenFuncType = std::function<void()>; std::map<Type, OpenFuncType> CreateValuesMap{ {Type::int8, [&]() { makeIt<std::int8_t>(Parent); }}, @@ -53,12 +50,10 @@ void initValueDataset(hdf5::node::Group const &Parent, Type ElementType) { {Type::float64, [&]() { makeIt<std::double_t>(Parent); }}, }; CreateValuesMap.at(ElementType)(); - Logger::Error("f144_Writer initValueDataset end"); } /// Parse the configuration for this stream. void f144_Writer::config_post_processing() { - Logger::Error("f144_Writer config_post_processing start"); auto ToLower = [](auto InString) { std::transform(InString.begin(), InString.end(), InString.begin(), [](auto C) { return std::tolower(C); }); @@ -80,14 +75,12 @@ void f144_Writer::config_post_processing() { Logger::Info(R"(Unknown data type with name "{}". Using double.)", DataType.get_value()); } - Logger::Error("f144_Writer config_post_processing end"); } /// \brief Implement the writer module interface, forward to the CREATE case /// of /// `init_hdf`. InitResult f144_Writer::init_hdf(hdf5::node::Group &HDFGroup) { - Logger::Error("f144_Writer init_hdf start"); auto Create = NeXusDataset::Mode::Create; try { NeXusDataset::Time(HDFGroup, Create, @@ -104,18 +97,15 @@ InitResult f144_Writer::init_hdf(hdf5::node::Group &HDFGroup) { auto message = hdf5::error::print_nested(E); Logger::Error("f144 writer could not init_hdf hdf_parent: {} trace: {}", static_cast<std::string>(HDFGroup.link().path()), message); - Logger::Error("f144_Writer init_hdf end"); return InitResult::ERROR; } - Logger::Error("f144_Writer init_hdf end"); return InitResult::OK; } /// \brief Implement the writer module interface, forward to the OPEN case of /// `init_hdf`. InitResult f144_Writer::reopen(hdf5::node::Group &HDFGroup) { - Logger::Error("f144_Writer reopen start"); auto Open = NeXusDataset::Mode::Open; try { Timestamp = NeXusDataset::Time(HDFGroup, Open); @@ -126,33 +116,26 @@ InitResult f144_Writer::reopen(hdf5::node::Group &HDFGroup) { Logger::Error( R"(Failed to reopen datasets in HDF file with error message: "{}")", std::string(E.what())); - Logger::Error("f144_Writer reopen end"); return InitResult::ERROR; } - Logger::Error("f144_Writer reopen end"); return InitResult::OK; } template <typename FBValueType, typename ReturnType> ReturnType extractScalarValue(const f144_LogData *LogDataMessage) { - Logger::Error("f144_Writer extractScalarValue start"); auto LogValue = LogDataMessage->value_as<FBValueType>(); - Logger::Error("f144_Writer extractScalarValue end"); return LogValue->value(); } template <typename DataType, typename ValueType, class DatasetType> ValuesInformation appendScalarData(DatasetType &Dataset, const f144_LogData *LogDataMessage) { - Logger::Error("f144_Writer appendScalarData start"); auto ScalarValue = extractScalarValue<ValueType, DataType>(LogDataMessage); Dataset.template appendElement<DataType>(ScalarValue); - Logger::Error("f144_Writer appendScalarData end"); return {double(ScalarValue), double(ScalarValue), double(ScalarValue), 1}; } void msgTypeIsConfigType(f144_Writer::Type ConfigType, Value MsgType) { - Logger::Error("f144_Writer msgTypeIsConfigType start"); std::unordered_map<Value, f144_Writer::Type> TypeComparison{ {Value::ArrayByte, f144_Writer::Type::int8}, {Value::Byte, f144_Writer::Type::int8}, @@ -209,12 +192,10 @@ void msgTypeIsConfigType(f144_Writer::Type ConfigType, Value MsgType) { } catch (std::out_of_range const &) { Logger::Error("Got out of range error when comparing types."); } - Logger::Error("f144_Writer msgTypeIsConfigType end"); } bool f144_Writer::writeImpl(FlatbufferMessage const &Message, [[maybe_unused]] bool is_buffered_message) { - Logger::Error("f144_Writer writeImpl start"); auto LogDataMessage = Getf144_LogData(Message.data()); Timestamp.appendElement(LogDataMessage->timestamp()); auto Type = LogDataMessage->value_type(); @@ -293,13 +274,11 @@ bool f144_Writer::writeImpl(FlatbufferMessage const &Message, MetaDataMax.setValue(Max); MetaDataMean.setValue(Sum / TotalNrOfElementsWritten); } - Logger::Error("f144_Writer writeImpl end"); return true; } void f144_Writer::register_meta_data(hdf5::node::Group const &HDFGroup, const MetaData::TrackerPtr &Tracker) { - Logger::Error("f144_Writer register_meta_data start"); if (MetaData.get_value()) { MetaDataMin = MetaData::Value<double>( HDFGroup, "minimum_value", MetaData::basicDatasetWriter<double>, @@ -319,7 +298,6 @@ void f144_Writer::register_meta_data(hdf5::node::Group const &HDFGroup, MetaDataMean.setAttribute("units", Unit.get_value()); Tracker->registerMetaData(MetaDataMean); } - Logger::Error("f144_Writer register_meta_data end"); } /// Register the writer module.