diff --git a/src/v/cluster/tx_gateway_frontend.cc b/src/v/cluster/tx_gateway_frontend.cc index de36c3de517e..d6e49d9997c0 100644 --- a/src/v/cluster/tx_gateway_frontend.cc +++ b/src/v/cluster/tx_gateway_frontend.cc @@ -1035,6 +1035,7 @@ ss::future tx_gateway_frontend::init_tm_tx( txlog.trace, "[tx_id: {}] waiting for {} topic to apper in metadata cache, " "retries left: {}", + tx_id, model::tx_manager_nt, retries); if (_metadata_cache.local().contains(model::tx_manager_nt)) { diff --git a/src/v/raft/consensus.cc b/src/v/raft/consensus.cc index 442b2b6a05ef..b20508d1bb03 100644 --- a/src/v/raft/consensus.cc +++ b/src/v/raft/consensus.cc @@ -983,7 +983,8 @@ consensus::update_group_member(model::broker broker) { if (!cfg.contains_broker(broker.id())) { vlog( _ctxlog.warn, - "Node with id {} does not exists in current configuration"); + "Node with id {} does not exists in current configuration", + broker.id()); return ss::make_ready_future( errc::node_does_not_exists); } diff --git a/src/v/storage/offset_translator_state.cc b/src/v/storage/offset_translator_state.cc index 6171114833c1..e5093f441894 100644 --- a/src/v/storage/offset_translator_state.cc +++ b/src/v/storage/offset_translator_state.cc @@ -115,6 +115,7 @@ model::offset offset_translator_state::to_log_offset( vassert( interval_end_it != _last_offset2batch.begin(), "ntp {}: log offset search start too small: {}", + _ntp, search_start); auto delta = std::prev(interval_end_it)->second.next_delta; diff --git a/src/v/storage/segment_appender.cc b/src/v/storage/segment_appender.cc index 616b1096049a..0713f7606645 100644 --- a/src/v/storage/segment_appender.cc +++ b/src/v/storage/segment_appender.cc @@ -297,6 +297,7 @@ ss::future<> segment_appender::truncate(size_t n) { n <= file_byte_offset(), "Cannot ask to truncate at:{} which is more bytes than we have:{} - {}", file_byte_offset(), + n, *this); return hard_flush() .then([this, n] { return do_truncation(n); }) @@ -663,7 +664,10 @@ bool segment_appender::inflight_write::try_merge( // writes should form a contiguous series of writes and we only check // the last write for merging. vassert( - committed_offset == pco, "in try_merge writes didn't touch: {} {}"); + committed_offset == pco, + "in try_merge writes didn't touch: {} {}", + committed_offset, + pco); // the lhs write cannot be full since then how could the next write // share its chunk: it must use a new chunk