diff --git a/components/raftstore/src/store/fsm/peer.rs b/components/raftstore/src/store/fsm/peer.rs index 27359ae397f..7c33bf66b87 100644 --- a/components/raftstore/src/store/fsm/peer.rs +++ b/components/raftstore/src/store/fsm/peer.rs @@ -2547,15 +2547,12 @@ where return Ok(()); } - debug!("!!!!!! handle 3"); - if msg.get_is_tombstone() { // we receive a message tells us to remove ourself. self.handle_gc_peer_msg(&msg); return Ok(()); } - debug!("!!!!!! handle 4"); if msg.has_merge_target() { fail_point!("on_has_merge_target", |_| Ok(())); if self.need_gc_merge(&msg)? { @@ -2564,12 +2561,10 @@ where return Ok(()); } - debug!("!!!!!! handle 5"); if self.check_msg(&msg) { return Ok(()); } - debug!("!!!!!! handle 6"); if msg.has_extra_msg() { self.on_extra_message(msg); return Ok(()); @@ -2577,7 +2572,6 @@ where let is_snapshot = msg.get_message().has_snapshot(); - debug!("!!!!!! handle 7"); // TODO: spin off the I/O code (delete_snapshot) let regions_to_destroy = match self.check_snapshot(&msg)? { Either::Left(key) => { @@ -2594,7 +2588,6 @@ where Either::Right(v) => v, }; - debug!("!!!!!! handle 8"); if util::is_vote_msg(msg.get_message()) || msg_type == MessageType::MsgTimeoutNow { if self.fsm.hibernate_state.group_state() != GroupState::Chaos { self.fsm.reset_hibernate_state(GroupState::Chaos); @@ -2607,12 +2600,10 @@ where let from_peer_id = msg.get_from_peer().get_id(); self.fsm.peer.insert_peer_cache(msg.take_from_peer()); - debug!("!!!!!! handle 8.1"); let result = if msg_type == MessageType::MsgTransferLeader { self.on_transfer_leader_msg(msg.get_message(), peer_disk_usage); Ok(()) } else { - debug!("!!!!!! handle 8.2"); // This can be a message that sent when it's still a follower. Nevertheleast, // it's meaningless to continue to handle the request as callbacks are cleared. if msg.get_message().get_msg_type() == MessageType::MsgReadIndex @@ -2620,22 +2611,14 @@ where && (msg.get_message().get_from() == raft::INVALID_ID || msg.get_message().get_from() == self.fsm.peer_id()) { - debug!( - "!!!!!! handle 8.2.2 {} {} {}", - self.fsm.peer.is_leader(), - msg.get_message().get_from(), - msg.get_message().get_from() - ); self.ctx.raft_metrics.message_dropped.stale_msg.inc(); return Ok(()); } - debug!("!!!!!! handle 8.3"); self.fsm.peer.step(self.ctx, msg.take_message()) }; stepped.set(result.is_ok()); - debug!("!!!!!! handle 9"); if is_snapshot { if !self.fsm.peer.has_pending_snapshot() { // This snapshot is rejected by raft-rs. diff --git a/components/raftstore/src/store/peer.rs b/components/raftstore/src/store/peer.rs index 2b2bfaab944..c0277cbc5a6 100644 --- a/components/raftstore/src/store/peer.rs +++ b/components/raftstore/src/store/peer.rs @@ -1777,7 +1777,6 @@ where ctx: &mut PollContext, mut m: eraftpb::Message, ) -> Result<()> { - info!("!!!!!! raft step {:?}", m); fail_point!( "step_message_3_1", self.peer.get_store_id() == 3 && self.region_id == 1, diff --git a/proxy_components/proxy_ffi/src/read_index_helper.rs b/proxy_components/proxy_ffi/src/read_index_helper.rs index 471a912f635..02b2666c67d 100644 --- a/proxy_components/proxy_ffi/src/read_index_helper.rs +++ b/proxy_components/proxy_ffi/src/read_index_helper.rs @@ -91,7 +91,6 @@ pub fn gen_read_index_raft_cmd_req(req: &mut ReadIndexRequest) -> RaftCmdRequest inner_req.set_cmd_type(CmdType::ReadIndex); inner_req.mut_read_index().set_start_ts(req.get_start_ts()); if !req.get_ranges().is_empty() { - tikv_util::info!("!!!!!! not empty"); let r = &mut req.mut_ranges()[0]; let mut range = kvproto::kvrpcpb::KeyRange::default(); range.set_start_key(r.take_start_key()); diff --git a/proxy_tests/proxy/shared/replica_read.rs b/proxy_tests/proxy/shared/replica_read.rs index e636bc9365f..ee75bc0e0e4 100644 --- a/proxy_tests/proxy/shared/replica_read.rs +++ b/proxy_tests/proxy/shared/replica_read.rs @@ -483,7 +483,6 @@ fn test_raft_cmd_request_learner_advanve_max_ts() { let region = cluster.get_region(b""); assert_eq!(region_id, 1); assert_eq!(region.get_id(), 1); - info!("!!!!! region {:?}", region); let leader = region.get_peers()[0].clone(); fail::cfg("on_pre_write_apply_state", "return(true)").unwrap(); @@ -552,7 +551,6 @@ fn test_raft_cmd_request_learner_advanve_max_ts() { } let resp = block_on(result_rx).unwrap(); - info!("!!!!!! XZZZDD {:?}", resp); (resp.get_responses()[0].get_read_index().clone(), start_ts) }; diff --git a/src/server/raftkv/mod.rs b/src/server/raftkv/mod.rs index 8a44fb7aa2b..9f42925b6d4 100644 --- a/src/server/raftkv/mod.rs +++ b/src/server/raftkv/mod.rs @@ -779,7 +779,6 @@ pub struct ReplicaReadLockChecker { impl ReplicaReadLockChecker { pub fn new(concurrency_manager: ConcurrencyManager) -> Self { - info!("!!!!!! ReplicaReadLockChecker::new"); ReplicaReadLockChecker { concurrency_manager, } @@ -798,10 +797,6 @@ impl ReadIndexObserver for ReplicaReadLockChecker { // Only check and return result if the current peer is a leader. // If it's not a leader, the read index request will be redirected to the leader // later. - info!( - "!!!!!! ReplicaReadLockChecker::on_step {:?} {:?}", - msg, role - ); if msg.get_msg_type() != MessageType::MsgReadIndex || role != StateRole::Leader { return; } @@ -813,7 +808,6 @@ impl ReadIndexObserver for ReplicaReadLockChecker { let start_ts = request.get_start_ts().into(); self.concurrency_manager.update_max_ts(start_ts); for range in request.mut_key_ranges().iter_mut() { - info!("!!!!!! ReplicaReadLockChecker::range {:?}", range); let key_bound = |key: Vec| { if key.is_empty() { None