Skip to content

Commit

Permalink
cosmetics
Browse files Browse the repository at this point in the history
  • Loading branch information
ssavenko-near committed Mar 4, 2025
1 parent 23c5ac8 commit 95146fd
Show file tree
Hide file tree
Showing 4 changed files with 40 additions and 44 deletions.
3 changes: 1 addition & 2 deletions chain/client/src/client.rs
Original file line number Diff line number Diff line change
Expand Up @@ -2279,8 +2279,7 @@ impl Client {
// Transactions only need to be recorded if the node is a validator.
if me.is_some() {
let mut pool = self.chunk_producer.sharded_tx_pool.lock().unwrap();
match pool.insert_transaction(shard_uid, tx.clone())
{
match pool.insert_transaction(shard_uid, tx.clone()) {
InsertTransactionResult::Success => {
trace!(target: "client", ?shard_uid, tx_hash = ?tx.get_hash(), "Recorded a transaction.");
}
Expand Down
16 changes: 7 additions & 9 deletions chain/client/src/tx_request_handler.rs
Original file line number Diff line number Diff line change
Expand Up @@ -178,12 +178,9 @@ impl TxRequestHandler {
cur_block.block_congestion_info().get(&receiver_shard).copied();
let protocol_version = self.epoch_manager.get_epoch_protocol_version(&epoch_id)?;

if let Err(err) = self.runtime.validate_tx(
&shard_layout,
tx,
protocol_version,
receiver_congestion_info,
) {
if let Err(err) =
self.runtime.validate_tx(&shard_layout, tx, protocol_version, receiver_congestion_info)
{
tracing::debug!(target: "client", tx_hash = ?tx.get_hash(), ?err, "Invalid tx during basic validation");
return Ok(ProcessTxResponse::InvalidTx(err));
}
Expand All @@ -206,7 +203,9 @@ impl TxRequestHandler {
// Not being able to fetch a state root most likely implies that we haven't
// caught up with the next epoch yet.
if is_forwarded {
return Err(near_client_primitives::types::Error::Other("Node has not caught up yet".to_string()));
return Err(near_client_primitives::types::Error::Other(
"Node has not caught up yet".to_string(),
));
} else {
self.forward_tx(&epoch_id, tx, signer)?;
return Ok(ProcessTxResponse::RequestRouted);
Expand All @@ -229,8 +228,7 @@ impl TxRequestHandler {
// Transactions only need to be recorded if the node is a validator.
if me.is_some() {
let mut pool = self.tx_pool.lock().unwrap();
match pool.insert_transaction(shard_uid, tx.clone())
{
match pool.insert_transaction(shard_uid, tx.clone()) {
InsertTransactionResult::Success => {
tracing::trace!(target: "client", ?shard_uid, tx_hash = ?tx.get_hash(), "Recorded a transaction.");
}
Expand Down
2 changes: 0 additions & 2 deletions chain/jsonrpc/src/lib.rs
Original file line number Diff line number Diff line change
Expand Up @@ -233,9 +233,7 @@ pub struct ClientSenderForRpc(
AsyncSender<DebugStatus, ActixResult<DebugStatus>>,
AsyncSender<GetClientConfig, ActixResult<GetClientConfig>>,
AsyncSender<GetNetworkInfo, ActixResult<GetNetworkInfo>>,
// AsyncSender<ProcessTxRequest, ActixResult<ProcessTxRequest>>,
AsyncSender<Status, ActixResult<Status>>,
// Sender<ProcessTxRequest>,
#[cfg(feature = "test_features")] Sender<near_client::NetworkAdversarialMessage>,
#[cfg(feature = "test_features")]
AsyncSender<
Expand Down
63 changes: 32 additions & 31 deletions integration-tests/src/env/setup.rs
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@ use near_chain_configs::{
ChunkDistributionNetworkConfig, ClientConfig, MutableConfigValue, ReshardingConfig,
};
use near_chunks::adapter::ShardsManagerRequestFromClient;
use near_chunks::client::{ShardedTransactionPool, ShardsManagerResponse};
use near_chunks::client::ShardsManagerResponse;
use near_chunks::shards_manager_actor::{ShardsManagerActor, start_shards_manager};
use near_chunks::test_utils::SynchronousShardsManagerAdapter;
use near_client::adversarial::Controls;
Expand Down Expand Up @@ -158,20 +158,7 @@ pub fn setup(

let mut rng_seed: RngSeed = [0; 32];
rand::thread_rng().fill(&mut rng_seed);
let dummy_tx_pool =
Arc::new(std::sync::Mutex::new(ShardedTransactionPool::new(rng_seed, Some(60000))));

let tx_processor_addr = spawn_tx_request_handler_actor(
clock.clone(),
config.clone(),
dummy_tx_pool.clone(),
epoch_manager.clone(),
shard_tracker.clone(),
signer.clone(),
runtime.clone(),
chain_genesis.clone(),
network_adapter.clone(),
);
let client_adapter_for_partial_witness_actor = LateBoundSender::new();
let (partial_witness_addr, _) = spawn_actix_actor(PartialWitnessActor::new(
clock.clone(),
Expand All @@ -190,28 +177,42 @@ pub fn setup(
let resharding_sender = resharding_sender_addr.with_auto_span_context();

let shards_manager_adapter_for_client = LateBoundSender::new();
let StartClientResult { client_actor, .. } = start_client(
clock,
let StartClientResult { client_actor, client_arbiter_handle: _, resharding_handle: _, tx_pool } =
start_client(
clock.clone(),
config.clone(),
chain_genesis.clone(),
epoch_manager.clone(),
shard_tracker.clone(),
runtime.clone(),
PeerId::new(PublicKey::empty(KeyType::ED25519)),
Arc::new(ActixFutureSpawner),
network_adapter.clone(),
shards_manager_adapter_for_client.as_sender(),
signer.clone(),
telemetry.with_auto_span_context().into_sender(),
None,
None,
adv,
None,
partial_witness_adapter.clone().into_multi_sender(),
enable_doomslug,
Some(TEST_SEED),
resharding_sender.into_multi_sender(),
);

let tx_processor_addr = spawn_tx_request_handler_actor(
clock.clone(),
config.clone(),
chain_genesis,
tx_pool.clone(),
epoch_manager.clone(),
shard_tracker.clone(),
runtime,
PeerId::new(PublicKey::empty(KeyType::ED25519)),
Arc::new(ActixFutureSpawner),
signer.clone(),
runtime.clone(),
chain_genesis.clone(),
network_adapter.clone(),
shards_manager_adapter_for_client.as_sender(),
signer,
telemetry.with_auto_span_context().into_sender(),
None,
None,
adv,
None,
partial_witness_adapter.clone().into_multi_sender(),
enable_doomslug,
Some(TEST_SEED),
resharding_sender.into_multi_sender(),
);

let validator_signer = Some(Arc::new(EmptyValidatorSigner::new(account_id)));
let (shards_manager_addr, _) = start_shards_manager(
epoch_manager.clone(),
Expand Down

0 comments on commit 95146fd

Please sign in to comment.