// Copyright 2017-2020 Parity Technologies (UK) Ltd. // This file is part of Polkadot. // Polkadot is free software: you can redistribute it and/or modify // it under the terms of the GNU General Public License as published by // the Free Software Foundation, either version 3 of the License, or // (at your option) any later version. // Polkadot is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU General Public License for more details. // You should have received a copy of the GNU General Public License // along with Polkadot. If not, see . //! Polkadot service. Specialized wrapper over substrate service. pub mod chain_spec; mod grandpa_support; mod client; use grandpa::{self, FinalityProofProvider as GrandpaFinalityProofProvider}; use log::info; use polkadot_node_core_proposer::ProposerFactory; use polkadot_overseer::{AllSubsystems, BlockInfo, Overseer, OverseerHandler}; use polkadot_subsystem::DummySubsystem; use prometheus_endpoint::Registry; use sc_client_api::ExecutorProvider; use sc_executor::native_executor_instance; use service::{error::Error as ServiceError, RpcHandlers}; use sp_blockchain::HeaderBackend; use sp_core::traits::SpawnNamed; use sp_trie::PrefixedMemoryDB; use std::sync::Arc; use std::time::Duration; pub use self::client::{AbstractClient, Client, ClientHandle, ExecuteWithClient, RuntimeApiCollection}; pub use chain_spec::{PolkadotChainSpec, KusamaChainSpec, WestendChainSpec, RococoChainSpec}; #[cfg(feature = "full-node")] pub use codec::Codec; pub use consensus_common::{Proposal, SelectChain, BlockImport, RecordProof, block_validation::Chain}; pub use polkadot_parachain::wasm_executor::run_worker as run_validation_worker; pub use polkadot_primitives::v1::{Block, BlockId, CollatorId, Id as ParaId}; pub use sc_client_api::{Backend, ExecutionStrategy, CallExecutor}; pub use sc_consensus::LongestChain; pub use sc_executor::NativeExecutionDispatch; pub use service::{ Role, PruningMode, TransactionPoolOptions, Error, RuntimeGenesis, TFullClient, TLightClient, TFullBackend, TLightBackend, TFullCallExecutor, TLightCallExecutor, Configuration, ChainSpec, TaskManager, }; pub use service::config::{DatabaseConfig, PrometheusConfig}; pub use sp_api::{ApiRef, Core as CoreApi, ConstructRuntimeApi, ProvideRuntimeApi, StateBackend}; pub use sp_runtime::traits::{DigestFor, HashFor, NumberFor, Block as BlockT, self as runtime_traits, BlakeTwo256}; pub use kusama_runtime; pub use polkadot_runtime; pub use rococo_runtime; pub use westend_runtime; native_executor_instance!( pub PolkadotExecutor, polkadot_runtime::api::dispatch, polkadot_runtime::native_version, frame_benchmarking::benchmarking::HostFunctions, ); native_executor_instance!( pub KusamaExecutor, kusama_runtime::api::dispatch, kusama_runtime::native_version, frame_benchmarking::benchmarking::HostFunctions, ); native_executor_instance!( pub WestendExecutor, westend_runtime::api::dispatch, westend_runtime::native_version, frame_benchmarking::benchmarking::HostFunctions, ); native_executor_instance!( pub RococoExecutor, rococo_runtime::api::dispatch, rococo_runtime::native_version, frame_benchmarking::benchmarking::HostFunctions, ); /// Can be called for a `Configuration` to check if it is a configuration for the `Kusama` network. pub trait IdentifyVariant { /// Returns if this is a configuration for the `Kusama` network. fn is_kusama(&self) -> bool; /// Returns if this is a configuration for the `Westend` network. fn is_westend(&self) -> bool; /// Returns if this is a configuration for the `Rococo` network. fn is_rococo(&self) -> bool; } impl IdentifyVariant for Box { fn is_kusama(&self) -> bool { self.id().starts_with("kusama") || self.id().starts_with("ksm") } fn is_westend(&self) -> bool { self.id().starts_with("westend") || self.id().starts_with("wnd") } fn is_rococo(&self) -> bool { self.id().starts_with("rococo") || self.id().starts_with("rco") } } // If we're using prometheus, use a registry with a prefix of `polkadot`. fn set_prometheus_registry(config: &mut Configuration) -> Result<(), ServiceError> { if let Some(PrometheusConfig { registry, .. }) = config.prometheus_config.as_mut() { *registry = Registry::new_custom(Some("polkadot".into()), None)?; } Ok(()) } pub type FullBackend = service::TFullBackend; type FullSelectChain = sc_consensus::LongestChain; pub type FullClient = service::TFullClient; type FullGrandpaBlockImport = grandpa::GrandpaBlockImport< FullBackend, Block, FullClient, FullSelectChain >; type LightBackend = service::TLightBackendWithHash; type LightClient = service::TLightClientWithBackend; #[cfg(feature = "full-node")] fn new_partial(config: &mut Configuration) -> Result< service::PartialComponents< FullClient, FullBackend, FullSelectChain, consensus_common::DefaultImportQueue>, sc_transaction_pool::FullPool>, ( impl Fn( polkadot_rpc::DenyUnsafe, polkadot_rpc::SubscriptionTaskExecutor, ) -> polkadot_rpc::RpcExtension, ( babe::BabeBlockImport< Block, FullClient, FullGrandpaBlockImport >, grandpa::LinkHalf, FullSelectChain>, babe::BabeLink ), ( grandpa::SharedVoterState, Arc>, ), ) >, Error > where RuntimeApi: ConstructRuntimeApi> + Send + Sync + 'static, RuntimeApi::RuntimeApi: RuntimeApiCollection>, Executor: NativeExecutionDispatch + 'static, { set_prometheus_registry(config)?; let inherent_data_providers = inherents::InherentDataProviders::new(); let (client, backend, keystore, task_manager) = service::new_full_parts::(&config)?; let client = Arc::new(client); let select_chain = sc_consensus::LongestChain::new(backend.clone()); let transaction_pool = sc_transaction_pool::BasicPool::new_full( config.transaction_pool.clone(), config.prometheus_registry(), task_manager.spawn_handle(), client.clone(), ); let grandpa_hard_forks = if config.chain_spec.is_kusama() { grandpa_support::kusama_hard_forks() } else { Vec::new() }; let (grandpa_block_import, grandpa_link) = grandpa::block_import_with_authority_set_hard_forks( client.clone(), &(client.clone() as Arc<_>), select_chain.clone(), grandpa_hard_forks, )?; let justification_import = grandpa_block_import.clone(); let (block_import, babe_link) = babe::block_import( babe::Config::get_or_compute(&*client)?, grandpa_block_import, client.clone(), )?; let import_queue = babe::import_queue( babe_link.clone(), block_import.clone(), Some(Box::new(justification_import)), None, client.clone(), select_chain.clone(), inherent_data_providers.clone(), &task_manager.spawn_handle(), config.prometheus_registry(), consensus_common::CanAuthorWithNativeVersion::new(client.executor().clone()), )?; let justification_stream = grandpa_link.justification_stream(); let shared_authority_set = grandpa_link.shared_authority_set().clone(); let shared_voter_state = grandpa::SharedVoterState::empty(); let finality_proof_provider = GrandpaFinalityProofProvider::new_for_service(backend.clone(), client.clone()); let import_setup = (block_import.clone(), grandpa_link, babe_link.clone()); let rpc_setup = (shared_voter_state.clone(), finality_proof_provider.clone()); let babe_config = babe_link.config().clone(); let shared_epoch_changes = babe_link.epoch_changes().clone(); let rpc_extensions_builder = { let client = client.clone(); let keystore = keystore.clone(); let transaction_pool = transaction_pool.clone(); let select_chain = select_chain.clone(); move |deny_unsafe, subscription_executor| -> polkadot_rpc::RpcExtension { let deps = polkadot_rpc::FullDeps { client: client.clone(), pool: transaction_pool.clone(), select_chain: select_chain.clone(), deny_unsafe, babe: polkadot_rpc::BabeDeps { babe_config: babe_config.clone(), shared_epoch_changes: shared_epoch_changes.clone(), keystore: keystore.clone(), }, grandpa: polkadot_rpc::GrandpaDeps { shared_voter_state: shared_voter_state.clone(), shared_authority_set: shared_authority_set.clone(), justification_stream: justification_stream.clone(), subscription_executor, finality_provider: finality_proof_provider.clone(), }, }; polkadot_rpc::create_full(deps) } }; Ok(service::PartialComponents { client, backend, task_manager, keystore, select_chain, import_queue, transaction_pool, inherent_data_providers, other: (rpc_extensions_builder, import_setup, rpc_setup) }) } fn real_overseer( leaves: impl IntoIterator, prometheus_registry: Option<&Registry>, s: S, ) -> Result<(Overseer, OverseerHandler), ServiceError> { let all_subsystems = AllSubsystems { candidate_validation: DummySubsystem, candidate_backing: DummySubsystem, candidate_selection: DummySubsystem, statement_distribution: DummySubsystem, availability_distribution: DummySubsystem, bitfield_signing: DummySubsystem, bitfield_distribution: DummySubsystem, provisioner: DummySubsystem, pov_distribution: DummySubsystem, runtime_api: DummySubsystem, availability_store: DummySubsystem, network_bridge: DummySubsystem, chain_api: DummySubsystem, collation_generation: DummySubsystem, collator_protocol: DummySubsystem, }; Overseer::new( leaves, all_subsystems, prometheus_registry, s, ).map_err(|e| ServiceError::Other(format!("Failed to create an Overseer: {:?}", e))) } #[cfg(feature = "full-node")] pub struct NewFull { pub task_manager: TaskManager, pub client: C, pub node_handles: OverseerHandler, pub network: Arc::Hash>>, pub network_status_sinks: service::NetworkStatusSinks, pub rpc_handlers: RpcHandlers, } #[cfg(feature = "full-node")] impl NewFull { fn with_client(self, func: impl FnOnce(C) -> Client) -> NewFull { NewFull { client: func(self.client), task_manager: self.task_manager, node_handles: self.node_handles, network: self.network, network_status_sinks: self.network_status_sinks, rpc_handlers: self.rpc_handlers, } } } /// Create a new full node of arbitrary runtime and executor. /// /// This is an advanced feature and not recommended for general use. Generally, `build_full` is /// a better choice. #[cfg(feature = "full-node")] pub fn new_full( mut config: Configuration, collating_for: Option<(CollatorId, ParaId)>, authority_discovery_enabled: bool, grandpa_pause: Option<(u32, u32)>, ) -> Result>>, Error> where RuntimeApi: ConstructRuntimeApi> + Send + Sync + 'static, RuntimeApi::RuntimeApi: RuntimeApiCollection>, Executor: NativeExecutionDispatch + 'static, { use sp_core::traits::BareCryptoStorePtr; let is_collator = collating_for.is_some(); let role = config.role.clone(); let is_authority = role.is_authority() && !is_collator; let force_authoring = config.force_authoring; let disable_grandpa = config.disable_grandpa; let name = config.network.node_name.clone(); let service::PartialComponents { client, backend, mut task_manager, keystore, select_chain, import_queue, transaction_pool, inherent_data_providers, other: (rpc_extensions_builder, import_setup, rpc_setup) } = new_partial::(&mut config)?; let prometheus_registry = config.prometheus_registry().cloned(); let (shared_voter_state, finality_proof_provider) = rpc_setup; let (network, network_status_sinks, system_rpc_tx, network_starter) = service::build_network(service::BuildNetworkParams { config: &config, client: client.clone(), transaction_pool: transaction_pool.clone(), spawn_handle: task_manager.spawn_handle(), import_queue, on_demand: None, block_announce_validator_builder: None, finality_proof_request_builder: None, finality_proof_provider: Some(finality_proof_provider.clone()), })?; if config.offchain_worker.enabled { service::build_offchain_workers( &config, backend.clone(), task_manager.spawn_handle(), client.clone(), network.clone(), ); } let telemetry_connection_sinks = service::TelemetryConnectionSinks::default(); let rpc_handlers = service::spawn_tasks(service::SpawnTasksParams { config, backend: backend.clone(), client: client.clone(), keystore: keystore.clone(), network: network.clone(), rpc_extensions_builder: Box::new(rpc_extensions_builder), transaction_pool: transaction_pool.clone(), task_manager: &mut task_manager, on_demand: None, remote_blockchain: None, telemetry_connection_sinks: telemetry_connection_sinks.clone(), network_status_sinks: network_status_sinks.clone(), system_rpc_tx, })?; let (block_import, link_half, babe_link) = import_setup; let overseer_client = client.clone(); let spawner = task_manager.spawn_handle(); let leaves: Vec<_> = select_chain.clone() .leaves() .unwrap_or_else(|_| vec![]) .into_iter() .filter_map(|hash| { let number = client.number(hash).ok()??; let parent_hash = client.header(&BlockId::Hash(hash)).ok()??.parent_hash; Some(BlockInfo { hash, parent_hash, number, }) }) .collect(); let (overseer, handler) = real_overseer(leaves, prometheus_registry.as_ref(), spawner)?; let handler_clone = handler.clone(); task_manager.spawn_essential_handle().spawn_blocking("overseer", Box::pin(async move { use futures::{pin_mut, select, FutureExt}; let forward = polkadot_overseer::forward_events(overseer_client, handler_clone); let forward = forward.fuse(); let overseer_fut = overseer.run().fuse(); pin_mut!(overseer_fut); pin_mut!(forward); loop { select! { _ = forward => break, _ = overseer_fut => break, complete => break, } } })); if role.is_authority() { let can_author_with = consensus_common::CanAuthorWithNativeVersion::new(client.executor().clone()); let proposer = ProposerFactory::new( client.clone(), transaction_pool, handler.clone(), ); let babe_config = babe::BabeParams { keystore: keystore.clone(), client: client.clone(), select_chain, block_import, env: proposer, sync_oracle: network.clone(), inherent_data_providers: inherent_data_providers.clone(), force_authoring, babe_link, can_author_with, }; let babe = babe::start_babe(babe_config)?; task_manager.spawn_essential_handle().spawn_blocking("babe", babe); } // if the node isn't actively participating in consensus then it doesn't // need a keystore, regardless of which protocol we use below. let keystore_opt = if is_authority { Some(keystore.clone() as BareCryptoStorePtr) } else { None }; let config = grandpa::Config { // FIXME substrate#1578 make this available through chainspec gossip_duration: Duration::from_millis(1000), justification_period: 512, name: Some(name), observer_enabled: false, keystore: keystore_opt, is_authority: role.is_network_authority(), }; let enable_grandpa = !disable_grandpa; if enable_grandpa { // start the full GRANDPA voter // NOTE: unlike in substrate we are currently running the full // GRANDPA voter protocol for all full nodes (regardless of whether // they're validators or not). at this point the full voter should // provide better guarantees of block and vote data availability than // the observer. // add a custom voting rule to temporarily stop voting for new blocks // after the given pause block is finalized and restarting after the // given delay. let voting_rule = match grandpa_pause { Some((block, delay)) => { info!("GRANDPA scheduled voting pause set for block #{} with a duration of {} blocks.", block, delay, ); grandpa::VotingRulesBuilder::default() .add(grandpa_support::PauseAfterBlockFor(block, delay)) .build() } None => grandpa::VotingRulesBuilder::default().build(), }; let grandpa_config = grandpa::GrandpaParams { config, link: link_half, network: network.clone(), inherent_data_providers: inherent_data_providers.clone(), telemetry_on_connect: Some(telemetry_connection_sinks.on_connect_stream()), voting_rule, prometheus_registry: prometheus_registry.clone(), shared_voter_state, }; task_manager.spawn_essential_handle().spawn_blocking( "grandpa-voter", grandpa::run_grandpa_voter(grandpa_config)? ); } else { grandpa::setup_disabled_grandpa( client.clone(), &inherent_data_providers, network.clone(), )?; } if matches!(role, Role::Authority{..} | Role::Sentry{..}) { use sc_network::Event; use futures::StreamExt; if authority_discovery_enabled { let (sentries, authority_discovery_role) = match role { Role::Authority { ref sentry_nodes } => ( sentry_nodes.clone(), authority_discovery::Role::Authority ( keystore.clone(), ), ), Role::Sentry {..} => ( vec![], authority_discovery::Role::Sentry, ), _ => unreachable!("Due to outer matches! constraint; qed."), }; let network_event_stream = network.event_stream("authority-discovery"); let dht_event_stream = network_event_stream.filter_map(|e| async move { match e { Event::Dht(e) => Some(e), _ => None, }}).boxed(); let (authority_discovery_worker, _service) = authority_discovery::new_worker_and_service( client.clone(), network.clone(), sentries, dht_event_stream, authority_discovery_role, prometheus_registry.clone(), ); task_manager.spawn_handle().spawn("authority-discovery-worker", authority_discovery_worker); } } network_starter.start_network(); Ok(NewFull { task_manager, client, node_handles: handler, network, network_status_sinks, rpc_handlers, }) } /// Builds a new service for a light client. fn new_light(mut config: Configuration) -> Result<(TaskManager, RpcHandlers), Error> where Runtime: 'static + Send + Sync + ConstructRuntimeApi>, >>::RuntimeApi: RuntimeApiCollection>, Dispatch: NativeExecutionDispatch + 'static, { set_prometheus_registry(&mut config)?; use sc_client_api::backend::RemoteBackend; let (client, backend, keystore, mut task_manager, on_demand) = service::new_light_parts::(&config)?; let select_chain = sc_consensus::LongestChain::new(backend.clone()); let transaction_pool = Arc::new(sc_transaction_pool::BasicPool::new_light( config.transaction_pool.clone(), config.prometheus_registry(), task_manager.spawn_handle(), client.clone(), on_demand.clone(), )); let grandpa_block_import = grandpa::light_block_import( client.clone(), backend.clone(), &(client.clone() as Arc<_>), Arc::new(on_demand.checker().clone()), )?; let finality_proof_import = grandpa_block_import.clone(); let finality_proof_request_builder = finality_proof_import.create_finality_proof_request_builder(); let (babe_block_import, babe_link) = babe::block_import( babe::Config::get_or_compute(&*client)?, grandpa_block_import, client.clone(), )?; let inherent_data_providers = inherents::InherentDataProviders::new(); // FIXME: pruning task isn't started since light client doesn't do `AuthoritySetup`. let import_queue = babe::import_queue( babe_link, babe_block_import, None, Some(Box::new(finality_proof_import)), client.clone(), select_chain.clone(), inherent_data_providers.clone(), &task_manager.spawn_handle(), config.prometheus_registry(), consensus_common::NeverCanAuthor, )?; let finality_proof_provider = GrandpaFinalityProofProvider::new_for_service(backend.clone(), client.clone()); let (network, network_status_sinks, system_rpc_tx, network_starter) = service::build_network(service::BuildNetworkParams { config: &config, client: client.clone(), transaction_pool: transaction_pool.clone(), spawn_handle: task_manager.spawn_handle(), import_queue, on_demand: Some(on_demand.clone()), block_announce_validator_builder: None, finality_proof_request_builder: Some(finality_proof_request_builder), finality_proof_provider: Some(finality_proof_provider), })?; if config.offchain_worker.enabled { service::build_offchain_workers( &config, backend.clone(), task_manager.spawn_handle(), client.clone(), network.clone(), ); } let light_deps = polkadot_rpc::LightDeps { remote_blockchain: backend.remote_blockchain(), fetcher: on_demand.clone(), client: client.clone(), pool: transaction_pool.clone(), }; let rpc_extensions = polkadot_rpc::create_light(light_deps); let rpc_handlers = service::spawn_tasks(service::SpawnTasksParams { on_demand: Some(on_demand), remote_blockchain: Some(backend.remote_blockchain()), rpc_extensions_builder: Box::new(service::NoopRpcExtensionBuilder(rpc_extensions)), task_manager: &mut task_manager, telemetry_connection_sinks: service::TelemetryConnectionSinks::default(), config, keystore, backend, transaction_pool, client, network, network_status_sinks, system_rpc_tx, })?; network_starter.start_network(); Ok((task_manager, rpc_handlers)) } /// Builds a new object suitable for chain operations. #[cfg(feature = "full-node")] pub fn new_chain_ops(mut config: &mut Configuration) -> Result< ( Arc, Arc, consensus_common::import_queue::BasicQueue>, TaskManager, ), ServiceError > { config.keystore = service::config::KeystoreConfig::InMemory; if config.chain_spec.is_rococo() { let service::PartialComponents { client, backend, import_queue, task_manager, .. } = new_partial::(config)?; Ok((Arc::new(Client::Rococo(client)), backend, import_queue, task_manager)) } else if config.chain_spec.is_kusama() { let service::PartialComponents { client, backend, import_queue, task_manager, .. } = new_partial::(config)?; Ok((Arc::new(Client::Kusama(client)), backend, import_queue, task_manager)) } else if config.chain_spec.is_westend() { let service::PartialComponents { client, backend, import_queue, task_manager, .. } = new_partial::(config)?; Ok((Arc::new(Client::Westend(client)), backend, import_queue, task_manager)) } else { let service::PartialComponents { client, backend, import_queue, task_manager, .. } = new_partial::(config)?; Ok((Arc::new(Client::Polkadot(client)), backend, import_queue, task_manager)) } } /// Build a new light node. pub fn build_light(config: Configuration) -> Result<(TaskManager, RpcHandlers), ServiceError> { if config.chain_spec.is_rococo() { new_light::(config) } else if config.chain_spec.is_kusama() { new_light::(config) } else if config.chain_spec.is_westend() { new_light::(config) } else { new_light::(config) } } #[cfg(feature = "full-node")] pub fn build_full( config: Configuration, collating_for: Option<(CollatorId, ParaId)>, authority_discovery_enabled: bool, grandpa_pause: Option<(u32, u32)>, ) -> Result, ServiceError> { if config.chain_spec.is_rococo() { new_full::( config, collating_for, authority_discovery_enabled, grandpa_pause, ).map(|full| full.with_client(Client::Rococo)) } else if config.chain_spec.is_kusama() { new_full::( config, collating_for, authority_discovery_enabled, grandpa_pause, ).map(|full| full.with_client(Client::Kusama)) } else if config.chain_spec.is_westend() { new_full::( config, collating_for, authority_discovery_enabled, grandpa_pause, ).map(|full| full.with_client(Client::Westend)) } else { new_full::( config, collating_for, authority_discovery_enabled, grandpa_pause, ).map(|full| full.with_client(Client::Polkadot)) } }