Skip to content
  1. Apr 17, 2024
    • Oliver Tale-Yazdi's avatar
      Fix nostd build of several crates (#4060) · 7a2c9d4a
      Oliver Tale-Yazdi authored
      
      
      Preparation for https://github.com/paritytech/polkadot-sdk/pull/3935
      
      Changes:
      - Add some `default-features = false` for the case that a crate and that
      dependency both support nostd builds.
      - Shuffle files around of some benchmarking-only crates. These
      conditionally disabled the `cfg_attr` for nostd and pulled in libstd.
      Example [here](https://github.com/ggwpez/zepter/pull/95). The actual
      logic is moved into a `inner.rs` to preserve nostd capability of the
      crate in case the benchmarking feature is disabled.
      - Add some `use sp_std::vec` where needed.
      - Remove some `optional = true` in cases where it was not optional.
      - Removed one superfluous `cfg_attr(not(feature = "std"), no_std..`.
      
      All in all this should be logical no-op.
      
      ---------
      
      Signed-off-by: default avatarOliver Tale-Yazdi <[email protected]>
      7a2c9d4a
    • Alexandru Vasile's avatar
      chainHead: Report unique hashes for pruned blocks (#3667) · bfbf7f5d
      Alexandru Vasile authored
      
      
      This PR ensures that the reported pruned blocks are unique.
      
      While at it, ensure that the best block event is properly generated when
      the last best block is a fork that will be pruned in the future.
      
      To achieve this, the chainHead keeps a LRU set of reported pruned blocks
      to ensure the following are not reported twice:
      
      ```bash
      	 finalized -> block 1 -> block 2 -> block 3
      	
      	                      -> block 2 -> block 4 -> block 5
      	
      	           -> block 1 -> block 2_f -> block 6 -> block 7 -> block 8
      ```
      
      When block 7 is finalized the branch [block 2; block 3] is reported as
      pruned.
      When block 8 is finalized the branch [block 2; block 4; block 5] should
      be reported as pruned, however block 2 was already reported as pruned at
      the previous step.
      
      This is a side-effect of the pruned blocks being reported at level N -
      1. For example, if all pruned forks would be reported with the first
      encounter (when block 6 is finalized we know that block 3 and block 5
      are stale), we would not need the LRU cache.
      
      cc @paritytech/subxt-team  
      
      Closes https://github.com/paritytech/polkadot-sdk/issues/3658
      
      ---------
      
      Signed-off-by: default avatarAlexandru Vasile <[email protected]>
      Co-authored-by: default avatarSebastian Kunert <[email protected]>
      bfbf7f5d
    • thiolliere's avatar
      Improve doc for pallet macro and config macro (#4146) · ca7c01c8
      thiolliere authored
      Improve doc:
      
      * the pallet macro is actually referring to 2 places, for the module and
      for the struct placeholder but doesn't really clarify it (I should have
      named the latter just `pallet_struct` or something but it is a bit late)
      
      * The doc of `with_default` is a bit confusing too IMO.
      
      CC @Kianenigma
      
      
      
      ---------
      
      Co-authored-by: default avatarLiam Aharon <[email protected]>
      ca7c01c8
    • Muharem Ismailov's avatar
      Asset Conversion: Pool Account ID derivation with additional Pallet ID seed (#3250) · 4e10d3b0
      Muharem Ismailov authored
      Introduce `PalletId` as an additional seed parameter for pool's account
      id derivation.
      
      The PR also introduces the `pallet_asset_conversion_ops` pallet with a
      call to migrate a given pool to thew new account. Additionally
      `fungibles::lifetime::ResetTeam` and `fungible::lifetime::Refund`
      traits, to facilitate the migration of pools.
      
      ---------
      
      Co-authored-by: command-bot <>
      4e10d3b0
    • Sergej Sakac's avatar
      XCM coretime region transfers (#3455) · e6f3106d
      Sergej Sakac authored
      
      
      This PR introduces changes enabling the transfer of coretime regions via
      XCM.
      
      TL;DR: There are two primary issues that are resolved in this PR:
      
      1. The `mint` and `burn` functions were not implemented for coretime
      regions. These operations are essential for moving assets to and from
      the XCM holding register.
      2. The transfer of non-fungible assets through XCM was previously
      disallowed. This was due to incorrectly benchmarking non-fungible asset
      transfers via XCM, which led to assigning it a weight of `Weight::Max`,
      effectively preventing its execution.
      
      ### `mint_into` and `burn` implementation
      
      This PR addresses the issue with cross-chain transferring regions back
      to the Coretime chain. Remote reserve transfers are performed by
      withdrawing and depositing the asset to and from the holding registry.
      This requires the asset to support burning and minting functionality.
      
      This PR adds burning and minting; however, they work a bit differently
      than usual so that the associated region record is not lost when
      burning. Instead of removing all the data, burning will set the owner of
      the region to `None`, and when minting it back, it will set it to an
      actual value. So, when cross-chain transferring, withdrawing into the
      registry will remove the region from its original owner, and when
      depositing it from the registry, it will set its owner to another
      account
      
      This was originally implemented in this PR: #3455, however we decided to
      move all of it to this single PR
      (https://github.com/paritytech/polkadot-sdk/pull/3455#discussion_r1547324892)
      
      ### Fixes made in this PR
      
      - Update the `XcmReserveTransferFilter` on coretime chain since it is
      meant as a reserve chain for coretime regions.
      - Update the XCM benchmark to use `AssetTransactor` instead of assuming
      `pallet-balances` for fungible transfers.
      - Update the XCM benchmark to properly measure weight consumption for
      nonfungible reserve asset transfers. ATM reserve transfers via the
      extrinsic do not work since the weight for it is set to `Weight::max()`.
      
      Closes: https://github.com/paritytech/polkadot-sdk/issues/865
      
      ---------
      
      Co-authored-by: default avatarBranislav Kontur <[email protected]>
      Co-authored-by: default avatarFrancisco Aguirre <[email protected]>
      Co-authored-by: default avatarDónal Murray <[email protected]>
      e6f3106d
    • Alexandru Vasile's avatar
      grandpa: Send neighbor packet to lightclients with every finalized head (#4135) · 8fd839df
      Alexandru Vasile authored
      
      
      This PR sends the GrandpaNeighbor packet to lightclients similarly to
      the full-nodes.
      
      Previously, the lightclient would receive a GrandpaNeigbor packet only
      when the note set changed.
      
      Related to: https://github.com/paritytech/polkadot-sdk/issues/4120
      
      Next steps:
      -  [ ] check with lightclient
      
      ---------
      
      Signed-off-by: default avatarAlexandru Vasile <[email protected]>
      8fd839df
  2. Apr 16, 2024
  3. Apr 15, 2024
    • thiolliere's avatar
      pallet assets: Fix errors (#4118) · a8f4f4f0
      thiolliere authored
      `LiveAsset` is an error to be returned when an asset is not supposed to
      be live.
      And `AssetNotLive` is an error to be returned when an asset is supposed
      to be live, I don't think frozen qualifies as live.
      a8f4f4f0
    • Dónal Murray's avatar
      [pallet-broker] add tests for renewing leases (#4099) · 0c9ad530
      Dónal Murray authored
      The first test proves that parachains who were migrated over on a legacy
      lease can renew without downtime.
      
      The exception is if their lease expires in period 0 - aka within
      `region_length` timeslices after `start_sales` is called. The second
      test is designed such that it passes if the issue exists and should be
      fixed.
      This will require an intervention on Kusama to add these renewals to
      storage as it is too tight to schedule a runtime upgrade before the
      start_sales call. All leases will still have at least two full regions
      of coretime.
      0c9ad530
    • Alexandru Vasile's avatar
      logging(fix): Use the proper log target for logging (#4124) · d1f9fe0a
      Alexandru Vasile authored
      This PR ensures the proper logging target (ie `libp2p_tcp` or `beefy`)
      is displayed.
      
      The issue has been introduced in:
      https://github.com/paritytech/polkadot-sdk/pull/4059, which removes the
      normalized metadata of logs.
      
      From
      [documentation](https://docs.rs/tracing-log/latest/tracing_log/trait.NormalizeEvent.html#tymethod.normalized_metadata):
      
      > In tracing-log, an Event produced by a log (through
      [AsTrace](https://docs.rs/tracing-log/latest/tracing_log/trait.AsTrace.html))
      has an hard coded “log” target
      
      >
      [normalized_metadata](https://docs.rs/tracing-log/latest/tracing_log/trait.NormalizeEvent.html#tymethod.normalized_metadata):
      If this Event comes from a log, this method provides a new normalized
      Metadata which has all available attributes from the original log,
      including file, line, module_path and target
      
      This has low implications if a version was deployed containing the
      mentioned pull request, as we'll lose the ability to distinguish between
      log targets.
      
      ### Before this PR
      
      ```
      2024-04-15 12:45:40.327  INFO main log: Parity Polkadot
      2024-04-15 12:45:40.328  INFO main log: ️  version 1.10.0-d1b0ef76
      2024-04-15 12:45:40.328  INFO main log: ️  by Parity Technologies <[email protected]>, 2017-2024
      2024-04-15 12:45:40.328  INFO main log: 📋 Chain specification: Development
      2024-04-15 12:45:40.328  INFO main log: 🏷  Node name: yellow-eyes-2963
      2024-04-15 12:45:40.328  INFO main log: 👤 Role: AUTHORITY
      2024-04-15 12:45:40.328  INFO main log: 💾 Database: RocksDb at /tmp/substrated39i9J/chains/rococo_dev/db/full
      2024-04-15 12:45:44.508  WARN main log: Took active validators from set with wrong size
      ...
      
      2024-04-15 12:45:45.805  INFO                 main log: 👶 Starting BABE Authorship worker
      2024-04-15 12:45:45.806  INFO tokio-runtime-worker log: 🥩 BEEFY gadget waiting for BEEFY pallet to become available...
      2024-04-15 12:45:45.806 DEBUG tokio-runtime-worker log: New listen address: /ip6/::1/tcp/30333
      2024-04-15 12:45:45.806 DEBUG tokio-runtime-worker log: New listen address: /ip4/127.0.0.1/tcp/30333
      ```
      
      ### After this PR
      
      ```
      2024-04-15 12:59:45.623  INFO main sc_cli::runner: Parity Polkadot
      2024-04-15 12:59:45.623  INFO main sc_cli::runner: ️  version 1.10.0-d1b0ef76
      2024-04-15 12:59:45.623  INFO main sc_cli::runner: ️  by Parity Technologies <[email protected]>, 2017-2024
      2024-04-15 12:59:45.623  INFO main sc_cli::runner: 📋 Chain specification: Development
      2024-04-15 12:59:45.623  INFO main sc_cli::runner: 🏷  Node name: helpless-lizards-0550
      2024-04-15 12:59:45.623  INFO main sc_cli::runner: 👤
      
       Role: AUTHORITY
      ...
      2024-04-15 12:59:50.204  INFO tokio-runtime-worker beefy: 🥩 BEEFY gadget waiting for BEEFY pallet to become available...
      2024-04-15 12:59:50.204 DEBUG tokio-runtime-worker libp2p_tcp: New listen address: /ip6/::1/tcp/30333
      2024-04-15 12:59:50.204 DEBUG tokio-runtime-worker libp2p_tcp: New listen address: /ip4/127.0.0.1/tcp/30333
      ```
      
      Signed-off-by: default avatarAlexandru Vasile <[email protected]>
      d1f9fe0a
    • Bastian Köcher's avatar
      pallet_broker: Support renewing leases expired in a previous period (#4089) · 6f73b746
      Bastian Köcher authored
      Part of: https://github.com/paritytech/polkadot-sdk/issues/4107
      6f73b746
    • Bastian Köcher's avatar
      sp-api: Use macro to detect if `frame-metadata` is enabled (#4117) · d1b0ef76
      Bastian Köcher authored
      While `sp-api-proc-macro` isn't used directly and thus, it should have
      the same features enabled as `sp-api`. However, I have seen issues
      around `frame-metadata` not being enabled for `sp-api`, but for
      `sp-api-proc-macro`. This can be prevented by using the
      `frame_metadata_enabled` macro from `sp-api` that ensures we have the
      same feature set between both crates.
      d1b0ef76
    • Alexandru Gheorghe's avatar
      Prevent accidental change of network-key for active authorities (#3852) · 2bc4ed11
      Alexandru Gheorghe authored
      
      
      As discovered during investigation of
      https://github.com/paritytech/polkadot-sdk/issues/3314 and
      https://github.com/paritytech/polkadot-sdk/issues/3673 there are active
      validators which accidentally might change their network key during
      restart, that's not a safe operation when you are in the active set
      because of distributed nature of DHT, so the old records would still
      exist in the network until they expire 36h, so unless they have a good
      reason validators should avoid changing their key when they restart
      their nodes.
      
      There is an effort in parallel to improve this situation
      https://github.com/paritytech/polkadot-sdk/pull/3786, but those changes
      are way more intrusive and will need more rigorous testing, additionally
      they will reduce the time to less than 36h, but the propagation won't be
      instant anyway, so not changing your network during restart should be
      the safest way to run your node, unless you have a really good reason to
      change it.
      
      ## Proposal
      1. Do not auto-generate the network if the network file does not exist
      in the provided path. Nodes where the key file does not exist will get
      the following error:
      ```
      Error: 
         0: Starting an authorithy without network key in /home/alexggh/.local/share/polkadot/chains/ksmcc3/network/secret_ed25519.
            
             This is not a safe operation because the old identity still lives in the dht for 36 hours.
            
             Because of it your node might suffer from not being properly connected to other nodes for validation purposes.
            
             If it is the first time running your node you could use one of the following methods.
            
             1. Pass --unsafe-force-node-key-generation and make sure you remove it for subsequent node restarts
            
             2. Separetly generate the key with: polkadot key generate-node-key --file <YOUR_PATH_TO_NODE_KEY>
      ```
      
      2. Add an explicit parameters for nodes that do want to change their
      network despite the warnings or if they run the node for the first time.
      `--unsafe-force-node-key-generation`
      
      3. For `polkadot key generate-node-key` add two new mutually exclusive
      parameters `base_path` and `default_base_path` to help with the key
      generation in the same path the polkadot main command would expect it.
       
      4. Modify the installation scripts to auto-generate a key in default
      path if one was not present already there, this should help with making
      the executable work out of the box after an instalation.
      
      ## Notes
      
      Nodes that do not have already the key persisted will fail to start
      after this change, however I do consider that better than the current
      situation where they start but they silently hide that they might not be
      properly connected to their peers.
      
      ## TODO
      - [x] Make sure only nodes that are authorities on producation chains
      will be affected by this restrictions.
      - [x] Proper PRDOC, to make sure node operators are aware this is
      coming.
      
      ---------
      
      Signed-off-by: default avatarAlexandru Gheorghe <[email protected]>
      Co-authored-by: default avatarDmitry Markin <[email protected]>
      Co-authored-by: default avatars0me0ne-unkn0wn <[email protected]>
      Co-authored-by: default avatarBastian Köcher <[email protected]>
      2bc4ed11
  4. Apr 13, 2024
  5. Apr 12, 2024
    • Vedhavyas Singareddi's avatar
      define block hash provider and default impl using frame_system (#4080) · 5b513cc0
      Vedhavyas Singareddi authored
      This PR introduces `BlockHashProvider` into `pallet_mmr::Config`
      This type is used to get `block_hash` for a given `block_number` rather
      than directly using `frame_system::Pallet::block_hash`
      
      The `DefaultBlockHashProvider` uses `frame_system::Pallet::block_hash`
      to get the `block_hash`
      
      Closes: #4062
      5b513cc0
    • Squirrel's avatar
      Remove redundent logging code (#4059) · b28ba4ae
      Squirrel authored
      
      
      1. The `CustomFmtContext::ContextWithFormatFields` enum arm isn't
      actually used and thus we don't need the enum anymore.
      
      2. We don't do anything with most of the normalized metadata that's
      created by calling `event.normalized_metadata();` - the `target` we can
      get from `event.metadata.target()` and level we can get from
      `event.metadata.level()` - let's just call them direct to simplify
      things. (`event.metadata()` is just a field call under the hood)
      
      Changelog: No functional changes, might run a tad faster with lots of
      logging turned on.
      
      ---------
      
      Co-authored-by: default avatarBastian Köcher <[email protected]>
      b28ba4ae
    • eskimor's avatar
      Improve docs of broker pallet (#3980) · a64009ae
      eskimor authored
      
      
      Small adjustments which should make understanding what is going on much
      easier for future readers.
      
      Initialization is a bit messy, the very least we should do is adding
      documentation to make it harder to use wrongly.
      
      I was thinking about calling `request_core_count` right from
      `start_sales`, but as explained in the docs, this is not necessarily
      what you want.
      
      ---------
      
      Co-authored-by: default avatareskimor <[email protected]>
      Co-authored-by: default avatarBastian Köcher <[email protected]>
      Co-authored-by: default avatarDónal Murray <[email protected]>
      a64009ae
    • PG Herveou's avatar
      033484c3
  6. Apr 11, 2024
  7. Apr 10, 2024
    • PG Herveou's avatar
      Contracts: Remove ED from base deposit (#3536) · 643aa2be
      PG Herveou authored
      
      
      - Update internal logic so that the storage_base_deposit does not
      include ED
      - add v16 migration to update ContractInfo struct with this change
      
      Before:
      <img width="820" alt="Screenshot 2024-03-21 at 11 23 29"
      src="https://github.com/paritytech/polkadot-sdk/assets/521091/a0a8df0d-e743-42c5-9e16-cf2ec1aa949c">
      
      After:
      ![Screenshot 2024-03-21 at 11 23
      42](https://github.com/paritytech/polkadot-sdk/assets/521091/593235b0-b866-4915-b653-2071d793228b)
      
      ---------
      
      Co-authored-by: default avatarCyrill Leutwiler <[email protected]>
      Co-authored-by: command-bot <>
      643aa2be
    • PG Herveou's avatar
      Contracts: Reuse module when validating (#3789) · 1da8d12d
      PG Herveou authored
      Reuse wasmi Module when validating.
      Prepare the code for 0.32 and the addition of Module::new_unchecked
      1da8d12d
    • Alexandru Vasile's avatar
      net/strategy: Log bad peerId from on_validated_block_announce (#4051) · cd010925
      Alexandru Vasile authored
      
      
      This tiny PR extends the `on_validated_block_announce` log with the bad
      PeerID.
      Used to identify if the peerID is malicious by correlating with other
      logs (ie peer-set).
      
      While at it, have removed the `\n` from a multiline log, which did not
      play well with
      [sub-triage-logs](https://github.com/lexnv/sub-triage-logs/tree/master).
      
      cc @paritytech/networking
      
      ---------
      
      Signed-off-by: default avatarAlexandru Vasile <[email protected]>
      Co-authored-by: default avatarBastian Köcher <[email protected]>
      cd010925
    • PG Herveou's avatar
      Contracts: Only exec parsed code in benchmarks (#3915) · 0d71753e
      PG Herveou authored
      [Weights
      compare](https://weights.tasty.limo/compare?unit=weight&ignore_errors=true&threshold=10&method=asymptotic&repo=polkadot-sdk&old=master&new=pg%2Fbench_tweaks&path_pattern=substrate%2Fframe%2F**%2Fsrc%2Fweights.rs%2Cpolkadot%2Fruntime%2F*%2Fsrc%2Fweights%2F**%2F*.rs%2Cpolkadot%2Fbridges%2Fmodules%2F*%2Fsrc%2Fweights.rs%2Ccumulus%2F**%2Fweights%2F*.rs%2Ccumulus%2F**%2Fweights%2Fxcm%2F*.rs%2Ccumulus%2F**%2Fsrc%2Fweights.rs)
      
      Note: Raw weights change does not mean much here, as this PR reduce the
      scope of what is benchmarked, they are therefore decreased by a good
      margin. One should instead print the Schedule using
      
      cargo test --features runtime-benchmarks bench_print_schedule --
      --nocapture
      or following the instructions from the
      [README](https://github.com/paritytech/polkadot-sdk/tree/pg/bench_tweaks/substrate/frame/contracts#schedule)
      for looking at the Schedule of a specific runtime
      
      ---------
      
      Co-authored-by: command-bot <>
      0d71753e
    • Serban Iorga's avatar
      Fix chain_spec_as_json_fails_with_invalid_config · fb84c461
      Serban Iorga authored and Bastian Köcher's avatar Bastian Köcher committed
      fb84c461
    • PG Herveou's avatar
      Update benchmarking macros (#3934) · d38f6e67
      PG Herveou authored
      
      
      Current benchmarking macro returns a closure with the captured
      benchmarked code.
      This can cause issues when the benchmarked code has complex lifetime
      requirements.
      
      This PR updates the existing macro by injecting the recording parameter
      and invoking the start / stop method around the benchmarked block
      instead of returning a closure
      
      One other added benefit is that you can write this kind of code now as
      well:
      
      ```rust
      let v;
      #[block]
      { v = func.call(); }
      dbg!(v); // or assert something on v
      ```
      
      
      [Weights compare
      link](https://weights.tasty.limo/compare?unit=weight&ignore_errors=true&threshold=10&method=asymptotic&repo=polkadot-sdk&old=pg/fix-weights&new=pg/bench_update&path_pattern=substrate/frame/**/src/weights.rs,polkadot/runtime/*/src/weights/**/*.rs,polkadot/bridges/modules/*/src/weights.rs,cumulus/**/weights/*.rs,cumulus/**/weights/xcm/*.rs,cumulus/**/src/weights.rs)
      
      ---------
      
      Co-authored-by: command-bot <>
      Co-authored-by: default avatarOliver Tale-Yazdi <oliver.tale-yazdi@pari...>
      d38f6e67
    • PG Herveou's avatar
      Contracts: Fix legacy uapi (#3994) · 2d927b07
      PG Herveou authored
      Fix some broken legacy definitions of pallet_contracts_uapi storage host
      functions
      2d927b07
  8. Apr 09, 2024
    • Sebastian Kunert's avatar
      Move cumulus zombienet tests to aura & async backing (#3568) · df818d29
      Sebastian Kunert authored
      Cumulus test-parachain node and test runtime were still using relay
      chain consensus and 12s blocktimes. With async backing around the corner
      on the major chains we should switch our tests too.
      
      Also needed to nicely test the changes coming to collators in #3168.
      
      ### Changes Overview
      - Followed the [migration
      guide](https://wiki.polkadot.network/docs/maintain-guides-async-backing)
      for async backing for the cumulus-test-runtime
      - Adjusted the cumulus-test-service to use the correct import-queue,
      lookahead collator etc.
      - The block validation function now uses the Aura Ext Executor so that
      the seal of the block is validated
      - Previous point requires that we seal block before calling into
      `validate_block`, I introduced a helper function for that
      - Test client adjusted to provide a slot to the relay chain proof and
      the aura pre-digest
      df818d29
    • Alexandru Vasile's avatar
      rpc-v2/transaction: Generate `Invalid` events and add tests (#3784) · 598e9557
      Alexandru Vasile authored
      
      
      This PR ensures that the transaction API generates an `Invalid` events
      for transaction bytes that fail to decode.
      
      The spec mentioned the `Invalid` event at the jsonrpc error section,
      however this spec PR makes things clearer:
      - https://github.com/paritytech/json-rpc-interface-spec/pull/146
      
      While at it have discovered an inconsistency with the generated events.
      The drop event from the transaction pool was incorrectly mapped to the
      `invalid` event.
      
      Added tests for the API stabilize the API soon:
      - https://github.com/paritytech/json-rpc-interface-spec/pull/144
      
      
      Closes: https://github.com/paritytech/polkadot-sdk/issues/3083
      
      
      cc @paritytech/subxt-team
      
      ---------
      
      Signed-off-by: default avatarAlexandru Vasile <[email protected]>
      598e9557
    • Dmitry Markin's avatar
      Detect closed notification substreams instead of evicting all peers (#3983) · a26d25d5
      Dmitry Markin authored
      
      
      This PR brings the fix
      https://github.com/paritytech/substrate/pull/13396 to polkadot-sdk.
      
      In the past, due to insufficient inbound slot count on polkadot &
      kusama, this fix led to low peer count. The situation has improved since
      then after changing the default ratio between `--in-peers` &
      `--out-peers`.
      
      Nevertheless, it's expected that the reported total peer count with this
      fix is going to be lower than without it. This should be seen as the
      correct number of working connections reported, as opposed to also
      reporting already closed connections, and not as lower count of working
      connections with peers.
      
      This PR also removes the peer eviction mechanism, as closed substream
      detection is a more granular way of detecting peers that stopped syncing
      with us.
      
      The burn-in has been already performed as part of testing these changes
      in https://github.com/paritytech/polkadot-sdk/pull/3426.
      
      ---------
      
      Co-authored-by: default avatarAaro Altonen <[email protected]>
      a26d25d5
    • Alexandru Vasile's avatar
      peer_store: Increase peer ban time until escapes banned threshold (#4031) · b1c9209a
      Alexandru Vasile authored
      
      
      This is a tiny PR to increase the time a peer remains banned.
      
      A peer is banned when the reputation drops below a threshold.
      With every second, the peer reputation is exponentially decayed towards
      zero.
      
      For the previous setup:
      - decaying to zero from (i32::MAX or i32::MIN) would take 948 seconds
      (15mins 48seconds)
      - from i32::MIN to escaping the banned threshold would take 10 seconds
      This means we are decaying reputation a bit too aggressive and
      misbehaving peers can misbehave again in 10 seconds.
      Another side effect of this is that we have encountered multiple
      warnings caused by a few misbehaving peers.
      
      In the new setup:
      - decaying to zero from (i32::MAX or i32::MIN) would take 3544 seconds
      (59 minutes)
      - from i32::MIN to escaping the banned threshold would take ~69 seconds
      
      This is a followup of:
      - https://github.com/paritytech/polkadot-sdk/pull/4000.
      
      ### Testing Done
      - Created a misbehaving client with
      [subp2p-explorer](https://github.com/lexnv/subp2p-explorer), the client
      is banned for approx 69seconds until it is allowed to connect again.
      
      cc @paritytech/networking
      
      ---------
      
      Signed-off-by: default avatarAlexandru Vasile <[email protected]>
      b1c9209a
    • Facundo Farall's avatar
      Upgrade `trie-db` from `0.28.0` to `0.29.0` (#3982) · 4e73c0fc
      Facundo Farall authored
      
      
      # Description
      - What does this PR do?
      1. Upgrades `trie-db`'s version to the latest release. This release
      includes, among others, an implementation of `DoubleEndedIterator` for
      the `TrieDB` struct, allowing to iterate both backwards and forwards
      within the leaves of a trie.
      2. Upgrades `trie-bench` to `0.39.0` for compatibility.
      3. Upgrades `criterion` to `0.5.1` for compatibility.
      - Why are these changes needed?
      Besides keeping up with the upgrade of `trie-db`, this specifically adds
      the functionality of iterating back on the leafs of a trie, with
      `sp-trie`. In a project we're currently working on, this comes very
      handy to verify a Merkle proof that is the response to a challenge. The
      challenge is a random hash that (most likely) will not be an existing
      leaf in the trie. So the challenged user, has to provide a Merkle proof
      of the previous and next existing leafs in the trie, that surround the
      random challenged hash.
      
      Without having DoubleEnded iterators, we're forced to iterate until we
      find the first existing leaf, like so:
      ```rust
              // ************* VERIFIER (RUNTIME) *************
              // Verify proof. This generates a partial trie based on the proof and
              // checks that the root hash matches the `expected_root`.
              let (memdb, root) = proof.to_memory_db(Some(&root)).unwrap();
              let trie = TrieDBBuilder::<LayoutV1<RefHasher>>::new(&memdb, &root).build();
      
              // Print all leaf node keys and values.
              println!("\nPrinting leaf nodes of partial tree...");
              for key in trie.key_iter().unwrap() {
                  if key.is_ok() {
                      println!("Leaf node key: {:?}", key.clone().unwrap());
      
                      let val = trie.get(&key.unwrap());
      
                      if val.is_ok() {
                          println!("Leaf node value: {:?}", val.unwrap());
                      } else {
                          println!("Leaf node value: None");
                      }
                  }
              }
      
              println!("RECONSTRUCTED TRIE {:#?}", trie);
      
              // Create an iterator over the leaf nodes.
              let mut iter = trie.iter().unwrap();
      
              // First element with a value should be the previous existing leaf to the challenged hash.
              let mut prev_key = None;
              for element in &mut iter {
                  if element.is_ok() {
                      let (key, _) = element.unwrap();
                      prev_key = Some(key);
                      break;
                  }
              }
              assert!(prev_key.is_some());
      
              // Since hashes are `Vec<u8>` ordered in big-endian, we can compare them directly.
              assert!(prev_key.unwrap() <= challenge_hash.to_vec());
      
              // The next element should exist (meaning there is no other existing leaf between the
              // previous and next leaf) and it should be greater than the challenged hash.
              let next_key = iter.next().unwrap().unwrap().0;
              assert!(next_key >= challenge_hash.to_vec());
      ```
      
      With DoubleEnded iterators, we can avoid that, like this:
      ```rust
              // ************* VERIFIER (RUNTIME) *************
              // Verify proof. This generates a partial trie based on the proof and
              // checks that the root hash matches the `expected_root`.
              let (memdb, root) = proof.to_memory_db(Some(&root)).unwrap();
              let trie = TrieDBBuilder::<LayoutV1<RefHasher>>::new(&memdb, &root).build();
      
              // Print all leaf node keys and values.
              println!("\nPrinting leaf nodes of partial tree...");
              for key in trie.key_iter().unwrap() {
                  if key.is_ok() {
                      println!("Leaf node key: {:?}", key.clone().unwrap());
      
                      let val = trie.get(&key.unwrap());
      
                      if val.is_ok() {
                          println!("Leaf node value: {:?}", val.unwrap());
                      } else {
                          println!("Leaf node value: None");
                      }
                  }
              }
      
              // println!("RECONSTRUCTED TRIE {:#?}", trie);
              println!("\nChallenged key: {:?}", challenge_hash);
      
              // Create an iterator over the leaf nodes.
              let mut double_ended_iter = trie.into_double_ended_iter().unwrap();
      
              // First element with a value should be the previous existing leaf to the challenged hash.
              double_ended_iter.seek(&challenge_hash.to_vec()).unwrap();
              let next_key = double_ended_iter.next_back().unwrap().unwrap().0;
              let prev_key = double_ended_iter.next_back().unwrap().unwrap().0;
      
              // Since hashes are `Vec<u8>` ordered in big-endian, we can compare them directly.
              println!("Prev key: {:?}", prev_key);
              assert!(prev_key <= challenge_hash.to_vec());
      
              println!("Next key: {:?}", next_key);
              assert!(next_key >= challenge_hash.to_vec());
      ```
      - How were these changes implemented and what do they affect?
      All that is needed for this functionality to be exposed is changing the
      version number of `trie-db` in all the `Cargo.toml`s applicable, and
      re-exporting some additional structs from `trie-db` in `sp-trie`.
      
      ---------
      
      Co-authored-by: default avatarBastian Köcher <[email protected]>
      4e73c0fc
    • PG Herveou's avatar
      Contracts: Refactor API to use WeightMeter (#2943) · b6231c79
      PG Herveou authored
      
      
      Update the Contracts API to use `WeightMeter`, as it simplifies the code
      and makes it easier to reason about, rather than taking a mutable weight
      or returning a tuple with the weight consumed
      
      ---------
      
      Co-authored-by: default avatarAlexander Theißen <[email protected]>
      b6231c79
    • Ankan's avatar
      Nomination pool configurations can be managed by custom origin (#3959) · 10ed7643
      Ankan authored
      
      
      closes https://github.com/paritytech/polkadot-sdk/issues/3894
      
      Allows Nomination Pool configuration to be set by a custom origin
      instead of root.
      
      In runtimes, we would set this to be `StakingAdmin`, same as for
      pallet-staking.
      
      ---------
      
      Co-authored-by: default avatarLiam Aharon <[email protected]>
      10ed7643
    • gupnik's avatar
      Fixes validation for `SkipCheckIfFeeless` extension (#3993) · 0e27b881
      gupnik authored
      During validation, `SkipCheckIfFeeless` should check if the call is
      `feeless` and delegate to the wrapped extension if not.
      0e27b881
  9. Apr 08, 2024
    • Léa Narzis's avatar
      Adapt `RemoteExternalities` and its related types to be used with generic hash parameters (#3953) · d733c77e
      Léa Narzis authored
      
      
      Closes  https://github.com/paritytech/polkadot-sdk/issues/3737
      
      ---------
      
      Co-authored-by: command-bot <>
      Co-authored-by: default avatarOliver Tale-Yazdi <[email protected]>
      Co-authored-by: default avatarBastian Köcher <[email protected]>
      d733c77e
    • Aaro Altonen's avatar
      Integrate litep2p into Polkadot SDK (#2944) · 80616f6d
      Aaro Altonen authored
      [litep2p](https://github.com/altonen/litep2p) is a libp2p-compatible P2P
      networking library. It supports all of the features of `rust-libp2p`
      that are currently being utilized by Polkadot SDK.
      
      Compared to `rust-libp2p`, `litep2p` has a quite different architecture
      which is why the new `litep2p` network backend is only able to use a
      little of the existing code in `sc-network`. The design has been mainly
      influenced by how we'd wish to structure our networking-related code in
      Polkadot SDK: independent higher-levels protocols directly communicating
      with the network over links that support bidirectional backpressure. A
      good example would be `NotificationHandle`/`RequestResponseHandle`
      abstractions which allow, e.g., `SyncingEngine` to directly communicate
      with peers to announce/request blocks.
      
      I've tried running `polkadot --network-backend litep2p` with a few
      different peer configurations and there is a noticeable reduction in
      networking CPU usage. For high load (`...
      80616f6d