Newer
Older
// This file is part of Substrate.
// Copyright (C) Parity Technologies (UK) Ltd.
// SPDX-License-Identifier: Apache-2.0
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
//! # Generalized Message Queue Pallet
//!
//! Provides generalized message queuing and processing capabilities on a per-queue basis for
//! arbitrary use-cases.
//!
//! # Design Goals
//!
//! 1. Minimal assumptions about `Message`s and `MessageOrigin`s. Both should be MEL bounded blobs.
//! This ensures the generality and reusability of the pallet.
//! 2. Well known and tightly limited pre-dispatch PoV weights, especially for message execution.
//! This is paramount for the success of the pallet since message execution is done in
//! `on_initialize` which must _never_ under-estimate its PoV weight. It also needs a frugal PoV
//! footprint since PoV is scarce and this is (possibly) done in every block. This must also hold
//! in the presence of unpredictable message size distributions.
//! 3. Usable as XCMP, DMP and UMP message/dispatch queue - possibly through adapter types.
//!
//! # Design
//!
//! The pallet has means to enqueue, store and process messages. This is implemented by having
//! *queues* which store enqueued messages and can be *served* to process said messages. A queue is
//! identified by its origin in the `BookStateFor`. Each message has an origin which defines into
//! which queue it will be stored. Messages are stored by being appended to the last [`Page`] of a
//! book. Each book keeps track of its pages by indexing `Pages`. The `ReadyRing` contains all
//! queues which hold at least one unprocessed message and are thereby *ready* to be serviced. The
//! `ServiceHead` indicates which *ready* queue is the next to be serviced.
//! The pallet implements [`frame_support::traits::EnqueueMessage`],
//! [`frame_support::traits::ServiceQueues`] and has [`frame_support::traits::ProcessMessage`] and
//! [`OnQueueChanged`] hooks to communicate with the outside world.
//!
//! NOTE: The storage items are not linked since they are not public.
//!
//! **Message Execution**
//!
//! Executing a message is offloaded to the [`Config::MessageProcessor`] which contains the actual
//! logic of how to handle the message since they are blobs. Storage changes are not rolled back on
//! error.
//!
//! A failed message can be temporarily or permanently overweight. The pallet will perpetually try
//! to execute a temporarily overweight message. A permanently overweight message is skipped and
//! must be executed manually.
//!
//! **Reentrancy**
//!
//! This pallet has two entry points for executing (possibly recursive) logic;
//! [`Pallet::service_queues`] and [`Pallet::execute_overweight`]. Both entry points are guarded by
//! the same mutex to error on reentrancy. The only functions that are explicitly **allowed** to be
//! called by a message processor are: [`Pallet::enqueue_message`] and
//! [`Pallet::enqueue_messages`]. All other functions are forbidden and error with
//! [`Error::RecursiveDisallowed`].
//!
//! **Pagination**
//!
//! Queues are stored in a *paged* manner by splitting their messages into [`Page`]s. This results
//! in a lot of complexity when implementing the pallet but is completely necessary to achieve the
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
//! second #[Design Goal](design-goals). The problem comes from the fact a message can *possibly* be
//! quite large, lets say 64KiB. This then results in a *MEL* of at least 64KiB which results in a
//! PoV of at least 64KiB. Now we have the assumption that most messages are much shorter than their
//! maximum allowed length. This would result in most messages having a pre-dispatch PoV size which
//! is much larger than their post-dispatch PoV size, possibly by a factor of thousand. Disregarding
//! this observation would cripple the processing power of the pallet since it cannot straighten out
//! this discrepancy at runtime. Conceptually, the implementation is packing as many messages into a
//! single bounded vec, as actually fit into the bounds. This reduces the wasted PoV.
//!
//! **Page Data Layout**
//!
//! A Page contains a heap which holds all its messages. The heap is built by concatenating
//! `(ItemHeader, Message)` pairs. The [`ItemHeader`] contains the length of the message which is
//! needed for retrieving it. This layout allows for constant access time of the next message and
//! linear access time for any message in the page. The header must remain minimal to reduce its PoV
//! impact.
//!
//! **Weight Metering**
//!
//! The pallet utilizes the [`sp_weights::WeightMeter`] to manually track its consumption to always
//! stay within the required limit. This implies that the message processor hook can calculate the
//! weight of a message without executing it. This restricts the possible use-cases but is necessary
//! since the pallet runs in `on_initialize` which has a hard weight limit. The weight meter is used
//! in a way that `can_accrue` and `check_accrue` are always used to check the remaining weight of
//! an operation before committing to it. The process of exiting due to insufficient weight is
//! termed "bailing".
//!
//! # Scenario: Message enqueuing
//!
//! A message `m` is enqueued for origin `o` into queue `Q[o]` through
//! [`frame_support::traits::EnqueueMessage::enqueue_message`]`(m, o)`.
//!
//! First the queue is either loaded if it exists or otherwise created with empty default values.
//! The message is then inserted to the queue by appended it into its last `Page` or by creating a
//! new `Page` just for `m` if it does not fit in there. The number of messages in the `Book` is
//! incremented.
//!
//! `Q[o]` is now *ready* which will eventually result in `m` being processed.
//!
//! # Scenario: Message processing
//!
//! The pallet runs each block in `on_initialize` or when being manually called through
//! [`frame_support::traits::ServiceQueues::service_queues`].
//!
//! First it tries to "rotate" the `ReadyRing` by one through advancing the `ServiceHead` to the
//! next *ready* queue. It then starts to service this queue by servicing as many pages of it as
//! possible. Servicing a page means to execute as many message of it as possible. Each executed
//! message is marked as *processed* if the [`Config::MessageProcessor`] return Ok. An event
//! [`Event::Processed`] is emitted afterwards. It is possible that the weight limit of the pallet
//! will never allow a specific message to be executed. In this case it remains as unprocessed and
//! is skipped. This process stops if either there are no more messages in the queue or the
//! remaining weight became insufficient to service this queue. If there is enough weight it tries
//! to advance to the next *ready* queue and service it. This continues until there are no more
//! queues on which it can make progress or not enough weight to check that.
//!
//! # Scenario: Overweight execution
//!
//! A permanently over-weight message which was skipped by the message processing will never be
//! executed automatically through `on_initialize` nor by calling
//! [`frame_support::traits::ServiceQueues::service_queues`].
//!
//! Manual intervention in the form of
//! [`frame_support::traits::ServiceQueues::execute_overweight`] is necessary. Overweight messages
//! emit an [`Event::OverweightEnqueued`] event which can be used to extract the arguments for
//! manual execution. This only works on permanently overweight messages. There is no guarantee that
//! this will work since the message could be part of a stale page and be reaped before execution
//! commences.
//!
//! # Terminology
//!
//! - `Message`: A blob of data into which the pallet has no introspection, defined as
//! [`BoundedSlice<u8, MaxMessageLenOf<T>>`]. The message length is limited by [`MaxMessageLenOf`]
//! which is calculated from [`Config::HeapSize`] and [`ItemHeader::max_encoded_len()`].
//! - `MessageOrigin`: A generic *origin* of a message, defined as [`MessageOriginOf`]. The
//! requirements for it are kept minimal to remain as generic as possible. The type is defined in
//! [`frame_support::traits::ProcessMessage::Origin`].
//! - `Page`: An array of `Message`s, see [`Page`]. Can never be empty.
//! - `Book`: A list of `Page`s, see [`BookState`]. Can be empty.
//! - `Queue`: A `Book` together with an `MessageOrigin` which can be part of the `ReadyRing`. Can
//! be empty.
//! - `ReadyRing`: A double-linked list which contains all *ready* `Queue`s. It chains together the
//! queues via their `ready_neighbours` fields. A `Queue` is *ready* if it contains at least one
//! `Message` which can be processed. Can be empty.
//! - `ServiceHead`: A pointer into the `ReadyRing` to the next `Queue` to be serviced.
//! - (`un`)`processed`: A message is marked as *processed* after it was executed by the pallet. A
//! message which was either: not yet executed or could not be executed remains as `unprocessed`
//! which is the default state for a message after being enqueued.
//! - `knitting`/`unknitting`: The means of adding or removing a `Queue` from the `ReadyRing`.
//! - `MEL`: The Max Encoded Length of a type, see [`codec::MaxEncodedLen`].
//! - `Reentrance`: To enter an execution context again before it has completed.
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
//!
//! # Properties
//!
//! **Liveness - Enqueueing**
//!
//! It is always possible to enqueue any message for any `MessageOrigin`.
//!
//! **Liveness - Processing**
//!
//! `on_initialize` always respects its finite weight-limit.
//!
//! **Progress - Enqueueing**
//!
//! An enqueued message immediately becomes *unprocessed* and thereby eligible for execution.
//!
//! **Progress - Processing**
//!
//! The pallet will execute at least one unprocessed message per block, if there is any. Ensuring
//! this property needs careful consideration of the concrete weights, since it is possible that the
//! weight limit of `on_initialize` never allows for the execution of even one message; trivially if
//! the limit is set to zero. `integrity_test` can be used to ensure that this property holds.
//!
//! **Fairness - Enqueuing**
//!
//! Enqueueing a message for a specific `MessageOrigin` does not influence the ability to enqueue a
//! message for the same of any other `MessageOrigin`; guaranteed by **Liveness - Enqueueing**.
//!
//! **Fairness - Processing**
//!
//! The average amount of weight available for message processing is the same for each queue if the
//! number of queues is constant. Creating a new queue must therefore be, possibly economically,
//! expensive. Currently this is archived by having one queue per para-chain/thread, which keeps the
//! number of queues within `O(n)` and should be "good enough".
#![deny(missing_docs)]
#![cfg_attr(not(feature = "std"), no_std)]
mod benchmarking;
mod integration_test;
mod mock;
pub mod mock_helpers;
mod tests;
pub mod weights;
use codec::{Codec, Decode, Encode, MaxEncodedLen};
use frame_support::{
defensive,
pallet_prelude::*,
traits::{
Defensive, DefensiveSaturating, DefensiveTruncateFrom, EnqueueMessage,
ExecuteOverweightError, Footprint, ProcessMessage, ProcessMessageError, QueueFootprint,
QueuePausedQuery, ServiceQueues,
},
BoundedSlice, CloneNoBound, DefaultNoBound,
};
use frame_system::pallet_prelude::*;
pub use pallet::*;
use scale_info::TypeInfo;
use sp_arithmetic::traits::{BaseArithmetic, Unsigned};
use sp_core::{defer, H256};
traits::{One, Zero},
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
SaturatedConversion, Saturating,
};
use sp_std::{fmt::Debug, ops::Deref, prelude::*, vec};
use sp_weights::WeightMeter;
pub use weights::WeightInfo;
/// Type for identifying a page.
type PageIndex = u32;
/// Data encoded and prefixed to the encoded `MessageItem`.
#[derive(Encode, Decode, PartialEq, MaxEncodedLen, Debug)]
pub struct ItemHeader<Size> {
/// The length of this item, not including the size of this header. The next item of the page
/// follows immediately after the payload of this item.
payload_len: Size,
/// Whether this item has been processed.
is_processed: bool,
}
/// A page of messages. Pages always contain at least one item.
#[derive(
CloneNoBound, Encode, Decode, RuntimeDebugNoBound, DefaultNoBound, TypeInfo, MaxEncodedLen,
)]
#[scale_info(skip_type_params(HeapSize))]
#[codec(mel_bound(Size: MaxEncodedLen))]
pub struct Page<Size: Into<u32> + Debug + Clone + Default, HeapSize: Get<Size>> {
/// Messages remaining to be processed; this includes overweight messages which have been
/// skipped.
remaining: Size,
/// The size of all remaining messages to be processed.
///
/// Includes overweight messages outside of the `first` to `last` window.
remaining_size: Size,
/// The number of items before the `first` item in this page.
first_index: Size,
/// The heap-offset of the header of the first message item in this page which is ready for
/// processing.
first: Size,
/// The heap-offset of the header of the last message item in this page.
last: Size,
/// The heap. If `self.offset == self.heap.len()` then the page is empty and should be deleted.
heap: BoundedVec<u8, IntoU32<HeapSize, Size>>,
}
impl<
Size: BaseArithmetic + Unsigned + Copy + Into<u32> + Codec + MaxEncodedLen + Debug + Default,
HeapSize: Get<Size>,
> Page<Size, HeapSize>
{
/// Create a [`Page`] from one unprocessed message.
fn from_message<T: Config>(message: BoundedSlice<u8, MaxMessageLenOf<T>>) -> Self {
let payload_len = message.len();
let data_len = ItemHeader::<Size>::max_encoded_len().saturating_add(payload_len);
let payload_len = payload_len.saturated_into();
let header = ItemHeader::<Size> { payload_len, is_processed: false };
let mut heap = Vec::with_capacity(data_len);
header.using_encoded(|h| heap.extend_from_slice(h));
heap.extend_from_slice(message.deref());
Page {
remaining: One::one(),
remaining_size: payload_len,
first_index: Zero::zero(),
first: Zero::zero(),
last: Zero::zero(),
heap: BoundedVec::defensive_truncate_from(heap),
}
}
/// Try to append one message to a page.
fn try_append_message<T: Config>(
&mut self,
message: BoundedSlice<u8, MaxMessageLenOf<T>>,
) -> Result<(), ()> {
let pos = self.heap.len();
let payload_len = message.len();
let data_len = ItemHeader::<Size>::max_encoded_len().saturating_add(payload_len);
let payload_len = payload_len.saturated_into();
let header = ItemHeader::<Size> { payload_len, is_processed: false };
let heap_size: u32 = HeapSize::get().into();
if (heap_size as usize).saturating_sub(self.heap.len()) < data_len {
// Can't fit.
return Err(())
}
let mut heap = sp_std::mem::take(&mut self.heap).into_inner();
header.using_encoded(|h| heap.extend_from_slice(h));
heap.extend_from_slice(message.deref());
self.heap = BoundedVec::defensive_truncate_from(heap);
self.last = pos.saturated_into();
self.remaining.saturating_inc();
self.remaining_size.saturating_accrue(payload_len);
Ok(())
}
/// Returns the first message in the page without removing it.
///
/// SAFETY: Does not panic even on corrupted storage.
fn peek_first(&self) -> Option<BoundedSlice<u8, IntoU32<HeapSize, Size>>> {
if self.first > self.last {
return None
}
let f = (self.first.into() as usize).min(self.heap.len());
let mut item_slice = &self.heap[f..];
if let Ok(h) = ItemHeader::<Size>::decode(&mut item_slice) {
let payload_len = h.payload_len.into() as usize;
if payload_len <= item_slice.len() {
// impossible to truncate since is sliced up from `self.heap: BoundedVec<u8,
// HeapSize>`
return Some(BoundedSlice::defensive_truncate_from(&item_slice[..payload_len]))
}
}
defensive!("message-queue: heap corruption");
None
}
/// Point `first` at the next message, marking the first as processed if `is_processed` is true.
fn skip_first(&mut self, is_processed: bool) {
let f = (self.first.into() as usize).min(self.heap.len());
if let Ok(mut h) = ItemHeader::decode(&mut &self.heap[f..]) {
if is_processed && !h.is_processed {
h.is_processed = true;
h.using_encoded(|d| self.heap[f..f + d.len()].copy_from_slice(d));
self.remaining.saturating_dec();
self.remaining_size.saturating_reduce(h.payload_len);
}
self.first
.saturating_accrue(ItemHeader::<Size>::max_encoded_len().saturated_into());
self.first.saturating_accrue(h.payload_len);
self.first_index.saturating_inc();
}
}
/// Return the message with index `index` in the form of `(position, processed, message)`.
fn peek_index(&self, index: usize) -> Option<(usize, bool, &[u8])> {
let mut pos = 0;
let mut item_slice = &self.heap[..];
let header_len: usize = ItemHeader::<Size>::max_encoded_len().saturated_into();
for _ in 0..index {
let h = ItemHeader::<Size>::decode(&mut item_slice).ok()?;
let item_len = h.payload_len.into() as usize;
if item_slice.len() < item_len {
return None
}
item_slice = &item_slice[item_len..];
pos.saturating_accrue(header_len.saturating_add(item_len));
}
let h = ItemHeader::<Size>::decode(&mut item_slice).ok()?;
if item_slice.len() < h.payload_len.into() as usize {
return None
}
item_slice = &item_slice[..h.payload_len.into() as usize];
Some((pos, h.is_processed, item_slice))
}
/// Set the `is_processed` flag for the item at `pos` to be `true` if not already and decrement
/// the `remaining` counter of the page.
///
/// Does nothing if no [`ItemHeader`] could be decoded at the given position.
fn note_processed_at_pos(&mut self, pos: usize) {
if let Ok(mut h) = ItemHeader::<Size>::decode(&mut &self.heap[pos..]) {
if !h.is_processed {
h.is_processed = true;
h.using_encoded(|d| self.heap[pos..pos + d.len()].copy_from_slice(d));
self.remaining.saturating_dec();
self.remaining_size.saturating_reduce(h.payload_len);
}
}
}
/// Returns whether the page is *complete* which means that no messages remain.
fn is_complete(&self) -> bool {
self.remaining.is_zero()
}
}
/// A single link in the double-linked Ready Ring list.
#[derive(Clone, Encode, Decode, MaxEncodedLen, TypeInfo, RuntimeDebug, PartialEq)]
pub struct Neighbours<MessageOrigin> {
/// The previous queue.
prev: MessageOrigin,
/// The next queue.
next: MessageOrigin,
}
/// The state of a queue as represented by a book of its pages.
///
/// Each queue has exactly one book which holds all of its pages. All pages of a book combined
/// contain all of the messages of its queue; hence the name *Book*.
/// Books can be chained together in a double-linked fashion through their `ready_neighbours` field.
#[derive(Clone, Encode, Decode, MaxEncodedLen, TypeInfo, RuntimeDebug)]
pub struct BookState<MessageOrigin> {
/// The first page with some items to be processed in it. If this is `>= end`, then there are
/// no pages with items to be processing in them.
begin: PageIndex,
/// One more than the last page with some items to be processed in it.
end: PageIndex,
/// The number of pages stored at present.
///
/// This might be larger than `end-begin`, because we keep pages with unprocessed overweight
/// messages outside of the end/begin window.
count: PageIndex,
/// If this book has any ready pages, then this will be `Some` with the previous and next
/// neighbours. This wraps around.
ready_neighbours: Option<Neighbours<MessageOrigin>>,
/// The number of unprocessed messages stored at present.
message_count: u64,
/// The total size of all unprocessed messages stored at present.
size: u64,
}
impl<MessageOrigin> Default for BookState<MessageOrigin> {
fn default() -> Self {
Self { begin: 0, end: 0, count: 0, ready_neighbours: None, message_count: 0, size: 0 }
}
}
impl<MessageOrigin> From<BookState<MessageOrigin>> for QueueFootprint {
fn from(book: BookState<MessageOrigin>) -> Self {
QueueFootprint {
pages: book.count,
ready_pages: book.end.defensive_saturating_sub(book.begin),
storage: Footprint { count: book.message_count, size: book.size },
}
}
}
/// Handler code for when the items in a queue change.
pub trait OnQueueChanged<Id> {
/// Note that the queue `id` now has `item_count` items in it, taking up `items_size` bytes.
fn on_queue_changed(id: Id, fp: QueueFootprint);
}
impl<Id> OnQueueChanged<Id> for () {
fn on_queue_changed(_: Id, _: QueueFootprint) {}
}
#[frame_support::pallet]
pub mod pallet {
use super::*;
#[pallet::pallet]
pub struct Pallet<T>(_);
/// The module configuration trait.
#[pallet::config]
pub trait Config: frame_system::Config {
/// The overarching event type.
type RuntimeEvent: From<Event<Self>> + IsType<<Self as frame_system::Config>::RuntimeEvent>;
/// Weight information for extrinsics in this pallet.
type WeightInfo: WeightInfo;
/// Processor for a message.
///
/// Storage changes are not rolled back on error.
///
/// # Benchmarking
///
/// Must be set to [`mock_helpers::NoopMessageProcessor`] for benchmarking.
/// Other message processors that consumes exactly (1, 1) weight for any give message will
/// work as well. Otherwise the benchmarking will also measure the weight of the message
/// processor, which is not desired.
type MessageProcessor: ProcessMessage;
/// Page/heap size type.
type Size: BaseArithmetic
+ Unsigned
+ Copy
+ Into<u32>
+ Member
+ Encode
+ Decode
+ MaxEncodedLen
+ TypeInfo
+ Default;
/// Code to be called when a message queue changes - either with items introduced or
/// removed.
type QueueChangeHandler: OnQueueChanged<<Self::MessageProcessor as ProcessMessage>::Origin>;
/// Queried by the pallet to check whether a queue can be serviced.
///
/// This also applies to manual servicing via `execute_overweight` and `service_queues`. The
/// value of this is only polled once before servicing the queue. This means that changes to
/// it that happen *within* the servicing will not be reflected.
type QueuePausedQuery: QueuePausedQuery<<Self::MessageProcessor as ProcessMessage>::Origin>;
/// The size of the page; this implies the maximum message size which can be sent.
///
/// A good value depends on the expected message sizes, their weights, the weight that is
/// available for processing them and the maximal needed message size. The maximal message
/// size is slightly lower than this as defined by [`MaxMessageLenOf`].
#[pallet::constant]
type HeapSize: Get<Self::Size>;
/// The maximum number of stale pages (i.e. of overweight messages) allowed before culling
/// can happen. Once there are more stale pages than this, then historical pages may be
/// dropped, even if they contain unprocessed overweight messages.
#[pallet::constant]
type MaxStale: Get<u32>;
/// The amount of weight (if any) which should be provided to the message queue for
/// servicing enqueued items `on_initialize`.
///
/// This may be legitimately `None` in the case that you will call
/// `ServiceQueues::service_queues` manually or set [`Self::IdleMaxServiceWeight`] to have
/// it run in `on_idle`.
#[pallet::constant]
type ServiceWeight: Get<Option<Weight>>;
/// The maximum amount of weight (if any) to be used from remaining weight `on_idle` which
/// should be provided to the message queue for servicing enqueued items `on_idle`.
/// Useful for parachains to process messages at the same block they are received.
///
/// If `None`, it will not call `ServiceQueues::service_queues` in `on_idle`.
#[pallet::constant]
type IdleMaxServiceWeight: Get<Option<Weight>>;
}
#[pallet::event]
#[pallet::generate_deposit(pub(super) fn deposit_event)]
pub enum Event<T: Config> {
/// Message discarded due to an error in the `MessageProcessor` (usually a format error).
ProcessingFailed {
/// The `blake2_256` hash of the message.
id: H256,
/// The queue of the message.
origin: MessageOriginOf<T>,
/// The error that occurred.
///
/// This error is pretty opaque. More fine-grained errors need to be emitted as events
/// by the `MessageProcessor`.
error: ProcessMessageError,
},
Processed {
/// The `blake2_256` hash of the message.
id: H256,
/// The queue of the message.
origin: MessageOriginOf<T>,
/// How much weight was used to process the message.
weight_used: Weight,
/// Whether the message was processed.
///
/// Note that this does not mean that the underlying `MessageProcessor` was internally
/// successful. It *solely* means that the MQ pallet will treat this as a success
/// condition and discard the message. Any internal error needs to be emitted as events
/// by the `MessageProcessor`.
success: bool,
},
/// Message placed in overweight queue.
OverweightEnqueued {
/// The `blake2_256` hash of the message.
id: [u8; 32],
/// The queue of the message.
/// The page of the message.
/// The index of the message within the page.
message_index: T::Size,
},
/// This page was reaped.
PageReaped {
/// The queue of the page.
origin: MessageOriginOf<T>,
/// The index of the page.
index: PageIndex,
},
}
#[pallet::error]
pub enum Error<T> {
/// Page is not reapable because it has items remaining to be processed and is not old
/// enough.
NotReapable,
/// Page to be reaped does not exist.
NoPage,
/// The referenced message could not be found.
NoMessage,
/// The message was already processed and cannot be processed again.
AlreadyProcessed,
/// The message is queued for future execution.
Queued,
/// There is temporarily not enough weight to continue servicing messages.
InsufficientWeight,
/// This message is temporarily unprocessable.
///
/// Such errors are expected, but not guaranteed, to resolve themselves eventually through
/// retrying.
TemporarilyUnprocessable,
/// The queue is paused and no message can be executed from it.
///
/// This can change at any time and may resolve in the future by re-trying.
QueuePaused,
/// Another call is in progress and needs to finish before this call can happen.
RecursiveDisallowed,
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
}
/// The index of the first and last (non-empty) pages.
#[pallet::storage]
pub(super) type BookStateFor<T: Config> =
StorageMap<_, Twox64Concat, MessageOriginOf<T>, BookState<MessageOriginOf<T>>, ValueQuery>;
/// The origin at which we should begin servicing.
#[pallet::storage]
pub(super) type ServiceHead<T: Config> = StorageValue<_, MessageOriginOf<T>, OptionQuery>;
/// The map of page indices to pages.
#[pallet::storage]
pub(super) type Pages<T: Config> = StorageDoubleMap<
_,
Twox64Concat,
MessageOriginOf<T>,
Twox64Concat,
PageIndex,
Page<T::Size, T::HeapSize>,
OptionQuery,
>;
#[pallet::hooks]
impl<T: Config> Hooks<BlockNumberFor<T>> for Pallet<T> {
fn on_initialize(_n: BlockNumberFor<T>) -> Weight {
if let Some(weight_limit) = T::ServiceWeight::get() {
Self::service_queues(weight_limit)
} else {
Weight::zero()
}
}
fn on_idle(_n: BlockNumberFor<T>, remaining_weight: Weight) -> Weight {
if let Some(weight_limit) = T::IdleMaxServiceWeight::get() {
// Make use of the remaining weight to process enqueued messages.
Self::service_queues(weight_limit.min(remaining_weight))
} else {
Weight::zero()
}
}
#[cfg(feature = "try-runtime")]
fn try_state(_: BlockNumberFor<T>) -> Result<(), sp_runtime::TryRuntimeError> {
Self::do_try_state()
}
/// Check all compile-time assumptions about [`crate::Config`].
#[cfg(test)]
Self::do_integrity_test().expect("Pallet config is valid; qed")
}
}
#[pallet::call]
impl<T: Config> Pallet<T> {
/// Remove a page which has no more messages remaining to be processed or is stale.
#[pallet::weight(T::WeightInfo::reap_page())]
pub fn reap_page(
origin: OriginFor<T>,
message_origin: MessageOriginOf<T>,
page_index: PageIndex,
) -> DispatchResult {
let _ = ensure_signed(origin)?;
Self::do_reap_page(&message_origin, page_index)
}
/// Execute an overweight message.
///
/// Temporary processing errors will be propagated whereas permanent errors are treated
/// as success condition.
///
/// - `origin`: Must be `Signed`.
/// - `message_origin`: The origin from which the message to be executed arrived.
/// - `page`: The page in the queue in which the message to be executed is sitting.
/// - `index`: The index into the queue of the message to be executed.
/// - `weight_limit`: The maximum amount of weight allowed to be consumed in the execution
/// of the message.
///
/// Benchmark complexity considerations: O(index + weight_limit).
#[pallet::weight(
T::WeightInfo::execute_overweight_page_updated().max(
T::WeightInfo::execute_overweight_page_removed()).saturating_add(*weight_limit)
)]
pub fn execute_overweight(
origin: OriginFor<T>,
message_origin: MessageOriginOf<T>,
page: PageIndex,
index: T::Size,
weight_limit: Weight,
) -> DispatchResultWithPostInfo {
let _ = ensure_signed(origin)?;
let actual_weight =
Self::do_execute_overweight(message_origin, page, index, weight_limit)?;
Ok(Some(actual_weight).into())
}
}
}
/// The status of a page after trying to execute its next message.
#[derive(PartialEq, Debug)]
enum PageExecutionStatus {
/// The execution bailed because there was not enough weight remaining.
Bailed,
/// The page did not make any progress on its execution.
///
/// This is a transient condition and can be handled by retrying - exactly like [Bailed].
NoProgress,
/// No more messages could be loaded. This does _not_ imply `page.is_complete()`.
///
/// The reasons for this status are:
/// - The end of the page is reached but there could still be skipped messages.
/// - The storage is corrupted.
NoMore,
}
/// The status after trying to execute the next item of a [`Page`].
#[derive(PartialEq, Debug)]
enum ItemExecutionStatus {
/// The execution bailed because there was not enough weight remaining.
Bailed,
/// The item did not make any progress on its execution.
///
/// This is a transient condition and can be handled by retrying - exactly like [Bailed].
NoProgress,
/// The item was not found.
NoItem,
/// Whether the execution of an item resulted in it being processed.
///
/// One reason for `false` would be permanently overweight.
Executed(bool),
}
/// The status of an attempt to process a message.
#[derive(PartialEq)]
enum MessageExecutionStatus {
/// There is not enough weight remaining at present.
InsufficientWeight,
/// There will never be enough weight.
Overweight,
/// The message was processed successfully.
Processed,
/// The message was processed and resulted in a, possibly permanent, error.
Unprocessable { permanent: bool },
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
}
impl<T: Config> Pallet<T> {
/// Knit `origin` into the ready ring right at the end.
///
/// Return the two ready ring neighbours of `origin`.
fn ready_ring_knit(origin: &MessageOriginOf<T>) -> Result<Neighbours<MessageOriginOf<T>>, ()> {
if let Some(head) = ServiceHead::<T>::get() {
let mut head_book_state = BookStateFor::<T>::get(&head);
let mut head_neighbours = head_book_state.ready_neighbours.take().ok_or(())?;
let tail = head_neighbours.prev;
head_neighbours.prev = origin.clone();
head_book_state.ready_neighbours = Some(head_neighbours);
BookStateFor::<T>::insert(&head, head_book_state);
let mut tail_book_state = BookStateFor::<T>::get(&tail);
let mut tail_neighbours = tail_book_state.ready_neighbours.take().ok_or(())?;
tail_neighbours.next = origin.clone();
tail_book_state.ready_neighbours = Some(tail_neighbours);
BookStateFor::<T>::insert(&tail, tail_book_state);
Ok(Neighbours { next: head, prev: tail })
} else {
ServiceHead::<T>::put(origin);
Ok(Neighbours { next: origin.clone(), prev: origin.clone() })
}
}
fn ready_ring_unknit(origin: &MessageOriginOf<T>, neighbours: Neighbours<MessageOriginOf<T>>) {
if origin == &neighbours.next {
debug_assert!(
origin == &neighbours.prev,
"unknitting from single item ring; outgoing must be only item"
);
// Service queue empty.
ServiceHead::<T>::kill();
} else {
BookStateFor::<T>::mutate(&neighbours.next, |book_state| {
if let Some(ref mut n) = book_state.ready_neighbours {
n.prev = neighbours.prev.clone()
}
});
BookStateFor::<T>::mutate(&neighbours.prev, |book_state| {
if let Some(ref mut n) = book_state.ready_neighbours {
n.next = neighbours.next.clone()
}
});
if let Some(head) = ServiceHead::<T>::get() {
if &head == origin {
ServiceHead::<T>::put(neighbours.next);
}
} else {
defensive!("`ServiceHead` must be some if there was a ready queue");
}
}
}
/// Tries to bump the current `ServiceHead` to the next ready queue.
///
/// Returns the current head if it got be bumped and `None` otherwise.
fn bump_service_head(weight: &mut WeightMeter) -> Option<MessageOriginOf<T>> {
if weight.try_consume(T::WeightInfo::bump_service_head()).is_err() {
return None
}
if let Some(head) = ServiceHead::<T>::get() {
let mut head_book_state = BookStateFor::<T>::get(&head);
if let Some(head_neighbours) = head_book_state.ready_neighbours.take() {
ServiceHead::<T>::put(&head_neighbours.next);
Some(head)
} else {
None
}
} else {
None
}
}
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
/// The maximal weight that a single message can consume.
///
/// Any message using more than this will be marked as permanently overweight and not
/// automatically re-attempted. Returns `None` if the servicing of a message cannot begin.
/// `Some(0)` means that only messages with no weight may be served.
fn max_message_weight(limit: Weight) -> Option<Weight> {
limit.checked_sub(&Self::single_msg_overhead())
}
/// The overhead of servicing a single message.
fn single_msg_overhead() -> Weight {
T::WeightInfo::bump_service_head()
.saturating_add(T::WeightInfo::service_queue_base())
.saturating_add(
T::WeightInfo::service_page_base_completion()
.max(T::WeightInfo::service_page_base_no_completion()),
)
.saturating_add(T::WeightInfo::service_page_item())
.saturating_add(T::WeightInfo::ready_ring_unknit())
}
/// Checks invariants of the pallet config.
///
/// The results of this can only be relied upon if the config values are set to constants.
#[cfg(test)]
fn do_integrity_test() -> Result<(), String> {
ensure!(!MaxMessageLenOf::<T>::get().is_zero(), "HeapSize too low");
if let Some(service) = T::ServiceWeight::get() {
if Self::max_message_weight(service).is_none() {
return Err(format!(
"ServiceWeight too low: {}. Must be at least {}",
service,
Self::single_msg_overhead(),
))
}
}
Ok(())
}
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
fn do_enqueue_message(
origin: &MessageOriginOf<T>,
message: BoundedSlice<u8, MaxMessageLenOf<T>>,
) {
let mut book_state = BookStateFor::<T>::get(origin);
book_state.message_count.saturating_inc();
book_state
.size
// This should be payload size, but here the payload *is* the message.
.saturating_accrue(message.len() as u64);
if book_state.end > book_state.begin {
debug_assert!(book_state.ready_neighbours.is_some(), "Must be in ready ring if ready");
// Already have a page in progress - attempt to append.
let last = book_state.end - 1;
let mut page = match Pages::<T>::get(origin, last) {
Some(p) => p,
None => {
defensive!("Corruption: referenced page doesn't exist.");
return
},
};
if page.try_append_message::<T>(message).is_ok() {
Pages::<T>::insert(origin, last, &page);
BookStateFor::<T>::insert(origin, book_state);
return
}
} else {
debug_assert!(
book_state.ready_neighbours.is_none(),
"Must not be in ready ring if not ready"
);
// insert into ready queue.
match Self::ready_ring_knit(origin) {
Ok(neighbours) => book_state.ready_neighbours = Some(neighbours),
Err(()) => {
defensive!("Ring state invalid when knitting");
},
}
}
// No room on the page or no page - link in a new page.
book_state.end.saturating_inc();
book_state.count.saturating_inc();
let page = Page::from_message::<T>(message);
Pages::<T>::insert(origin, book_state.end - 1, page);
// NOTE: `T::QueueChangeHandler` is called by the caller.
BookStateFor::<T>::insert(origin, book_state);
}
/// Try to execute a single message that was marked as overweight.
///
/// The `weight_limit` is the weight that can be consumed to execute the message. The base
/// weight of the function it self must be measured by the caller.
pub fn do_execute_overweight(
origin: MessageOriginOf<T>,
page_index: PageIndex,
index: T::Size,
weight_limit: Weight,
) -> Result<Weight, Error<T>> {
match with_service_mutex(|| {
Self::do_execute_overweight_inner(origin, page_index, index, weight_limit)
}) {
Err(()) => Err(Error::<T>::RecursiveDisallowed),
Ok(x) => x,
}
}
/// Same as `do_execute_overweight` but must be called while holding the `service_mutex`.
fn do_execute_overweight_inner(
origin: MessageOriginOf<T>,
page_index: PageIndex,
index: T::Size,
weight_limit: Weight,
) -> Result<Weight, Error<T>> {
let mut book_state = BookStateFor::<T>::get(&origin);
ensure!(!T::QueuePausedQuery::is_paused(&origin), Error::<T>::QueuePaused);
let mut page = Pages::<T>::get(&origin, page_index).ok_or(Error::<T>::NoPage)?;
let (pos, is_processed, payload) =
page.peek_index(index.into() as usize).ok_or(Error::<T>::NoMessage)?;
let payload_len = payload.len() as u64;
ensure!(
page_index < book_state.begin ||
(page_index == book_state.begin && pos < page.first.into() as usize),
Error::<T>::Queued
);
ensure!(!is_processed, Error::<T>::AlreadyProcessed);
use MessageExecutionStatus::*;
let mut weight_counter = WeightMeter::with_limit(weight_limit);
match Self::process_message_payload(
origin.clone(),
page_index,
index,
payload,
&mut weight_counter,
Weight::MAX,
// ^^^ We never recognise it as permanently overweight, since that would result in an
// additional overweight event being deposited.
) {
Overweight | InsufficientWeight => Err(Error::<T>::InsufficientWeight),
Unprocessable { permanent: false } => Err(Error::<T>::TemporarilyUnprocessable),
Unprocessable { permanent: true } | Processed => {
page.note_processed_at_pos(pos);
book_state.message_count.saturating_dec();
book_state.size.saturating_reduce(payload_len);
let page_weight = if page.remaining.is_zero() {
debug_assert!(
page.remaining_size.is_zero(),
"no messages remaining; no space taken; qed"
);
Pages::<T>::remove(&origin, page_index);
debug_assert!(book_state.count >= 1, "page exists, so book must have pages");
book_state.count.saturating_dec();
T::WeightInfo::execute_overweight_page_removed()