Newer
Older
// This file is part of Substrate.
// Copyright (C) Parity Technologies (UK) Ltd.
// SPDX-License-Identifier: Apache-2.0
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
//! # Generalized Message Queue Pallet
//!
//! Provides generalized message queuing and processing capabilities on a per-queue basis for
//! arbitrary use-cases.
//!
//! # Design Goals
//!
//! 1. Minimal assumptions about `Message`s and `MessageOrigin`s. Both should be MEL bounded blobs.
//! This ensures the generality and reusability of the pallet.
//! 2. Well known and tightly limited pre-dispatch PoV weights, especially for message execution.
//! This is paramount for the success of the pallet since message execution is done in
//! `on_initialize` which must _never_ under-estimate its PoV weight. It also needs a frugal PoV
//! footprint since PoV is scarce and this is (possibly) done in every block. This must also hold
//! in the presence of unpredictable message size distributions.
//! 3. Usable as XCMP, DMP and UMP message/dispatch queue - possibly through adapter types.
//!
//! # Design
//!
//! The pallet has means to enqueue, store and process messages. This is implemented by having
//! *queues* which store enqueued messages and can be *served* to process said messages. A queue is
//! identified by its origin in the `BookStateFor`. Each message has an origin which defines into
//! which queue it will be stored. Messages are stored by being appended to the last [`Page`] of a
//! book. Each book keeps track of its pages by indexing `Pages`. The `ReadyRing` contains all
//! queues which hold at least one unprocessed message and are thereby *ready* to be serviced. The
//! `ServiceHead` indicates which *ready* queue is the next to be serviced.
//! The pallet implements [`frame_support::traits::EnqueueMessage`],
//! [`frame_support::traits::ServiceQueues`] and has [`frame_support::traits::ProcessMessage`] and
//! [`OnQueueChanged`] hooks to communicate with the outside world.
//!
//! NOTE: The storage items are not linked since they are not public.
//!
//! **Message Execution**
//!
//! Executing a message is offloaded to the [`Config::MessageProcessor`] which contains the actual
//! logic of how to handle the message since they are blobs. A message can be temporarily or
//! permanently overweight. The pallet will perpetually try to execute a temporarily overweight
//! message. A permanently overweight message is skipped and must be executed manually.
//!
//! **Pagination**
//!
//! Queues are stored in a *paged* manner by splitting their messages into [`Page`]s. This results
//! in a lot of complexity when implementing the pallet but is completely necessary to achieve the
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
//! second #[Design Goal](design-goals). The problem comes from the fact a message can *possibly* be
//! quite large, lets say 64KiB. This then results in a *MEL* of at least 64KiB which results in a
//! PoV of at least 64KiB. Now we have the assumption that most messages are much shorter than their
//! maximum allowed length. This would result in most messages having a pre-dispatch PoV size which
//! is much larger than their post-dispatch PoV size, possibly by a factor of thousand. Disregarding
//! this observation would cripple the processing power of the pallet since it cannot straighten out
//! this discrepancy at runtime. Conceptually, the implementation is packing as many messages into a
//! single bounded vec, as actually fit into the bounds. This reduces the wasted PoV.
//!
//! **Page Data Layout**
//!
//! A Page contains a heap which holds all its messages. The heap is built by concatenating
//! `(ItemHeader, Message)` pairs. The [`ItemHeader`] contains the length of the message which is
//! needed for retrieving it. This layout allows for constant access time of the next message and
//! linear access time for any message in the page. The header must remain minimal to reduce its PoV
//! impact.
//!
//! **Weight Metering**
//!
//! The pallet utilizes the [`sp_weights::WeightMeter`] to manually track its consumption to always
//! stay within the required limit. This implies that the message processor hook can calculate the
//! weight of a message without executing it. This restricts the possible use-cases but is necessary
//! since the pallet runs in `on_initialize` which has a hard weight limit. The weight meter is used
//! in a way that `can_accrue` and `check_accrue` are always used to check the remaining weight of
//! an operation before committing to it. The process of exiting due to insufficient weight is
//! termed "bailing".
//!
//! # Scenario: Message enqueuing
//!
//! A message `m` is enqueued for origin `o` into queue `Q[o]` through
//! [`frame_support::traits::EnqueueMessage::enqueue_message`]`(m, o)`.
//!
//! First the queue is either loaded if it exists or otherwise created with empty default values.
//! The message is then inserted to the queue by appended it into its last `Page` or by creating a
//! new `Page` just for `m` if it does not fit in there. The number of messages in the `Book` is
//! incremented.
//!
//! `Q[o]` is now *ready* which will eventually result in `m` being processed.
//!
//! # Scenario: Message processing
//!
//! The pallet runs each block in `on_initialize` or when being manually called through
//! [`frame_support::traits::ServiceQueues::service_queues`].
//!
//! First it tries to "rotate" the `ReadyRing` by one through advancing the `ServiceHead` to the
//! next *ready* queue. It then starts to service this queue by servicing as many pages of it as
//! possible. Servicing a page means to execute as many message of it as possible. Each executed
//! message is marked as *processed* if the [`Config::MessageProcessor`] return Ok. An event
//! [`Event::Processed`] is emitted afterwards. It is possible that the weight limit of the pallet
//! will never allow a specific message to be executed. In this case it remains as unprocessed and
//! is skipped. This process stops if either there are no more messages in the queue or the
//! remaining weight became insufficient to service this queue. If there is enough weight it tries
//! to advance to the next *ready* queue and service it. This continues until there are no more
//! queues on which it can make progress or not enough weight to check that.
//!
//! # Scenario: Overweight execution
//!
//! A permanently over-weight message which was skipped by the message processing will never be
//! executed automatically through `on_initialize` nor by calling
//! [`frame_support::traits::ServiceQueues::service_queues`].
//!
//! Manual intervention in the form of
//! [`frame_support::traits::ServiceQueues::execute_overweight`] is necessary. Overweight messages
//! emit an [`Event::OverweightEnqueued`] event which can be used to extract the arguments for
//! manual execution. This only works on permanently overweight messages. There is no guarantee that
//! this will work since the message could be part of a stale page and be reaped before execution
//! commences.
//!
//! # Terminology
//!
//! - `Message`: A blob of data into which the pallet has no introspection, defined as
//! [`BoundedSlice<u8, MaxMessageLenOf<T>>`]. The message length is limited by [`MaxMessageLenOf`]
//! which is calculated from [`Config::HeapSize`] and [`ItemHeader::max_encoded_len()`].
//! - `MessageOrigin`: A generic *origin* of a message, defined as [`MessageOriginOf`]. The
//! requirements for it are kept minimal to remain as generic as possible. The type is defined in
//! [`frame_support::traits::ProcessMessage::Origin`].
//! - `Page`: An array of `Message`s, see [`Page`]. Can never be empty.
//! - `Book`: A list of `Page`s, see [`BookState`]. Can be empty.
//! - `Queue`: A `Book` together with an `MessageOrigin` which can be part of the `ReadyRing`. Can
//! be empty.
//! - `ReadyRing`: A double-linked list which contains all *ready* `Queue`s. It chains together the
//! queues via their `ready_neighbours` fields. A `Queue` is *ready* if it contains at least one
//! `Message` which can be processed. Can be empty.
//! - `ServiceHead`: A pointer into the `ReadyRing` to the next `Queue` to be serviced.
//! - (`un`)`processed`: A message is marked as *processed* after it was executed by the pallet. A
//! message which was either: not yet executed or could not be executed remains as `unprocessed`
//! which is the default state for a message after being enqueued.
//! - `knitting`/`unknitting`: The means of adding or removing a `Queue` from the `ReadyRing`.
//! - `MEL`: The Max Encoded Length of a type, see [`codec::MaxEncodedLen`].
//!
//! # Properties
//!
//! **Liveness - Enqueueing**
//!
//! It is always possible to enqueue any message for any `MessageOrigin`.
//!
//! **Liveness - Processing**
//!
//! `on_initialize` always respects its finite weight-limit.
//!
//! **Progress - Enqueueing**
//!
//! An enqueued message immediately becomes *unprocessed* and thereby eligible for execution.
//!
//! **Progress - Processing**
//!
//! The pallet will execute at least one unprocessed message per block, if there is any. Ensuring
//! this property needs careful consideration of the concrete weights, since it is possible that the
//! weight limit of `on_initialize` never allows for the execution of even one message; trivially if
//! the limit is set to zero. `integrity_test` can be used to ensure that this property holds.
//!
//! **Fairness - Enqueuing**
//!
//! Enqueueing a message for a specific `MessageOrigin` does not influence the ability to enqueue a
//! message for the same of any other `MessageOrigin`; guaranteed by **Liveness - Enqueueing**.
//!
//! **Fairness - Processing**
//!
//! The average amount of weight available for message processing is the same for each queue if the
//! number of queues is constant. Creating a new queue must therefore be, possibly economically,
//! expensive. Currently this is archived by having one queue per para-chain/thread, which keeps the
//! number of queues within `O(n)` and should be "good enough".
#![cfg_attr(not(feature = "std"), no_std)]
mod benchmarking;
mod integration_test;
mod mock;
pub mod mock_helpers;
mod tests;
pub mod weights;
use codec::{Codec, Decode, Encode, MaxEncodedLen};
use frame_support::{
defensive,
pallet_prelude::*,
traits::{
DefensiveTruncateFrom, EnqueueMessage, ExecuteOverweightError, Footprint, ProcessMessage,
ProcessMessageError, ServiceQueues,
},
BoundedSlice, CloneNoBound, DefaultNoBound,
};
use frame_system::pallet_prelude::*;
pub use pallet::*;
use scale_info::TypeInfo;
use sp_arithmetic::traits::{BaseArithmetic, Unsigned};
use sp_runtime::{
traits::{Hash, One, Zero},
SaturatedConversion, Saturating,
};
use sp_std::{fmt::Debug, ops::Deref, prelude::*, vec};
use sp_weights::WeightMeter;
pub use weights::WeightInfo;
/// Type for identifying a page.
type PageIndex = u32;
/// Data encoded and prefixed to the encoded `MessageItem`.
#[derive(Encode, Decode, PartialEq, MaxEncodedLen, Debug)]
pub struct ItemHeader<Size> {
/// The length of this item, not including the size of this header. The next item of the page
/// follows immediately after the payload of this item.
payload_len: Size,
/// Whether this item has been processed.
is_processed: bool,
}
/// A page of messages. Pages always contain at least one item.
#[derive(
CloneNoBound, Encode, Decode, RuntimeDebugNoBound, DefaultNoBound, TypeInfo, MaxEncodedLen,
)]
#[scale_info(skip_type_params(HeapSize))]
#[codec(mel_bound(Size: MaxEncodedLen))]
pub struct Page<Size: Into<u32> + Debug + Clone + Default, HeapSize: Get<Size>> {
/// Messages remaining to be processed; this includes overweight messages which have been
/// skipped.
remaining: Size,
/// The size of all remaining messages to be processed.
///
/// Includes overweight messages outside of the `first` to `last` window.
remaining_size: Size,
/// The number of items before the `first` item in this page.
first_index: Size,
/// The heap-offset of the header of the first message item in this page which is ready for
/// processing.
first: Size,
/// The heap-offset of the header of the last message item in this page.
last: Size,
/// The heap. If `self.offset == self.heap.len()` then the page is empty and should be deleted.
heap: BoundedVec<u8, IntoU32<HeapSize, Size>>,
}
impl<
Size: BaseArithmetic + Unsigned + Copy + Into<u32> + Codec + MaxEncodedLen + Debug + Default,
HeapSize: Get<Size>,
> Page<Size, HeapSize>
{
/// Create a [`Page`] from one unprocessed message.
fn from_message<T: Config>(message: BoundedSlice<u8, MaxMessageLenOf<T>>) -> Self {
let payload_len = message.len();
let data_len = ItemHeader::<Size>::max_encoded_len().saturating_add(payload_len);
let payload_len = payload_len.saturated_into();
let header = ItemHeader::<Size> { payload_len, is_processed: false };
let mut heap = Vec::with_capacity(data_len);
header.using_encoded(|h| heap.extend_from_slice(h));
heap.extend_from_slice(message.deref());
Page {
remaining: One::one(),
remaining_size: payload_len,
first_index: Zero::zero(),
first: Zero::zero(),
last: Zero::zero(),
heap: BoundedVec::defensive_truncate_from(heap),
}
}
/// Try to append one message to a page.
fn try_append_message<T: Config>(
&mut self,
message: BoundedSlice<u8, MaxMessageLenOf<T>>,
) -> Result<(), ()> {
let pos = self.heap.len();
let payload_len = message.len();
let data_len = ItemHeader::<Size>::max_encoded_len().saturating_add(payload_len);
let payload_len = payload_len.saturated_into();
let header = ItemHeader::<Size> { payload_len, is_processed: false };
let heap_size: u32 = HeapSize::get().into();
if (heap_size as usize).saturating_sub(self.heap.len()) < data_len {
// Can't fit.
return Err(())
}
let mut heap = sp_std::mem::take(&mut self.heap).into_inner();
header.using_encoded(|h| heap.extend_from_slice(h));
heap.extend_from_slice(message.deref());
self.heap = BoundedVec::defensive_truncate_from(heap);
self.last = pos.saturated_into();
self.remaining.saturating_inc();
self.remaining_size.saturating_accrue(payload_len);
Ok(())
}
/// Returns the first message in the page without removing it.
///
/// SAFETY: Does not panic even on corrupted storage.
fn peek_first(&self) -> Option<BoundedSlice<u8, IntoU32<HeapSize, Size>>> {
if self.first > self.last {
return None
}
let f = (self.first.into() as usize).min(self.heap.len());
let mut item_slice = &self.heap[f..];
if let Ok(h) = ItemHeader::<Size>::decode(&mut item_slice) {
let payload_len = h.payload_len.into() as usize;
if payload_len <= item_slice.len() {
// impossible to truncate since is sliced up from `self.heap: BoundedVec<u8,
// HeapSize>`
return Some(BoundedSlice::defensive_truncate_from(&item_slice[..payload_len]))
}
}
defensive!("message-queue: heap corruption");
None
}
/// Point `first` at the next message, marking the first as processed if `is_processed` is true.
fn skip_first(&mut self, is_processed: bool) {
let f = (self.first.into() as usize).min(self.heap.len());
if let Ok(mut h) = ItemHeader::decode(&mut &self.heap[f..]) {
if is_processed && !h.is_processed {
h.is_processed = true;
h.using_encoded(|d| self.heap[f..f + d.len()].copy_from_slice(d));
self.remaining.saturating_dec();
self.remaining_size.saturating_reduce(h.payload_len);
}
self.first
.saturating_accrue(ItemHeader::<Size>::max_encoded_len().saturated_into());
self.first.saturating_accrue(h.payload_len);
self.first_index.saturating_inc();
}
}
/// Return the message with index `index` in the form of `(position, processed, message)`.
fn peek_index(&self, index: usize) -> Option<(usize, bool, &[u8])> {
let mut pos = 0;
let mut item_slice = &self.heap[..];
let header_len: usize = ItemHeader::<Size>::max_encoded_len().saturated_into();
for _ in 0..index {
let h = ItemHeader::<Size>::decode(&mut item_slice).ok()?;
let item_len = h.payload_len.into() as usize;
if item_slice.len() < item_len {
return None
}
item_slice = &item_slice[item_len..];
pos.saturating_accrue(header_len.saturating_add(item_len));
}
let h = ItemHeader::<Size>::decode(&mut item_slice).ok()?;
if item_slice.len() < h.payload_len.into() as usize {
return None
}
item_slice = &item_slice[..h.payload_len.into() as usize];
Some((pos, h.is_processed, item_slice))
}
/// Set the `is_processed` flag for the item at `pos` to be `true` if not already and decrement
/// the `remaining` counter of the page.
///
/// Does nothing if no [`ItemHeader`] could be decoded at the given position.
fn note_processed_at_pos(&mut self, pos: usize) {
if let Ok(mut h) = ItemHeader::<Size>::decode(&mut &self.heap[pos..]) {
if !h.is_processed {
h.is_processed = true;
h.using_encoded(|d| self.heap[pos..pos + d.len()].copy_from_slice(d));
self.remaining.saturating_dec();
self.remaining_size.saturating_reduce(h.payload_len);
}
}
}
/// Returns whether the page is *complete* which means that no messages remain.
fn is_complete(&self) -> bool {
self.remaining.is_zero()
}
}
/// A single link in the double-linked Ready Ring list.
#[derive(Clone, Encode, Decode, MaxEncodedLen, TypeInfo, RuntimeDebug, PartialEq)]
pub struct Neighbours<MessageOrigin> {
/// The previous queue.
prev: MessageOrigin,
/// The next queue.
next: MessageOrigin,
}
/// The state of a queue as represented by a book of its pages.
///
/// Each queue has exactly one book which holds all of its pages. All pages of a book combined
/// contain all of the messages of its queue; hence the name *Book*.
/// Books can be chained together in a double-linked fashion through their `ready_neighbours` field.
#[derive(Clone, Encode, Decode, MaxEncodedLen, TypeInfo, RuntimeDebug)]
pub struct BookState<MessageOrigin> {
/// The first page with some items to be processed in it. If this is `>= end`, then there are
/// no pages with items to be processing in them.
begin: PageIndex,
/// One more than the last page with some items to be processed in it.
end: PageIndex,
/// The number of pages stored at present.
///
/// This might be larger than `end-begin`, because we keep pages with unprocessed overweight
/// messages outside of the end/begin window.
count: PageIndex,
/// If this book has any ready pages, then this will be `Some` with the previous and next
/// neighbours. This wraps around.
ready_neighbours: Option<Neighbours<MessageOrigin>>,
/// The number of unprocessed messages stored at present.
message_count: u64,
/// The total size of all unprocessed messages stored at present.
size: u64,
}
impl<MessageOrigin> Default for BookState<MessageOrigin> {
fn default() -> Self {
Self { begin: 0, end: 0, count: 0, ready_neighbours: None, message_count: 0, size: 0 }
}
}
/// Handler code for when the items in a queue change.
pub trait OnQueueChanged<Id> {
/// Note that the queue `id` now has `item_count` items in it, taking up `items_size` bytes.
fn on_queue_changed(id: Id, items_count: u64, items_size: u64);
}
impl<Id> OnQueueChanged<Id> for () {
fn on_queue_changed(_: Id, _: u64, _: u64) {}
}
#[frame_support::pallet]
pub mod pallet {
use super::*;
#[pallet::pallet]
#[pallet::generate_store(pub(super) trait Store)]
pub struct Pallet<T>(_);
/// The module configuration trait.
#[pallet::config]
pub trait Config: frame_system::Config {
/// The overarching event type.
type RuntimeEvent: From<Event<Self>> + IsType<<Self as frame_system::Config>::RuntimeEvent>;
/// Weight information for extrinsics in this pallet.
type WeightInfo: WeightInfo;
/// Processor for a message.
///
/// Must be set to [`mock_helpers::NoopMessageProcessor`] for benchmarking.
/// Other message processors that consumes exactly (1, 1) weight for any give message will
/// work as well. Otherwise the benchmarking will also measure the weight of the message
/// processor, which is not desired.
type MessageProcessor: ProcessMessage;
/// Page/heap size type.
type Size: BaseArithmetic
+ Unsigned
+ Copy
+ Into<u32>
+ Member
+ Encode
+ Decode
+ MaxEncodedLen
+ TypeInfo
+ Default;
/// Code to be called when a message queue changes - either with items introduced or
/// removed.
type QueueChangeHandler: OnQueueChanged<<Self::MessageProcessor as ProcessMessage>::Origin>;
/// The size of the page; this implies the maximum message size which can be sent.
///
/// A good value depends on the expected message sizes, their weights, the weight that is
/// available for processing them and the maximal needed message size. The maximal message
/// size is slightly lower than this as defined by [`MaxMessageLenOf`].
#[pallet::constant]
type HeapSize: Get<Self::Size>;
/// The maximum number of stale pages (i.e. of overweight messages) allowed before culling
/// can happen. Once there are more stale pages than this, then historical pages may be
/// dropped, even if they contain unprocessed overweight messages.
#[pallet::constant]
type MaxStale: Get<u32>;
/// The amount of weight (if any) which should be provided to the message queue for
/// servicing enqueued items.
///
/// This may be legitimately `None` in the case that you will call
/// `ServiceQueues::service_queues` manually.
#[pallet::constant]
type ServiceWeight: Get<Option<Weight>>;
}
#[pallet::event]
#[pallet::generate_deposit(pub(super) fn deposit_event)]
pub enum Event<T: Config> {
/// Message discarded due to an inability to decode the item. Usually caused by state
/// corruption.
Discarded { hash: T::Hash },
/// Message discarded due to an error in the `MessageProcessor` (usually a format error).
ProcessingFailed { hash: T::Hash, origin: MessageOriginOf<T>, error: ProcessMessageError },
/// Message is processed.
Processed { hash: T::Hash, origin: MessageOriginOf<T>, weight_used: Weight, success: bool },
/// Message placed in overweight queue.
OverweightEnqueued {
hash: T::Hash,
origin: MessageOriginOf<T>,
page_index: PageIndex,
message_index: T::Size,
},
/// This page was reaped.
PageReaped { origin: MessageOriginOf<T>, index: PageIndex },
}
#[pallet::error]
pub enum Error<T> {
/// Page is not reapable because it has items remaining to be processed and is not old
/// enough.
NotReapable,
/// Page to be reaped does not exist.
NoPage,
/// The referenced message could not be found.
NoMessage,
/// The message was already processed and cannot be processed again.
AlreadyProcessed,
/// The message is queued for future execution.
Queued,
/// There is temporarily not enough weight to continue servicing messages.
InsufficientWeight,
/// This message is temporarily unprocessable.
///
/// Such errors are expected, but not guaranteed, to resolve themselves eventually through
/// retrying.
TemporarilyUnprocessable,
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
}
/// The index of the first and last (non-empty) pages.
#[pallet::storage]
pub(super) type BookStateFor<T: Config> =
StorageMap<_, Twox64Concat, MessageOriginOf<T>, BookState<MessageOriginOf<T>>, ValueQuery>;
/// The origin at which we should begin servicing.
#[pallet::storage]
pub(super) type ServiceHead<T: Config> = StorageValue<_, MessageOriginOf<T>, OptionQuery>;
/// The map of page indices to pages.
#[pallet::storage]
pub(super) type Pages<T: Config> = StorageDoubleMap<
_,
Twox64Concat,
MessageOriginOf<T>,
Twox64Concat,
PageIndex,
Page<T::Size, T::HeapSize>,
OptionQuery,
>;
#[pallet::hooks]
impl<T: Config> Hooks<BlockNumberFor<T>> for Pallet<T> {
fn on_initialize(_n: BlockNumberFor<T>) -> Weight {
if let Some(weight_limit) = T::ServiceWeight::get() {
Self::service_queues(weight_limit)
} else {
Weight::zero()
}
}
/// Check all assumptions about [`crate::Config`].
fn integrity_test() {
assert!(!MaxMessageLenOf::<T>::get().is_zero(), "HeapSize too low");
}
}
#[pallet::call]
impl<T: Config> Pallet<T> {
/// Remove a page which has no more messages remaining to be processed or is stale.
#[pallet::weight(T::WeightInfo::reap_page())]
pub fn reap_page(
origin: OriginFor<T>,
message_origin: MessageOriginOf<T>,
page_index: PageIndex,
) -> DispatchResult {
let _ = ensure_signed(origin)?;
Self::do_reap_page(&message_origin, page_index)
}
/// Execute an overweight message.
///
/// Temporary processing errors will be propagated whereas permanent errors are treated
/// as success condition.
///
/// - `origin`: Must be `Signed`.
/// - `message_origin`: The origin from which the message to be executed arrived.
/// - `page`: The page in the queue in which the message to be executed is sitting.
/// - `index`: The index into the queue of the message to be executed.
/// - `weight_limit`: The maximum amount of weight allowed to be consumed in the execution
/// of the message.
///
/// Benchmark complexity considerations: O(index + weight_limit).
#[pallet::weight(
T::WeightInfo::execute_overweight_page_updated().max(
T::WeightInfo::execute_overweight_page_removed()).saturating_add(*weight_limit)
)]
pub fn execute_overweight(
origin: OriginFor<T>,
message_origin: MessageOriginOf<T>,
page: PageIndex,
index: T::Size,
weight_limit: Weight,
) -> DispatchResultWithPostInfo {
let _ = ensure_signed(origin)?;
let actual_weight =
Self::do_execute_overweight(message_origin, page, index, weight_limit)?;
Ok(Some(actual_weight).into())
}
}
}
/// The status of a page after trying to execute its next message.
#[derive(PartialEq, Debug)]
enum PageExecutionStatus {
/// The execution bailed because there was not enough weight remaining.
Bailed,
/// The page did not make any progress on its execution.
///
/// This is a transient condition and can be handled by retrying - exactly like [Bailed].
NoProgress,
/// No more messages could be loaded. This does _not_ imply `page.is_complete()`.
///
/// The reasons for this status are:
/// - The end of the page is reached but there could still be skipped messages.
/// - The storage is corrupted.
NoMore,
}
/// The status after trying to execute the next item of a [`Page`].
#[derive(PartialEq, Debug)]
enum ItemExecutionStatus {
/// The execution bailed because there was not enough weight remaining.
Bailed,
/// The item did not make any progress on its execution.
///
/// This is a transient condition and can be handled by retrying - exactly like [Bailed].
NoProgress,
/// The item was not found.
NoItem,
/// Whether the execution of an item resulted in it being processed.
///
/// One reason for `false` would be permanently overweight.
Executed(bool),
}
/// The status of an attempt to process a message.
#[derive(PartialEq)]
enum MessageExecutionStatus {
/// There is not enough weight remaining at present.
InsufficientWeight,
/// There will never be enough weight.
Overweight,
/// The message was processed successfully.
Processed,
/// The message was processed and resulted in a, possibly permanent, error.
Unprocessable { permanent: bool },
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
}
impl<T: Config> Pallet<T> {
/// Knit `origin` into the ready ring right at the end.
///
/// Return the two ready ring neighbours of `origin`.
fn ready_ring_knit(origin: &MessageOriginOf<T>) -> Result<Neighbours<MessageOriginOf<T>>, ()> {
if let Some(head) = ServiceHead::<T>::get() {
let mut head_book_state = BookStateFor::<T>::get(&head);
let mut head_neighbours = head_book_state.ready_neighbours.take().ok_or(())?;
let tail = head_neighbours.prev;
head_neighbours.prev = origin.clone();
head_book_state.ready_neighbours = Some(head_neighbours);
BookStateFor::<T>::insert(&head, head_book_state);
let mut tail_book_state = BookStateFor::<T>::get(&tail);
let mut tail_neighbours = tail_book_state.ready_neighbours.take().ok_or(())?;
tail_neighbours.next = origin.clone();
tail_book_state.ready_neighbours = Some(tail_neighbours);
BookStateFor::<T>::insert(&tail, tail_book_state);
Ok(Neighbours { next: head, prev: tail })
} else {
ServiceHead::<T>::put(origin);
Ok(Neighbours { next: origin.clone(), prev: origin.clone() })
}
}
fn ready_ring_unknit(origin: &MessageOriginOf<T>, neighbours: Neighbours<MessageOriginOf<T>>) {
if origin == &neighbours.next {
debug_assert!(
origin == &neighbours.prev,
"unknitting from single item ring; outgoing must be only item"
);
// Service queue empty.
ServiceHead::<T>::kill();
} else {
BookStateFor::<T>::mutate(&neighbours.next, |book_state| {
if let Some(ref mut n) = book_state.ready_neighbours {
n.prev = neighbours.prev.clone()
}
});
BookStateFor::<T>::mutate(&neighbours.prev, |book_state| {
if let Some(ref mut n) = book_state.ready_neighbours {
n.next = neighbours.next.clone()
}
});
if let Some(head) = ServiceHead::<T>::get() {
if &head == origin {
ServiceHead::<T>::put(neighbours.next);
}
} else {
defensive!("`ServiceHead` must be some if there was a ready queue");
}
}
}
/// Tries to bump the current `ServiceHead` to the next ready queue.
///
/// Returns the current head if it got be bumped and `None` otherwise.
fn bump_service_head(weight: &mut WeightMeter) -> Option<MessageOriginOf<T>> {
if !weight.check_accrue(T::WeightInfo::bump_service_head()) {
return None
}
if let Some(head) = ServiceHead::<T>::get() {
let mut head_book_state = BookStateFor::<T>::get(&head);
if let Some(head_neighbours) = head_book_state.ready_neighbours.take() {
ServiceHead::<T>::put(&head_neighbours.next);
Some(head)
} else {
None
}
} else {
None
}
}
fn do_enqueue_message(
origin: &MessageOriginOf<T>,
message: BoundedSlice<u8, MaxMessageLenOf<T>>,
) {
let mut book_state = BookStateFor::<T>::get(origin);
book_state.message_count.saturating_inc();
book_state
.size
// This should be payload size, but here the payload *is* the message.
.saturating_accrue(message.len() as u64);
if book_state.end > book_state.begin {
debug_assert!(book_state.ready_neighbours.is_some(), "Must be in ready ring if ready");
// Already have a page in progress - attempt to append.
let last = book_state.end - 1;
let mut page = match Pages::<T>::get(origin, last) {
Some(p) => p,
None => {
defensive!("Corruption: referenced page doesn't exist.");
return
},
};
if page.try_append_message::<T>(message).is_ok() {
Pages::<T>::insert(origin, last, &page);
BookStateFor::<T>::insert(origin, book_state);
return
}
} else {
debug_assert!(
book_state.ready_neighbours.is_none(),
"Must not be in ready ring if not ready"
);
// insert into ready queue.
match Self::ready_ring_knit(origin) {
Ok(neighbours) => book_state.ready_neighbours = Some(neighbours),
Err(()) => {
defensive!("Ring state invalid when knitting");
},
}
}
// No room on the page or no page - link in a new page.
book_state.end.saturating_inc();
book_state.count.saturating_inc();
let page = Page::from_message::<T>(message);
Pages::<T>::insert(origin, book_state.end - 1, page);
// NOTE: `T::QueueChangeHandler` is called by the caller.
BookStateFor::<T>::insert(origin, book_state);
}
/// Try to execute a single message that was marked as overweight.
///
/// The `weight_limit` is the weight that can be consumed to execute the message. The base
/// weight of the function it self must be measured by the caller.
pub fn do_execute_overweight(
origin: MessageOriginOf<T>,
page_index: PageIndex,
index: T::Size,
weight_limit: Weight,
) -> Result<Weight, Error<T>> {
let mut book_state = BookStateFor::<T>::get(&origin);
let mut page = Pages::<T>::get(&origin, page_index).ok_or(Error::<T>::NoPage)?;
let (pos, is_processed, payload) =
page.peek_index(index.into() as usize).ok_or(Error::<T>::NoMessage)?;
let payload_len = payload.len() as u64;
ensure!(
page_index < book_state.begin ||
(page_index == book_state.begin && pos < page.first.into() as usize),
Error::<T>::Queued
);
ensure!(!is_processed, Error::<T>::AlreadyProcessed);
use MessageExecutionStatus::*;
let mut weight_counter = WeightMeter::from_limit(weight_limit);
match Self::process_message_payload(
origin.clone(),
page_index,
index,
payload,
&mut weight_counter,
Weight::MAX,
// ^^^ We never recognise it as permanently overweight, since that would result in an
// additional overweight event being deposited.
) {
Overweight | InsufficientWeight => Err(Error::<T>::InsufficientWeight),
Unprocessable { permanent: false } => Err(Error::<T>::TemporarilyUnprocessable),
Unprocessable { permanent: true } | Processed => {
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
page.note_processed_at_pos(pos);
book_state.message_count.saturating_dec();
book_state.size.saturating_reduce(payload_len);
let page_weight = if page.remaining.is_zero() {
debug_assert!(
page.remaining_size.is_zero(),
"no messages remaining; no space taken; qed"
);
Pages::<T>::remove(&origin, page_index);
debug_assert!(book_state.count >= 1, "page exists, so book must have pages");
book_state.count.saturating_dec();
T::WeightInfo::execute_overweight_page_removed()
// no need to consider .first or ready ring since processing an overweight page
// would not alter that state.
} else {
Pages::<T>::insert(&origin, page_index, page);
T::WeightInfo::execute_overweight_page_updated()
};
BookStateFor::<T>::insert(&origin, &book_state);
T::QueueChangeHandler::on_queue_changed(
origin,
book_state.message_count,
book_state.size,
);
Ok(weight_counter.consumed.saturating_add(page_weight))
},
}
}
/// Remove a stale page or one which has no more messages remaining to be processed.
fn do_reap_page(origin: &MessageOriginOf<T>, page_index: PageIndex) -> DispatchResult {
let mut book_state = BookStateFor::<T>::get(origin);
// definitely not reapable if the page's index is no less than the `begin`ning of ready
// pages.
ensure!(page_index < book_state.begin, Error::<T>::NotReapable);
let page = Pages::<T>::get(origin, page_index).ok_or(Error::<T>::NoPage)?;
// definitely reapable if the page has no messages in it.
let reapable = page.remaining.is_zero();
// also reapable if the page index has dropped below our watermark.
let cullable = || {
let total_pages = book_state.count;
let ready_pages = book_state.end.saturating_sub(book_state.begin).min(total_pages);
// The number of stale pages - i.e. pages which contain unprocessed overweight messages.
// We would prefer to keep these around but will restrict how far into history they can
// extend if we notice that there's too many of them.
//
// We don't know *where* in history these pages are so we use a dynamic formula which
// reduces the historical time horizon as the stale pages pile up and increases it as
// they reduce.
let stale_pages = total_pages - ready_pages;
// The maximum number of stale pages (i.e. of overweight messages) allowed before
// culling can happen at all. Once there are more stale pages than this, then historical
// pages may be dropped, even if they contain unprocessed overweight messages.
let max_stale = T::MaxStale::get();
// The amount beyond the maximum which are being used. If it's not beyond the maximum
// then we exit now since no culling is needed.
let overflow = match stale_pages.checked_sub(max_stale + 1) {
Some(x) => x + 1,
None => return false,
};
// The special formula which tells us how deep into index-history we will pages. As
// the overflow is greater (and thus the need to drop items from storage is more urgent)
// this is reduced, allowing a greater range of pages to be culled.
// With a minimum `overflow` (`1`), this returns `max_stale ** 2`, indicating we only
// cull beyond that number of indices deep into history.
// At this overflow increases, our depth reduces down to a limit of `max_stale`. We
// never want to reduce below this since this will certainly allow enough pages to be
// culled in order to bring `overflow` back to zero.
let backlog = (max_stale * max_stale / overflow).max(max_stale);
let watermark = book_state.begin.saturating_sub(backlog);
page_index < watermark
};
ensure!(reapable || cullable(), Error::<T>::NotReapable);
Pages::<T>::remove(origin, page_index);
debug_assert!(book_state.count > 0, "reaping a page implies there are pages");
book_state.count.saturating_dec();
book_state.message_count.saturating_reduce(page.remaining.into() as u64);
book_state.size.saturating_reduce(page.remaining_size.into() as u64);
BookStateFor::<T>::insert(origin, &book_state);
T::QueueChangeHandler::on_queue_changed(
origin.clone(),
book_state.message_count,
book_state.size,
);
Self::deposit_event(Event::PageReaped { origin: origin.clone(), index: page_index });
Ok(())
}
/// Execute any messages remaining to be processed in the queue of `origin`, using up to
/// `weight_limit` to do so. Any messages which would take more than `overweight_limit` to
/// execute are deemed overweight and ignored.
fn service_queue(
origin: MessageOriginOf<T>,
weight: &mut WeightMeter,
overweight_limit: Weight,
) -> (bool, Option<MessageOriginOf<T>>) {
use PageExecutionStatus::*;
if !weight.check_accrue(
T::WeightInfo::service_queue_base().saturating_add(T::WeightInfo::ready_ring_unknit()),
) {
return (false, None)
}
let mut book_state = BookStateFor::<T>::get(&origin);
let mut total_processed = 0;
while book_state.end > book_state.begin {
let (processed, status) =
Self::service_page(&origin, &mut book_state, weight, overweight_limit);
total_processed.saturating_accrue(processed);
match status {
// Store the page progress and do not go to the next one.
Bailed | NoProgress => break,
// Go to the next page if this one is at the end.
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
};
book_state.begin.saturating_inc();
}
let next_ready = book_state.ready_neighbours.as_ref().map(|x| x.next.clone());
if book_state.begin >= book_state.end && total_processed > 0 {
// No longer ready - unknit.
if let Some(neighbours) = book_state.ready_neighbours.take() {
Self::ready_ring_unknit(&origin, neighbours);
} else {
defensive!("Freshly processed queue must have been ready");
}
}
BookStateFor::<T>::insert(&origin, &book_state);
if total_processed > 0 {
T::QueueChangeHandler::on_queue_changed(
origin,
book_state.message_count,
book_state.size,
);
}
(total_processed > 0, next_ready)
}
/// Service as many messages of a page as possible.
///
/// Returns how many messages were processed and the page's status.
fn service_page(
origin: &MessageOriginOf<T>,
book_state: &mut BookStateOf<T>,
weight: &mut WeightMeter,
overweight_limit: Weight,
) -> (u32, PageExecutionStatus) {
use PageExecutionStatus::*;
if !weight.check_accrue(
T::WeightInfo::service_page_base_completion()
.max(T::WeightInfo::service_page_base_no_completion()),
) {
return (0, Bailed)
}
let page_index = book_state.begin;