Store, Table-Based Introspectable Storage
Abstract
This standard introduces a flexible on-chain storage pattern that organizes data into structured tables that consist of records with fixed key and value schemas, similar to a traditional database. This storage pattern consists of a unified contract interface for data access, along with a compact binary encoding format for both static and dynamic data types. State changes are tracked through standardized events that enable automatic, schema-aware state replication by off-chain indexers. New tables can be dynamically registered at runtime through a special table that stores schema metadata for all tables, allowing the system to evolve without breaking existing contracts or integrations.
Motivation
The absence of consistent standards for on-chain data management in smart contracts can lead to rigid implementations, tightly coupled contract logic with off-chain services, and challenges in updating or extending a contract’s data layout without breaking existing integrations.
Using the storage mechanism defined in this ERC provides the following benefits:
- Automatic Indexing: By emitting consistent, standardized events during state changes, off-chain services can automatically track on-chain state and provide schema-aware indexer APIs.
- Elimination of Custom Getter Functions: Any contract or off-chain service can read stored data through a consistent interface, decoupling smart contract implementation from specific data access patterns and reducing development overhead.
- Simpler Upgradability: This pattern leverages unstructured storage, making it easier to upgrade contract logic without the risks associated with using a fixed storage layout.
- Flexible Data Extensions: New tables can be added at runtime without without breaking existing integrations with other data consumers.
- Reduced gas costs: Using efficient data packing reduces gas costs for both storage and event emissions.
Specification
Definitions
Store
A smart contract that implements the interface proposed by this ERC and organizes data in Tables. It emits events for each data operation so that off-chain components can replicate the state of all tables.
Table
A storage structure that holds Records sharing the same Schema.
- On-chain Table: Stores its state on-chain and emits events for off-chain- indexers.
- Off-chain Table: Does not store state on-chain but emits events for off-chain indexers.
Record
A piece of data stored in a Table, addressed by one or more keys.
ResourceId
A 32-byte value that uniquely identifies each Table within the Store.
type ResourceId is bytes32;
Encoding:
Bytes (from left to right) | Description |
---|---|
0-1 | Table type identifier |
2-31 | Unique identifier |
Table Type Identifiers:
0x7462
("tb"
) for on-chain tables0x6f74
("ot"
) for off-chain tables
Schema
Used to represent the layout of Records within a table.
type Schema is bytes32;
Each Table defines two schemas:
- Key Schema: the types of the keys used to uniquely identify a Record within a table. It consists only of fixed-length data types.
- Value Schema: the types of the value fields of a Record within a table, which can include both fixed-length and variable-length data types.
Byte(s) from left to right | Value | Constraint |
---|---|---|
0-1 | Total byte length of static fields | |
2 | Number of static length fields | ≤ (28 - number of dynamic length fields) |
3 | Number of dynamic length fields | For the key schema, 0 |
For the value schema, ≤5 | ||
4-31 | Each byte encodes a SchemaType |
Dynamic-length types MUST come after all the static-length types. |
SchemaType
Single byte that represents the type of a specific static or dynamic field.
enum SchemaType { ... }
Type Encoding:
Value Range | Type |
---|---|
0x00 to 0x1F |
uint8 to uint256 (increments of 8 bits) |
0x20 to 0x3F |
int8 to int256 (increments of 8 bits) |
0x40 to 0x5F |
bytes1 to bytes32 |
0x60 |
bool |
0x61 |
address |
0x62 to 0x81 |
uint8[] to uint256[] |
0x82 to 0xA1 |
int8[] to int256[] |
0xA2 to 0xC1 |
bytes1[] to bytes32[] |
0xC2 |
bool[] |
0xC3 |
address[] |
0xC4 |
bytes |
0xC5 |
string |
FieldLayout
Encodes the concrete value Schema
information, specifically the total byte length of the static fields, the number of dynamic fields and the length of each static field on its own.
This encoding serves as an optimization for on-chain operations. By having the exact lengths readily available, the Store doesn’t need to repeatedly compute or translate the schema definitions into actual field lengths during execution.
type FieldLayout is bytes32;
Byte(s) from left to right | Value | Constraint |
---|---|---|
0-1 | Total length of static fields | |
2 | Number of static length fields | ≤ (28 - number of dynamic length fields) |
3 | Number of dynamic length fields | For the key schema, 0 |
For the value schema, ≤5 | ||
4-31 | Each byte encodes the byte length of the corresponding static field |
EncodedLengths
Encodes the byte length of all the dynamic fields of a specific Record. It is returned by the Store methods when reading a Record, as it is needed for decoding dynamic fields.
type EncodedLengths is bytes32;
Bytes (from least to most significant) | Type | Description |
---|---|---|
0x00-0x06 | uint56 | Total byte length of dynamic data |
0x07-0xB | uint40 | Length of the first dynamic field |
0x0C-0x10 | uint40 | Length of the second dynamic field |
0x11-0x15 | uint40 | Length of the third dynamic field |
0x16-0x1A | uint40 | Length of the fourth dynamic field |
0x1B-0x1F | uint40 | Length of the fifth dynamic field |
Packed Data Encoding
Record data returned by Store methods and included in Store events uses the following encoding rules.
Field Limits
- Maximum Total Fields: A record can contain up to 28 fields in total (both static and dynamic fields combined).
- This limit is due to the
Schema
type structure, which uses 28 bytes (bytes 4 to 31) to define field types, with one byte per field (SchemaType
).
- This limit is due to the
- Dynamic Fields Limit: A record can have up to 5 dynamic fields.
- This is due to the fact that a single 32 bytes word (
EncodedLengths
) to encode the byte lengths of each dynamic field, instead of encoding each length separately as Solidity’sabi.encode
would.
- This is due to the fact that a single 32 bytes word (
- Static Fields Limit: The maximum number of static fields is 28 minus the number of dynamic fields.
- For example, if there are 5 dynamic fields, the maximum number of static fields is 23 (28 - 5).
Encoding Rules
- Static-length fields are encoded without any padding, and concatenated in the order they are defined in the schema, which is equivalent to using Solidity’s
abi.encodePacked
. - For dynamic-length fields (arrays,
bytes
, andstring
s):- If the field is an array, its elements are tightly packed without padding.
- All dynamic fields are concatenated together without padding and without including their lengths.
- The lengths of all dynamic fields are encoded into a single
EncodedLengths
.
Example
Suppose a table has the following value schema:
(uint256 id, address owner, string description, uint8[] scores)
Encoding (Pseudocode):
bytes memory staticData = abi.encodePacked(id, owner);
// This is a custom function as Solidity does not provide a way to tightly pack array elements
bytes memory packedScores = packElementsWithoutPadding(scores);
// abi.encodePacked concatenates both description and packedScores without including their lengths
bytes memory dynamicData = abi.encodePacked(description, packedScores);
// Total length is encoded in the 56 least significant bits
EncodedLengths encodedLengths = dynamicData.length;
// Each length is encoded using 5 bytes
encodedLengths |= (description.length << (56));
encodedLengths |= (encodedData.length << (56 + 8 * 5));
// The full encoded record data is represented by the following tuple:
// (staticData, encodedLengths, dynamicData)
Store Interface
All Stores MUST implement the following interface.
interface IStore {
/**
* Get full encoded record (all fields, static and dynamic data) for the given tableId and key tuple.
*/
function getRecord(
ResourceId tableId,
bytes32[] calldata keyTuple
) external view returns (bytes memory staticData, EncodedLengths encodedLengths, bytes memory dynamicData);
/**
* Get a single encoded field from the given tableId and key tuple.
*/
function getField(
ResourceId tableId,
bytes32[] calldata keyTuple,
uint8 fieldIndex
) external view returns (bytes memory data);
/**
* Get the byte length of a single field from the given tableId and key tuple
*/
function getFieldLength(
ResourceId tableId,
bytes32[] memory keyTuple,
uint8 fieldIndex
) external view returns (uint256);
}
The return values of both getRecord
and getField
use the encoding rules previously defined in the Packed Data Encoding section. More specifically, getRecord
returns the fully encoded record tuple, and the data returned by getField
is encoded using the encoding rules as if the field was being encoded on its own.
Store Operations and Events
This standard defines three core operations for manipulating records in a table: setting, updating, and deleting. For each operation, specific events must be emitted. The implementation details of these operations are left to the discretion of each Store implementation.
The fundamental requirement is that for on-chain tables the Record data retrieved through the Store interface methods at any given block MUST be consistent with the Record data that would be obtained by applying the operations implied by the Store events up to that block. This ensures data integrity and allows for accurate off-chain state reconstruction.
Store_SetRecord
Setting a Record means overwriting all of its fields. This operation can be performed whether the record has been set before or not (the standard does not enforce existence checks).
The Store_SetRecord
event MUST be emitted whenever the full data of a record has been overwritten.
event Store_SetRecord(
ResourceId indexed tableId,
bytes32[] keyTuple,
bytes staticData,
EncodedLengths encodedLengths,
bytes dynamicData
);
Parameters:
Name | Type | Description |
---|---|---|
tableId | ResourceId | The ID of the table where the record is set |
keyTuple | bytes32[] | An array representing the composite key for the record |
staticData | bytes | The static data of the record using packed encoding |
encodedLengths | EncodedLengths | The encoded lengths of the dynamic data of the record |
dynamicData | bytes | The dynamic data of the record, using custom packed encoding |
Store_SpliceStaticData
Splicing the static data of a Record consists in overwriting bytes of the packed encoded static fields. The total length of static data does not change as it is determined by the table’s value schema.
The Store_SpliceStaticData
event MUST be emitted whenever the static data of the Record has been spliced.
event Store_SpliceStaticData(
ResourceId indexed tableId,
bytes32[] keyTuple,
uint48 start,
bytes data
);
Parameters:
Name | Type | Description |
---|---|---|
tableId | ResourceId | The ID of the table where the data is spliced |
keyTuple | bytes32[] | An array representing the key for the record |
start | uint48 | The start position in bytes for the splice operation |
data | bytes | Packed ABI encoding of a tuple with the value’s static fields |
Store_SpliceDynamicData
Splicing the dynamic data of a Record involves modifying the packed encoded representation of its dynamic fields by removing, replacing, and/or inserting new bytes in place.
The Store_SpliceDynamicData
event MUST be emitted whenever the dynamic data of the Record has been spliced.
event Store_SpliceDynamicData(
ResourceId indexed tableId,
bytes32[] keyTuple,
uint8 dynamicFieldIndex,
uint48 start,
uint40 deleteCount,
EncodedLengths encodedLengths,
bytes data
);
Parameters:
Name | Type | Description |
---|---|---|
tableId | ResourceId | The ID of the table where the data is spliced |
keyTuple | bytes32[] | An array representing the composite key for the record |
dynamicFieldIndex | uint8 | The index of the dynamic field to splice data, relative to the start of the dynamic fields (Dynamic field index = field index - number of static fields) |
start | uint48 | The start position in bytes for the splice operation |
deleteCount | uint40 | The number of bytes to delete in the splice operation |
encodedLengths | EncodedLengths | The resulting encoded lengths of the dynamic data of the record |
data | bytes | The data to insert into the dynamic data of the record at the start byte |
Store_DeleteRecord
The Store_DeleteRecord
event MUST be emitted whenever the Record has been deleted from the Table.
event Store_DeleteRecord(ResourceId indexed tableId, bytes32[] keyTuple);
Parameters:
Name | Type | Description |
---|---|---|
tableId | ResourceId | The ID of the table where the record is deleted |
keyTuple | bytes32[] | An array representing the composite key for the record |
See the reference implementation section for an example on how to index store events.
The Tables
table
To keep track of the information of each table and support registering new tables at runtime, the Store implementation MUST include a special on-chain Tables
table, which behaves the same way as other on-chain tables except for the special constraints mentioned below.
The Tables
table MUST use the following Schema
s:
- Key Schema:
tableId
(ResourceId
):ResourceId
of the table this record describes.
- Value Schema:
fieldLayout
(FieldLayout
): encodes the byte length of each static data type in the table.keySchema
(Schema
): represents the data types of the (composite) key of the table.valueSchema
(Schema
): represents the data types of the value fields of the table.abiEncodedKeyNames
(bytes
): ABI encoded string array of key names.abiEncodedFieldNames
(bytes
): ABI encoded string array of field names.
Records stored in the Tables
table are considered immutable:
- The
Store
MUST emit a singleStore_SetRecord
event for each table being registered. - The
Store
SHOULD NOT emit any otherStore
events for aTable
registered in theTables
table.
The Tables
table MUST store a record that describes itself before any other table is registered, emitting the corresponding Store_SetRecord
event. The record must use the following tableId
:
// First two bytes indicates that this is an on-chain table
// The next 30 bytes are the unique identifier for the Tables table
// bytes32("tb") | bytes32("store") >> (2 * 8) | bytes32("Tables") >> (2 * 8 + 14 * 8)
ResourceId tableId = ResourceId.wrap(0x746273746f72650000000000000000005461626c657300000000000000000000);
By using a predefined ResourceId
and Schema
for the Tables
table, off-chain indexers can interpret store events for all registered tables. This enables the development of advanced off-chain services that operate on structured data rather than raw encoded data like in the previous indexer implementation example.
Rationale
Splice Events
While the Store_SetRecord
event suffices for tracking the data of each record off-chain, including Splice
events (Store_SpliceStaticData
and Store_SpliceDynamicData
) allows for more efficient partial updates. When only a portion of a record changes, emitting a full SetRecord
event would be inefficient because the entire record data would need to be read from storage and emitted. Splice
events enable the store to emit only the minimal necessary data for the update, reducing gas consumption. This is particularly important for records with large dynamic fields, as the cost of updating them doesn’t grow with the field’s size.
Disallowing Arrays of Dynamic Types
Arrays of dynamic types (e.g., string[]
, bytes[]
) are intentionally not included as supported SchemaType
s. This restriction enforces a flat data schema, which simplifies the store implementation and enhances efficiency. If users need to store such data structures, they can model them using a separate table with a schema like { index: uint256, data: bytes }
, where each array element is represented as an individual record.
FieldLayout Optimization
Including the FieldLayout
in the Tables
schema provides an on-chain optimization by precomputing and storing the exact byte lengths of static fields. This eliminates the need to repeatedly compute field lengths and offsets during runtime, which can be gas-intensive. By having this information readily available, the store can perform storage operations more efficiently, while components reading from the store can retrieve it from the Tables
table to decode the corresponding records.
Special Tables
table
Including a special Tables
table provides significant benefits for off-chain indexers. While emitting events for table registration isn’t strictly necessary for basic indexers that operate on raw encoded data, doing so makes indexers aware of the schemas used by each table. This awareness enables the development of more advanced, schema-aware indexer APIs (e.g., SQL-like query capabilities), enhancing the utility and flexibility of off-chain data interactions.
By reusing existing Store abstractions for table registration, we also simplify the implementation and eliminate the need for additional, specific table registration events. Indexers can leverage the standard Store events to access schema information, ensuring consistency and reducing complexity.
Reference Implementation
Store Event Indexing
The following example shows how a simple in-memory indexer can use the Store events to replicate the Store state off-chain. It is important to note that this indexer operates over raw encoded data which is not that useful on its own, but can be improved as we will explain in the next section.
We use TypeScript for this example but it can easily be replicated with other languages.
type Hex = `0x${string}`;
type Record = {
staticData: Hex;
encodedLengths: Hex;
dynamicData: Hex;
};
const store = new Map<string, Record>();
// Create a key string from a table ID and key tuple to use in our store Map above
function storeKey(tableId: Hex, keyTuple: Hex[]): string {
return `${tableId}:${keyTuple.join(",")}`;
}
// Like `Array.splice`, but for strings of bytes
function bytesSplice(
data: Hex,
start: number,
deleteCount = 0,
newData: Hex = "0x"
): Hex {
const dataNibbles = data.replace(/^0x/, "").split("");
const newDataNibbles = newData.replace(/^0x/, "").split("");
return `0x${dataNibbles
.splice(start, deleteCount * 2)
.concat(newDataNibbles)
.join("")}`;
}
function bytesLength(data: Hex): number {
return data.replace(/^0x/, "").length / 2;
}
function processStoreEvent(log: StoreEvent) {
if (log.eventName === "Store_SetRecord") {
const key = storeKey(log.args.tableId, log.args.keyTuple);
// Overwrite all of the Record's fields
store.set(key, {
staticData: log.args.staticData,
encodedLengths: log.args.encodedLengths,
dynamicData: log.args.dynamicData,
});
} else if (log.eventName === "Store_SpliceStaticData") {
const key = storeKey(log.args.tableId, log.args.keyTuple);
const record = store.get(key) ?? {
staticData: "0x",
encodedLengths: "0x",
dynamicData: "0x",
};
// Splice the static field data of the Record
store.set(key, {
staticData: bytesSplice(
record.staticData,
log.args.start,
bytesLength(log.args.data),
log.args.data
),
encodedLengths: record.encodedLengths,
dynamicData: record.dynamicData,
});
} else if (log.eventName === "Store_SpliceDynamicData") {
const key = storeKey(log.args.tableId, log.args.keyTuple);
const record = store.get(key) ?? {
staticData: "0x",
encodedLengths: "0x",
dynamicData: "0x",
};
// Splice the dynamic field data of the Record
store.set(key, {
staticData: record.staticData,
encodedLengths: log.args.encodedLengths,
dynamicData: bytesSplice(
record.dynamicData,
log.args.start,
log.args.deleteCount,
log.args.data
),
});
} else if (log.eventName === "Store_DeleteRecord") {
const key = storeKey(log.args.tableId, log.args.keyTuple);
// Delete the whole Record
store.delete(key);
}
}
Security Considerations
Access Control
This standard only defines functions to read from the Store (getRecord
, getField
, and getFieldLength
). The methods for setting or modifying records in the store are left to each specific implementation. Therefore, implementations must provide appropriate access control mechanisms for writing to the store, tailored to their specific use cases.
On-Chain Data Accessibility
All data stored within a store is accessible not only off-chain but also on-chain by other smart contracts through the provided read functions (getRecord
, getField
, and getFieldLength
). This differs from the typical behavior of smart contracts, where internal storage variables are private by default and cannot be directly read by other contracts unless explicit getter functions are provided. Thus, developers must be mindful that any data stored in the store is openly accessible to other smart contracts.
Copyright
Copyright and related rights waived via CC0.