Skip to main content

SQL specification

The Tableland SQL language specification offers a unique set of SQL that is a subset of the SQLite language.


Author(s): @carsonfarmer, @brunocalza, @jsign

Synopsis

Tableland understands a small subset of the standard SQL language. It does omit many features while at the same time adding a few features of its own. This document attempts to describe precisely what parts of the SQL language Tableland does and does not support. A list of supported data types is also provided. The SQL language supported by Tableland is a subset of the SQLite SQL language specification (and as such, we borrow heavily from their documentation with attribution), with additional constraints specific to Tableland operations.

This general SQL specification is broken down into two core sub-documents (which are linked below). This specification is a living document, and as such, may be updated over time. Proposals for the addition of SQL language features and data types may be submitted by the Tableland community over time. These proposals will be evaluated for technical feasibility, utility to the community, and longer-term sustainability.

Table of Contents

Statement Types

The core Tableland SQL parser accepts an SQL statement list which is a semicolon-separated list of statements. Each SQL statement in the statement list is an instance of one of the following specific statement types. All other standard SQL statement types are unavailable (at the moment). Each statement type is associated with a well-known SQL command (see following sections). In general, the entire Tableland SQL API can be summarized in eight command/statement types: CREATE TABLE, ALTER TABLE, INSERT, UPDATE, DELETE, SELECT, GRANT, REVOKE.

⚠️ The statement and data types provided here are part of the official minimal Tableland SQL specification. Additional functionality may be available in practice. However, it is not recommended that developers rely on SQL features outside of this minimal specification in the long-term.

CREATE TABLE

The CREATE TABLE command is used to create a new table on Tableland. A CREATE TABLE command specifies the following attributes of the new table:

  • The name of the new table (table_name).
  • The name of each column in the table (column_name).
  • The declared type of each column in the table (data_type).
  • A default value or expression for each column in the table.
  • Optionally, a PRIMARY KEY for the table. Both single column and composite (multiple column) primary keys are supported.
  • A set of SQL constraints for the table. Tableland supports UNIQUE, NOT NULL, CHECK and PRIMARY KEY constraints (see previous bullet).
  • Optionally, a generated column constraint.

Structure

CREATE TABLE *table_name* ( [
{ *column_name* *data_type* [ *column_constraint* [, ... ] ]
| table_constraint }
[, ...]
] );

where column_constraint has structure

[ CONSTRAINT constraint_name ]
{ NOT NULL |
CHECK ( expression ) |
DEFAULT default_expr |
UNIQUE index_parameters |
PRIMARY KEY index_parameters |
}

and table_constraint has structure

[ CONSTRAINT constraint_name ]
{ CHECK ( expression ) |
UNIQUE ( column_name [, ... ] ) |
PRIMARY KEY ( column_name [, ... ] )

Details

Table Identifiers/Names

Every CREATE TABLE statement must specify a fully-qualified table name (name) as the name of the new table. The fully-qualified table name has the following structure:

tableName=prefix+chainId+tableId\mathtt{tableName} = \mathtt{prefix} + \mathtt{chainId} + \mathtt{tableId}

Where prefix\mathtt{prefix} is optional. When a prefix is included, it must start with a letter and be followed by any combination of (zero or more) letters, numbers, and/or underscores. A prefix string may be up to 32 bytes in length. In practice, long names with spaces must be slug-ified with underscores. For example, "my amazing table" would become "my_amazing_table". The last two components of the table name, must be the chain id and the table id, which are numeric values separated by an underscore. For example, a valid table name without a prefix looks like _42_0 (or 42_1), whereas a valid table name with a prefix might look like dogs_42_0.

⚠️ It is not up to the caller to determine what table id to use in a CREATE TABLE statement. The table id is a monotonically-increasing numeric value which is provided by the smart contract that is processing the create statements. See the Onchain API Specification for details on the smart contract calls used to generate CREATE TABLE statements in practice.

Table names are globally unique. The combination of chain id and monotonically increasing table id ensures this is the case in practice. As such, the addition of a user-defined prefix string is an aesthetic feature that most developers will find useful (but is not required). The maximum (slug-ified) prefix length is 32 bytes.

ℹ️ Tableland also supports quoted identifiers (for table names, column names, etc). This allows callers to use SQL Keywords (see next section) as part of identifiers, etc. There are some limitations to this, and it does not circumvent any other naming constraints.

Reserved Keywords

The SQL standard specifies a large number of keywords which may not be used as the names of tables, indices, columns, databases, or any other named object. The list of keywords is often so long that few people can remember them all. For most SQL code, your safest bet is to never use any English language word as the name of a user-defined object.

If you want to use a keyword as a name, you need to quote it. There are four ways of quoting keywords in SQLite:

  • 'keyword' — A keyword in single quotes is a string literal.
  • "keyword" — A keyword in double-quotes is an identifier.
  • [keyword] — A keyword enclosed in square brackets is an identifier. This is not standard SQL, it is included in Tableland for compatibility.
  • keyword — A keyword enclosed in grave accents (ASCII code 96) is an identifier. This is not standard SQL, it is included in Tableland for compatibility.

The list below shows all possible reserved keywords used by Tableland (or SQLite). Any identifier that is not on the following element list is not considered a keyword to the SQL parser in Tableland:

TRUE, FALSE, AND, OR, NOT, NULL, NONE, INTEGER, NUMERIC, REAL, TEXT, CAST, AS, IS, ISNULL, NOTNULL, COLLATE, LIKE, IN, REGEXP, GLOB, MATCH, ESCAPE, BETWEEN, CASE, WHEN, THEN, ELSE, END, SELECT, FROM, WHERE, GROUP, BY, HAVING, LIMIT, OFFSET, ORDER, ASC, DESC, NULLS, FIRST, LAST, DISTINCT, ALL, JOIN, ON, USING, EXISTS, FILTER, BLOB, INT, ANY, CREATE, TABLE, PRIMARY, KEY, UNIQUE, CHECK, DEFAULT, GENERATED, ALWAYS, STORED, VIRTUAL, CONSTRAINT, INSERT, VALUES, INTO, DELETE, UPDATE, SET, GRANT, TO, REVOKE, CONFLICT, DO, NOTHING

ℹ️ You can also find the most up to date list of keywords used by Tableland in the reference parser implementation. See Implementation

⚠️ Table names that begin with sqlite, system or registry are also reserved for internal use. It is an error to attempt to create a table with a name that starts with these reserved names.

Column Definitions and Constraints

Every CREATE TABLE statement includes one or more column definitions, optionally followed by a list of table constraints. Each column definition consists of the name of the column, followed by the declared type of the column (see Data Types), then one or more optional column constraints. Included in the definition of column constraints for the purposes of the previous statement is the DEFAULT clause, even though this is not really a constraint in the sense that it does not restrict the data that the table may contain. The other constraints, NOT NULL, CHECK, UNIQUE, and PRIMARY KEY constraints, impose restrictions on the table data.

⚠️ The number of columns in a table is limited by the MaxColumns validator configuration parameters (defaults to 24). A single character fields in a table cannot store more than MaxTextLength bytes of data (defaults to 1024). The number of rows in a table is limited by the MaxRowCount validator configuration parameter (defaults to 100,000). This values are all configurable at the network-level, and may change in the future.

⚠️ In practice, a CREATE TABLE statement must be sent as a single top-level statement (i.e., it must be provided in a statement list of length one).

🚧 Feature At Risk: FOREIGN KEY constraints of the form FOREIGN KEY(column_name) REFERENCES table_id(column_name) are currently not supported across Tableland tables. Instead, dynamic JOINs can be used to reference columns in remote tables. However, inclusion of FOREIGN KEY constraints are being considered for inclusion in the Tableland SQL specification with some specific limitations. In particular, key constraint actions would be restricted to SET NULL or SET DEFAULT (see the section called SQLite foreign key constraint actions at the link below). See SQLite Foreign Key

Column Defaults

The DEFAULT clause specifies a default value to use for the column if no value is explicitly provided by the user when doing an INSERT. If there is no explicit DEFAULT clause attached to a column definition, then the default value of the column is NULL. An explicit DEFAULT clause may specify that the default value is NULL, a string constant, a blob constant, a signed-number, or any constant expression enclosed in parentheses. For the purposes of the DEFAULT clause, an expression is considered constant if it contains no sub-queries, column, or table references, or string literals enclosed in double-quotes instead of single-quotes.

Each time a row is inserted into the table by an INSERT statement that does not provide explicit values for all table columns the values stored in the new row are determined by their default values, as follows:

  • If the default value of the column is a constant NULL, text, blob or signed-number value, then that value is used directly in the new row.
  • If the default value of a column is an expression in parentheses, then the expression is evaluated once for each row inserted and the results used in the new row.

Generated Columns

A column that includes a GENERATED ALWAYS AS clause is a generated column:

CREATE TABLE table_id (
...,
column_name data_type { GENERATED ALWAYS } AS (*expression*) { STORED | VIRTUAL }
);

Generated columns (also sometimes called "computed columns") are columns of a table whose values are a function of other columns in the same row. Generated columns can be read, but their values can not be directly written. The only way to change the value of a generated column is to modify the values of the other columns used to calculate the generated column.

The GENERATED ALWAYS keywords at the beginning of the constraint and the VIRTUAL or STORED keyword at the end are all optional. Only the AS keyword and the parenthesized expression are required. If the trailing VIRTUAL or STORED keyword is omitted, then VIRTUAL is the default.

The value of a VIRTUAL column is computed when read, whereas the value of a STORED column is computed when the row is written. STORED columns take up space in the database file, whereas VIRTUAL columns use more CPU cycles when being read.

Features and Limitations

  • Generated columns must also have a defined data type (just like all columns in Tableland). Tableland will attempt to transform the result of the generating expression into that data type using the same affinity rules as for ordinary columns.
  • Generated columns may have NOT NULL, CHECK, and UNIQUE constraints, just like ordinary columns.
  • The expression of a generated column can refer to any of the other declared columns in the table, including other generated columns, as long as the expression does not directly or indirectly refer back to itself.
  • Generated columns may not have a DEFAULT clause. The value of a generated column is always the value specified by the expression that follows the AS keyword.
  • Generated columns may not be used as part of the PRIMARY KEY.
  • The expression of a generated column may only reference constant literals and columns within the same row, and may only use scalar deterministic functions. The expression may not use sub-queries, aggregate functions, etc.
  • The expression of a generated column may refer to other generated columns in the same row, but no generated column can depend upon itself, either directly or indirectly.
  • Every table must have at least one non-generated column.
  • The data type of the generated column is determined only by the declared data type from the column definition. The datatype of the GENERATED ALWAYS AS expression has no affect on the data type of the column data itself.

Primary Key

Each table in Tableland may have at most one PRIMARY KEY. If the keywords PRIMARY KEY are added to a column definition, then the primary key for the table consists of that single column. Or, if a PRIMARY KEY clause is specified as a separate table constraint, then the primary key of the table consists of the list of columns specified as part of the PRIMARY KEY clause. The PRIMARY KEY clause must contain only column names. An error is raised if more than one PRIMARY KEY clause appears in a CREATE TABLE statement. The PRIMARY KEY is optional.

If a table has a single column primary key and the declared type of that column is INTEGER, then the column is known as an INTEGER PRIMARY KEY. See below for a description of the special properties and behaviors associated with an INTEGER PRIMARY KEY.

Each row in a table with a primary key must have a unique combination of values in its primary key columns. If an INSERT or UPDATE statement attempts to modify the table content so that two or more rows have identical primary key values, that is a constraint violation. Related, the SQL standard is that a PRIMARY KEY should always be NOT NULL, so Tableland enforces this constraint.

Integer Primary Key

All rows within Tableland tables have a 64-bit signed integer key that uniquely identifies the row within its table. This integer is usually called the ROWID. The ROWID value can be accessed using one of the special case-independent names "rowid", "oid", or "_rowid_" in place of a column name. As such, these values are not allowed as identifiers for columns in a CREATE TABLE statement.

The data for Tableland tables are stored in sorted order, by ROWID. This means that retrieving or sorting records by ROWID is fast. Searching for a record with a specific ROWID, or for all records with ROWIDs within a specified range is around twice as fast as a similar search made by specifying any other PRIMARY KEY. This ROWID sorting is also required for sub-queries and other SQL features on Tableland, to ensure deterministic ordering of results.

With one exception noted below, if a table has a primary key that consists of a single column and the declared type of that column is INTEGER, then the column becomes an alias for the ROWID. Such a column is usually referred to as an "integer primary key". A PRIMARY KEY column only becomes an integer primary key if the declared type name is exactly INTEGER. Other integer type names like INT causes the primary key column to behave as an ordinary table column with integer affinity and a unique index, not as an alias for the ROWID.

In the above case of an integer primary key, there is an additional implied AUTOINCREMENT constraint, which forces the integer primary key to behave as if it were specified with INTEGER PRIMARY KEY AUTOINCREMENT. See Autoincrement for further details.

The exception mentioned above is that if the declaration of a column with declared type INTEGER includes an PRIMARY KEY DESC clause, it does not become an alias for the ROWID and is not classified as an integer primary key. In practice, this means that the following three table declarations all cause the column "x" to be an alias for the ROWID (an integer primary key):

  • CREATE TABLE t(x INTEGER PRIMARY KEY ASC, y, z);
  • CREATE TABLE t(x INTEGER, y, z, PRIMARY KEY(x ASC));
  • CREATE TABLE t(x INTEGER, y, z, PRIMARY KEY(x DESC));

But the following declaration does not result in "x" being an alias for the ROWID:

  • CREATE TABLE t(x INTEGER PRIMARY KEY DESC, y, z);

Given this, and the implied autoincrement behavior, the following transformation rules are enforced by the Tableland Parser to maintain the correct ROWID alias behavior:

StatementTransformation
CREATE TABLE (a INTEGER PRIMARY KEY);Inject AUTOINCREMENT
CREATE TABLE (a INTEGER PRIMARY KEY DESC);Unchanged and not an alias
CREATE TABLE (a INTEGER, PRIMARY KEY(x ASC);Transformed to first row version with injected AUTOINCREMENT
CREATE TABLE (a INTEGER, PRIMARY KEY(x DESC);Transformed to second row version and no longer an alias

These transformations to the more "canonical" direct constraint on the primary key are required to enforce the implied AUTOINCREMENT behavior on the special integer primary keys.

Rowid values may not be modified using an UPDATE statement by attempting to assign to one of the built-in aliases ("rowid", "oid" or "_rowid_"). However, it is possible to UPDATE an integer primary key value (which is an alias to ROWID) by specifying a value directly. Similarly, an INSERT statement may be used to directly provide a value to use as the ROWID for any row inserted. For example, the following statements are allowed, and will update/set the ROWID value directly:

  • INSERT INTO a VALUES (2, 'Hello');
  • UPDATE a SET a = 10 WHERE b = 'Hello';

⚠️ If an UPDATE or INSERT sets a given ROWID to the largest possible value, then new INSERTs are not allowed and any attempt to insert a new row will fail with an error. As such, use caution when directly assigning values to a ROWID alias in the form of an integer primary key.

Autoincrement

In Tableland, a column with type INTEGER PRIMARY KEY is an alias for the ROWID which is always a 64-bit signed integer. It is implied that this column will behave as INTEGER PRIMARY KEY AUTOINCREMENT. This is a special feature of the Tableland SQL Specification, and helps to ensure deterministic ordering of values within a table.

ℹ️ While the AUTOINCREMENT keyword is implied with INTEGER PRIMARY KEY columns, the keyword itself is not allowed in this specification. Any attempt to use the AUTOINCREMENT keyword on any column results in an error.

On an INSERT, the ROWID or INTEGER PRIMARY KEY column will be filled automatically with a monotonically increasing integer value, usually one more than the largest ROWID currently in use.

In practice, this prevents the reuse of ROWIDs over the lifetime of the table. In other words, the purpose of the implied AUTOINCREMENT is to prevent the reuse of ROWIDs from previously deleted rows.

The ROWID chosen for the new row is at least one larger than the largest ROWID that has ever before existed in that same table. If the table has never before contained any data, then a ROWID of 1 is used. If the largest possible ROWID has previously been inserted, then new INSERTs are not allowed and any attempt to insert a new row will fail with an error. Only ROWID values from previous transactions that were committed are considered. ROWID values that were rolled back are ignored and can be reused.

Rows with automatically selected ROWIDs are guaranteed to have ROWIDs that have never been used before by the same table. And the automatically generated ROWIDs are guaranteed to be monotonically increasing. These are important properties for blockchain applications.

Note that "monotonically increasing" does not imply that the ROWID always increases by exactly one. One is the usual increment. However, if an insert fails due to (for example) a uniqueness constraint, the ROWID of the failed insertion attempt might not be reused on subsequent inserts, resulting in gaps in the ROWID sequence. Tableland guarantees that automatically chosen ROWIDs will be increasing but not that they will be sequential.

ALTER TABLE

The ALTER TABLE command allows the following alterations of an existing table: renaming a column, adding a column, and dropping a column.

Structure

ALTER TABLE table_name *action*

where action is one of:

-- For renaming a column
RENAME [ COLUMN ] *column_name* TO *new_column_name*

-- For adding a column
ADD [ COLUMN ] *column_name* *data_type* [ *column_constraint* [, ... ] ]

-- For dropping a dolumn
DROP [ COLUMN ] *column_name*

Details

The ADD COLUMN syntax is used to add a new column to an existing table. The new column is always appended to the end of the list of existing columns. The new column may take any of the forms permissible in a CREATE TABLE statement, with the following restrictions:

  • The column may not have a PRIMARY KEY or UNIQUE constraint.
  • If a NOT NULL constraint is specified, then the column must have a default value other than NULL.
  • The column may not be GENERATED ALWAYS ... STORED, though VIRTUAL columns are allowed.

The DROP COLUMN syntax is used to remove an existing column from a table. The DROP COLUMN command removes the named column from the table, and rewrites its content to purge the data associated with that column. The DROP COLUMN command only works if the column is not referenced by any other parts of the schema and is not a PRIMARY KEY and does not have a UNIQUE constraint. Possible reasons why the DROP COLUMN command can fail include:

  • The column is a PRIMARY KEY or part of one.
  • The column has a UNIQUE constraint.
  • The column is named in a table or column CHECK constraint not associated with the column being dropped.
  • The column is used in the expression of a generated column.

DELETE

The DELETE command removes records from the table identified by the table id.

Structure

DELETE FROM table_name [ WHERE condition ]

Details

If the WHERE clause is not present, all records in the table are deleted. If a WHERE clause is supplied, then only those rows for which the WHERE clause boolean expression is true are deleted. Rows for which the expression is false or NULL are retained.

INSERT

The INSERT command creates new rows in a table identified by the table name.

Structure

INSERT INTO table_name [ ( *column_name* [, ...] ) ] VALUES (
{ expression } [, ...]
);

or

INSERT INTO table_name DEFAULT VALUES;

or, the following limited sub-query syntax

INSERT INTO table_name [ ( *column_name* [, ...] ) ] SELECT [ * | expression [, ...] ]
[ FROM from_clause [, ...] ]
[ WHERE where_clause ];

Details

An INSERT statement creates one or more new rows in an existing table. If the column_name list after table_name is omitted then the number of values inserted into each row must be the same as the number of columns in the table. In this case the result of evaluating the left-most expression from each term of the VALUES list is inserted into the left-most column of each new row, and so forth for each subsequent expression. If a column_name list is specified, then the number of values in each term of the VALUE list must match the number of specified columns. Each of the named columns of the new row is populated with the results of evaluating the corresponding VALUES expression. Table columns that do not appear in the column list are populated with the default column value (specified as part of the CREATE TABLE statement), or with NULL if no default value is specified.

The alternative INSERT ... DEFAULT VALUES statement inserts a single new row into the named table. Each column of the new row is populated with its default value, or with a NULL if no default value is specified as part of the column definition in the CREATE TABLE statement.

The last form of the INSERT statement contains a SELECT statement instead of a VALUES clause. A new entry is inserted into the table for each row of data returned by executing the SELECT statement. If a column name list is specified, the number of columns in the result of the SELECT must be the same as the number of items in the column name list. Otherwise, if no column name list is specified, the number of columns in the result of the SELECT must be the same as the number of columns in the table. Only simple (flattened) SELECT statement may be used in an INSERT statement of this form. This means the SELECT statements cannot include UNIONs, JOINs, or further sub-queries. Additionally, only direct references to tables on the same chain are supported.

⚠️ Although the GROUP BY clause is supported, HAVING is not allowed in any SELECT statements within an INSERT. Additionally, under the hood, the Tableland Specification forces an implicit ORDER BY rowid clause on the SELECT statement.

UPSERT

UPSERT is a special syntax addition to INSERT that causes the INSERT to behave as an UPDATE or a no-op if the INSERT would violate a uniqueness constraint. UPSERT is not standard SQL. UPSERT in Tableland follows the syntax used in SQLite.

Structure

INSERT INTO table_name [ ( *column_name* [, ...] ) ] VALUES (
{ expression } [, ...]
) [upsert_clause];

where upsert_clause has structure

ON CONFLICT [ conflict_target ] conflict_action

where conflict_target has structure

[ ( *column_name* [, ...] ) ]  [ WHERE condition ]

and conflict_action can be one of

DO NOTHING

or

DO UPDATE SET { column_name = { expression | DEFAULT } } [, ...]
[ WHERE condition ];

Details

An UPSERT is an ordinary INSERT statement that is followed by the special ON CONFLICT clause shown above.

The syntax that occurs in between the "ON CONFLICT" and "DO" keywords is called the "conflict target". The conflict target specifies a specific uniqueness constraint that will trigger the upsert. The conflict target is required for DO UPDATE upserts, but is optional for DO NOTHING. When the conflict target is omitted, the upsert behavior is triggered by a violation of any uniqueness constraint on the table of the INSERT.

If the insert operation would cause the uniqueness constraint identified by the conflict_target clause to fail, then the insert is omitted and either the DO NOTHING or DO UPDATE operation is performed instead. In the case of a multi-row insert, this decision is made separately for each row of the insert.

The special UPSERT processing happens only for uniqueness constraint on the table that is receiving the INSERT. A "uniqueness constraint" is an explicit UNIQUE or PRIMARY KEY constraint within the CREATE TABLE statement, or a unique index. UPSERT does not intervene for failed NOT NULL constraints.

Column names in the expressions of a DO UPDATE refer to the original unchanged value of the column, before the attempted INSERT.

Note that the DO UPDATE clause acts only on the single row that experienced the constraint error during INSERT. It is not necessary to include a WHERE clause that restricts the action to that one row. The only use for the WHERE clause at the end of the DO UPDATE is to optionally change the DO UPDATE into a no-op depending on the original and/or new values.

UPDATE

An UPDATE statement is used to modify a subset of the values stored in zero or more rows of the database table identified by the table name.

Structure

UPDATE table_name
SET { column_name = { expression | DEFAULT } } [, ...]
[ WHERE condition ];

Details

If the UPDATE statement does not have a WHERE clause, all rows in the table are modified by the UPDATE. Otherwise, the UPDATE affects only those rows for which the WHERE clause boolean expression is true. It is not an error if the WHERE clause does not evaluate to true for any row in the table; this just means that the UPDATE statement affects zero rows.

The modifications made to each row affected by an UPDATE statement are determined by the list of assignments following the SET keyword. Each assignment specifies a column-name to the left of the equals sign and a scalar expression to the right. For each affected row, the named columns are set to the values found by evaluating the corresponding scalar expressions. If a single column-name appears more than once in the list of assignment expressions, all but the rightmost occurrence is ignored. Columns that do not appear in the list of assignments are left unmodified. The scalar expressions may refer to columns of the row being updated. In this case all scalar expressions are evaluated before any assignments are made.

ℹ️ An assignment in the SET clause can be a parenthesized list of column names on the left and a ROW value of the same size on the right. For example, consider the following two “styles” of UPDATE statements: UPDATE table_id SET (a,b)=(b,a); or UPDATE table_id SET a=b, b=a;.

GRANT/REVOKE

The GRANT and REVOKE commands are used to define low-level access privileges for a table identified by table name and id.

Structure

GRANT { INSERT | UPDATE | DELETE } [, ...]
ON { [ TABLE ] table_name [, ...] }
TO role [, ...]

REVOKE { INSERT | UPDATE | DELETE } [, ...]
ON { [ TABLE ] table_name [, ...] }
FROM role [, ...]

Details

The GRANT command gives specific privileges on a table to one or more role. These privileges are added to those already granted, if any. By default, the creator of a table (as specified by a public ETH address) has all (valid) privileges on creation. The owner could, however, choose to revoke some of their own privileges for safety reasons.

Related, if a table is created with an access controller contract specified, or if an address with sufficient privileges updates a table’s access control rules to use a controller contract, then all command-based access control rules are ignored in favor of the controller contract access control. In other words, if a controller contract is set, GRANT/REVOKE is disabled. See Onchain API Specification for further details on specifying and controlling access via a controller smart contract.

⚠️ Currently, the only allowable privileges for granting are INSERT, UPDATE and DELETE. Note that SELECT privileges are not required at this time, as SELECT statements are not access controlled (all reads are allowed). See the SELECT section for further details.

Roles (role) in Tableland are defined by an Ethereum public-key based address. Any (hex string encoded) ETH address is a valid Tableland role, and as such, privileges can be granted to any valid ETH address. In practice, ETH address strings must be specified as string literals using single quotes (e.g., '0x181Ec6E8f49A1eEbcf8969e88189EA2EFC9108dD').

ℹ️ Only a table owner has permission to GRANT or REVOKE access privileges to other roles/accounts.

Conversely to the GRANT command, the REVOKE command removes specific, previously granted access privileges on a table from one or more roles. All role definitions and allowable privileges associated with granting privileges also apply to revoking them.

SELECT

The SELECT statement is used to query the database. The result of a SELECT is zero or more rows of data where each row has a fixed number of columns. A SELECT statement does not make any changes to the database.

Structure

The SELECT statement is the work-house of the SQL query model, and as such, the available syntax is extremely complex. In practice, most SELECT statements are simple SELECT statements of the form:

SELECT [ ALL | DISTINCT ]
[ * | expression [, ...] ]
[ FROM from_clause [, ...] ]
[ WHERE where_clause ]
[ GROUP BY [ expression [, ...] ]
[ ORDER BY expression [ ASC | DESC ]
[ LIMIT { count | ALL } ]
[ OFFSET { number } ]

See the standalone sections on WHERE, FROM, and JOIN clauses for further details on query structure.

Details

Generating the results of a simple SELECT statement is presented as a four step process in the description below:

  1. FROM clause processing: The input data for the simple SELECT is determined. The input data is either implicitly a single row with 0 columns (if there is no FROM clause) or is determined by the FROM clause.
  2. WHERE clause processing: The input data is filtered using the WHERE clause expression.
  3. Result set processing (GROUP BY and result expression processing): The set of result rows is computed by aggregating the data according to any GROUP BY clause and calculating the result set expressions for the rows of the filtered input dataset.
  4. DISTINCT/ALL keyword processing: If the query is a SELECT DISTINCT query (see further details below), duplicate rows are removed from the set of result rows.

There are two types of simple SELECT statement — aggregate and non-aggregate queries. A simple SELECT statement is an aggregate query if it contains either a GROUP BY clause or one or more aggregate functions in the result set. Otherwise, if a simple SELECT contains no aggregate functions or a GROUP BY clause, it is a non-aggregate query.

Once the input data from the FROM clause has been filtered by the WHERE clause expression (if any), the set of result rows for the simple SELECT are calculated. Exactly how this is done depends on whether the simple SELECT is an aggregate or non-aggregate query, and whether or not a GROUP BY clause was specified.

One of the ALL or DISTINCT keywords may follow the SELECT keyword in a simple SELECT statement. If the simple SELECT is a SELECT ALL, then the entire set of result rows are returned by the SELECT. If neither ALL or DISTINCT are present, then the behavior is as if ALL were specified. If the simple SELECT is a SELECT DISTINCT, then duplicate rows are removed from the set of result rows before it is returned. For the purposes of detecting duplicate rows, two NULL values are considered to be equal.

WHERE clause

Structure

WHERE condition

Details

The SQL WHERE clause is an optional clause of the SELECT, DELETE, and/or UPDATE statements. It appears after the primary clauses of the corresponding statement. For example in a SELECT statement, the WHERE clause can be added after the FROM clause to filter rows returned by the query. Only rows for which the WHERE clause expression evaluates to true are included from the dataset before continuing. Rows are excluded from the result if the WHERE clause evaluates to either false or NULL.

When evaluating a SELECT statement with a WHERE clause, Tableland uses the following steps:

  1. Determine the table(s) in the FROM clause,
  2. Evaluate the conditions in the WHERE clause to determine the rows that meet the given condition,
  3. Generate the final result set based on the rows in the previous step, with columns subset to match the SELECT statement.

ℹ️ The search condition in the WHERE clause is made up of any number of comparisons (=, <, >, LIKE, IN, etc), combined using a range of logical operators (e.g., OR, AND, ALL, ANY, etc).

FROM clause

Structure

FROM { table_name [ * ] [ [ AS ] alias ] | ( sub_select ) [ AS ] alias }

Details

The input data used by a simple SELECT query is a set of N rows each M columns wide. If the FROM clause is omitted from a simple SELECT statement, then the input data is implicitly a single row zero columns wide (i.e. N=1 and M=0).

If a FROM clause is specified, the data on which a simple SELECT query operates comes from the one or more tables or sub-queries (SELECT statements in parentheses) specified following the FROM keyword. A sub-query specified in the table or sub-query clause following the FROM clause in a simple SELECT statement is handled as if it was a table containing the data returned by executing the sub-query statement.

If there is only a single table or sub-query in the FROM clause (a common case), then the input data used by the SELECT statement is the contents of the named table. If there is more than one table or sub-query in FROM clause, then the contents of all tables and/or sub-queries are joined into a single dataset for the simple SELECT statement to operate on. Exactly how the data is combined depends on the specific JOIN clause (i.e., the combination of join operator and join constraint) used to connect the tables or sub-queries together.

JOIN clause

Structure

[ NATURAL ] join_type table_or_subquery [ ON on_expression | USING ( column_name [, ...] ) ]

where join_type is one of

  • [ INNER ] JOIN
  • LEFT [ OUTER ] JOIN
  • RIGHT [ OUTER ] JOIN
  • FULL [ OUTER ] JOIN

ℹ️ There is no difference between the "INNER JOIN", "JOIN" and "," join operators. They are completely interchangeable in Tableland.

or,

CROSS JOIN table_or_subquery [ ON on_expression | USING ( column_name [, ...] ) ]

or, a series of comma-separated tables or sub-queries followed by an optional join constraint as in above.

The table_or_subquery is a table or sub-query of the form:

{ table_name [ [ AS ] alias ] | ( sub_select ) [ AS ] alias }

Details

All joins in Tableland are based on the cartesian product of the left- and right-want databsets. The columns of the cartesian product dataset are, in order, all the columns of the left-hand dataset followed by all the columns of the right-hand dataset. This is a row in the cartesian product dataset formed by combining each unique combination of a raw from the left-hand and right-hand datasets. In other words, if the left-hand dataset consists of NlN_{l} rows and MlM_{l} columns, and the right-hand dataset of NrN_{r} rows of MrM_{r} columns, then the cartesian product is a dataset of Nl×NrN_{l} \times N_{r} rows, each containing Nl+NrN_{l} + N_{r} columns.

If the join operator is "CROSS JOIN", "INNER JOIN", "JOIN" or a comma (",") and there is no ON or USING clause, then the result of the join is simply the cartesian product of the left and right-hand datasets. If join operator does have ON or USING clauses, those are handled according to the following bullet points:

  • If there is an ON clause then the ON expression is evaluated for each row of the cartesian product as a boolean expression. Only rows for which the expression evaluates to true are included from the dataset.
  • If there is a USING clause then each of the column names specified must exist in the datasets to both the left and right of the join operator. For each pair of named columns, the expression Xlhs=XrhsX_{lhs} = X_{rhs} is evaluated for each row of the cartesian product as a boolean expression. Only rows for which all such expressions evaluates to true are included from the result set. When comparing values as a result of a USING clause, the normal rules for handling affinities, collation sequences and NULL values in comparisons apply. The column from the dataset on the left-hand side of the join operator is considered to be on the left-hand side of the comparison operator (=) for the purposes of collation sequence and affinity precedence.
  • For each pair of columns identified by a USING clause, the column from the right-hand dataset is omitted from the joined dataset. This is the only difference between a USING clause and its equivalent ON constraint.
  • If the NATURAL keyword is in the join operator then an implicit USING clause is added to the join constraints. The implicit USING clause contains each of the column names that appear in both the left and right-hand input datasets. If the left and right-hand input datasets feature no common column names, then the NATURAL keyword has no effect on the results of the join. A USING or ON clause may not be added to a join that specifies the NATURAL keyword.
  • If the join operator is a "LEFT JOIN" or "LEFT OUTER JOIN", then after the ON or USING filtering clauses have been applied, an extra row is added to the output for each row in the original left-hand input dataset that does not match any row in the right-hand dataset. The added rows contain NULL values in the columns that would normally contain values copied from the right-hand input dataset
  • If the join operator is a "RIGHT JOIN" or "RIGHT OUTER JOIN", then after the ON or USING filtering clauses have been applied, an extra row is added to the output for each row in the original right-hand input dataset that does not match any row in the left-hand dataset. The added rows contain NULL values in the columns that would normally contain values copied from the left-hand input dataset.
  • A "FULL JOIN" or "FULL OUTER JOIN" is a combination of a "LEFT JOIN" and a "RIGHT JOIN". Extra rows of output are added for each row in left dataset that matches no rows in the right, and for each row in the right dataset that matches no rows in the left. Unmatched columns are filled in with NULL.

When more than two tables are joined together as part of a FROM clause, the join operations are processed in order from left to right. In other words, the FROM clause (A+B+C)(A + B + C) is computed as ((A+B)+C)((A + B) + C).

⚠️ The "CROSS JOIN" join operator produces the same result as the "INNER JOIN", "JOIN" and "," operators, but is handled differently by the query optimizer in that it prevents the query optimizer from reordering the tables in the join. An application programmer can use the CROSS JOIN operator to directly influence the algorithm that is chosen to implement the SELECT statement. Avoid using CROSS JOIN except in specific situations where manual control of the query optimizer is desired. Avoid using CROSS JOIN early in the development of an application as doing so is a premature optimization. The special handling of CROSS JOIN is an implementation detail. It is not a part of standard SQL, and should not be relied upon.

Compound Select Statements

Two or more simple SELECT statements may be connected together to form a compound SELECT using the UNION, UNION ALL, INTERSECT or EXCEPT operator.

In a compound SELECT, all the constituent SELECTs must return the same number of result columns. As the components of a compound SELECT must be simple SELECT statements, they may not contain ORDER BY or LIMIT clauses. ORDER BY and LIMIT clauses may only occur at the end of the entire compound SELECT, and then only if the final element of the compound is not a VALUES clause.

A compound SELECT created using UNION ALL operator returns all the rows from the SELECT to the left of the UNION ALL operator, and all the rows from the SELECT to the right of it. The UNION operator works the same way as UNION ALL, except that duplicate rows are removed from the final result set. The INTERSECT operator returns the intersection of the results of the left and right SELECTs. The EXCEPT operator returns the subset of rows returned by the left SELECT that are not also returned by the right-hand SELECT. Duplicate rows are removed from the results of INTERSECT and EXCEPT operators before the result set is returned.

For the purposes of determining duplicate rows for the results of compound SELECT operators, NULL values are considered equal to other NULL values and distinct from all non-NULL values. The collation sequence used to compare two text values is determined as if the columns of the left and right-hand SELECT statements were the left and right-hand operands of the equals (=) operator, except that greater precedence is not assigned to a collation sequence specified with the postfix COLLATE operator. No affinity transformations are applied to any values when comparing rows as part of a compound SELECT.

When three or more simple SELECTs are connected into a compound SELECT, they group from left to right. In other words, if AA, BB and CC are all simple SELECT statements, (ABC)(A * B * C) is processed as ((AB)C)((A * B) * C).

Custom functions

The Tableland SQL Specification includes several web3 native functions that simplify working with blockchain transactions. The list of custom functions may grow over time.

TXN_HASH()

The Validator will replace this text with the hash of the transaction that delivered the SQL event (only available in write queries).

INSERT INTO {table_name} VALUES (TXN_HASH());

BLOCK_NUM()

The Validator will replace this text with the number of the block that delivered the SQL event (only available in write queries).

INSERT INTO {table_name} VALUES (BLOCK_NUM());

If BLOCK_NUM is called with an integer argument (i.e., BLOCK_NUM(<chain_id>)), the Validator will replace this text with the number of the last seen block for the given chain (only available to read queries).

Data Types

Tableland supports a small set of accepted column types in user-defined tables. The currently supported types are listed below and can be used to represent most, if not all, common SQL types:

TypeDescription
INTSigned integer values, stored in 0, 1, 2, 3, 4, 6, or 8 bytes depending on the magnitude of the value.
INTEGERSame as INT, except it may also be used to represent an auto-incrementing PRIMARY KEY field.
TEXTText string, stored using the database encoding (UTF-8).
BLOBA blob of data, stored exactly as it was input. Useful for byte slices etc.

Details

When creating tables, every column definition must specify a data type for that column, and the data type must be one of the above types. No other data type names are allowed, though new types might be added in future versions of the Tableland SQL specification.

Content inserted into a column must be either a NULL (assuming there is no NOT NULL constraint on the column) or the type specified. Tableland will attempt to coerce input data into the appropriate type using the usual affinity rules, as most SQL engines all do. However, if the value cannot be losslessly converted in the specified datatype, then an error will be raised.

Common Types

For users looking for more nuanced data types in tables, the following set of recommendations will help guide table schema design. Additionally, new types might be added in future versions of the Tableland SQL Specification, and users are able to make requests/suggestions via Tableland TIPs.

Character

Tableland represents all character/text types using the single variable-length TEXT type. Although the type TEXT is not in any SQL standard, several other SQL database management systems have it as well. You can store any text/character-based data as TEXT. Additionally, more complex data types such as dates, timestamps, JSON strings, and more can be represented using TEXT (or in some cases BLOB).

Integers

Numeric types often consist of integer and floating-point (float/real) numbers. On Tableland, two-, four-, and eight-byte integers are all represented by the INTEGER type, and their storage size depends on the magnitude of the value itself.

Floats

Tableland does not have a separate data type to represent float/real types. This is because in practice floating point values are approximate, which may lead to non-deterministic behavior across compute platforms. If you need an exact answer, you should not use floating-point values, in Tableland or in any other software. This is not a Tableland limitation per se, but a mathematical limitation inherent in the design of floating-point numbers.

See the SQLite documentation about issues with floating-point numbers, or learn more about why floating-point math is hard.

⚠️ In addition to not supporting floating point values (REAL) as a storage data type in create statements, the Tableland specification also does not allow REAL value literals in read or write queries.

Boolean

Tableland does not have a separate data type to represent boolean values. Instead, Tableland users should represent true and false values using the integers 1 (true) and 0 (false).

Date/Time

Tableland does not have a storage class set aside for storing dates and/or times. Instead, users of Tableland can store dates and times as TEXT or INTEGER values:

  • TEXT as ISO-8601 strings.
  • INTEGER as Unix Time (number of seconds since (or before) 1970-01-01 00:00:00 UTC).

Tableland does not support any of the date nor time functions provided by the SQLite database engine. Namely, these functions can lead to non-deterministic behavior, so they are not available.

JSON

JSON data types are for storing JSON (JavaScript Object Notation) data, as specified in RFC 7159. In practice, Tableland stores JSON as ordinary TEXT. Users are able to manipulate JSON data using a number of functions that make working with JSON data much easier. For example, the json(X) function verifies that its argument X is a valid JSON string and returns a minified version of that JSON string (with all unnecessary whitespace removed). If X is not a well-formed JSON string, then this routine throws an error. The JSON manipulation functions supported by Tableland is derived from SQLite, which in turn is generally compatible (in terms of syntax) with PostgresQL. Tableland currently supports the 15 scalar functions and operators and two aggregate SQL functions for JSON data provided by the SQLite database engine.

🚧 Feature At Risk: Note that these JSON scalar functions, operators, and aggregate functions have not yet been formalized into the Tableland SQL language specification. You are welcome to use them for now, but they should be considered unstable features.

Solidity

To prevent overflows while working with Solidity numbers, it is recommended to use a text type in certain scenarios. Anything larger than a uint64 / int32 could lead to an overflow in the Tableland database. Note that in many use cases, it is unlikely overflows will happen due to the extremely large size of these numbers.

Alternatively, consider casting the overflow-able numbers to or simply use a int64 in smart contracts if it makes sense for the use case. See the following tables for how each Solidity number should be defined in Tableland schemas:

Solidity TypeSQL Type
uint256text
uint128text
uint64text
uint32integer
uint16integer
uint8integer
int256text
int128text
int64integer
int32integer
int16integer
int8integer

Other best practices have also been defined below:

Solidity TypeSQL Type
stringtext
addresstext
bytesblob
boolinteger

⚠️ Tableland doesn’t support boolean values, but TRUE/FALSE keywords are supported since they're aliased to a 1 or 0. Thus, a Solidity bool should be defined as an integer in Tableland.

Encoding

As mentioned in the section on Statement Types, the core Tableland SQL parser accepts a semicolon-separated list of statements, which are then parsed and evaluated according to this Tableland SQL Specification. Internally, the statements are represented using an abstract syntax tree (AST). The internal representation of the nodes of this AST is outside the bounds of the Tableland SQL Specification, however, further details can be found in the Go Tableland SQL Parser reference implementation.

With the above caveat in mind, the Tableland SQL Specification does define a canonical string encoding of a set of (compliant) SQL statements that have passed through the Tableland SQL Parser (and have been represented via the Parser's AST). That is, this Specification outlines — in general terms — the string encoding produced by parsing a set of Tableland SQL Specification compliant statements and re-encoding them into a canonical (string) format.

There are some nuances and corner cases that affect the final encoded string. For example, the statement UPDATE t SET (A, b) = (1, 2); is ultimately encoded as update t set A = 1, b = 2. This is because the two statements are equivalent, and their representation within the AST is identical. As such, when producing the canonical string for such a statement, the parser outputs the most conventional form.

In general, any (set of) statement(s) processed by the parser should be encoded such that,

  • All SQL language components are specified using lower case ASCII characters,
  • The execution of the (set of) statement(s) after encoding is equivalent to the original set of statements, and
  • The encoding of a (set of) statements(s) is as close as possible to the original (set of) statement(s).

Any further guarantees are left outside the scope of this specification.