Tutorial¶

Welcome to ONNX Script Tutorials¶

For extended tutorials on how to use the Optimizer and Rewriter tools, refer to the relevant sub-sections within the Tutorial section.

In this tutorial, we illustrate the features supported by ONNX Script using examples.

Basic Features¶

The example below shows a definition of Softplus as an ONNX Script function.

# Copyright (c) Microsoft Corporation.
# Licensed under the MIT License.
# We use ONNX opset 15 to define the function below.
from onnxscript import opset15 as op
from onnxscript import script


# We use the script decorator to indicate that the following function is meant
# to be translated to ONNX.
@script()
def Softplus(X):
    return op.Log(op.Exp(X) + 1.0)

In the above example, expressions such as op.Log(...) and op.Exp(...) represent a call to an ONNX operator (and is translated into an ONNX NodeProto). Here, op serves to identify the opset containing the called operator. In this example, we are using the standard ONNX opset version 15 (as identified by the import statement from onnxscript.onnx_opset import opset15 as op).

Operators such as + are supported as syntactic shorthand and are mapped to a corresponding standard ONNX operator (such as Add) in an appropriate opset. In the above example, the use of op indicates opset 15 is to be used. If the example does not make use of an opset explicitly in this fashion, it must be specified via the parameter default_opset to the @script() invocation.

Similarly, constant literals such as 1.0 are allowed as syntactic shorthand (in contexts such as in the above example) and are implicitly promoted into an ONNX tensor constant.

Omitting optional inputs¶

Some of the input arguments of ONNX ops are optional: for example, the min and max inputs of the Clip operator. The value None can be used to indicate an omitted optional input, as shown below, or it can be simply omitted in the case of trailing inputs:

# Copyright (c) Microsoft Corporation.
# Licensed under the MIT License.
from onnxscript import opset15 as op
from onnxscript import script


@script()
def omitted_input(x):
    # The following two statements are equivalent:
    y1 = op.Clip(x)
    y2 = op.Clip(x, None, None)
    # The following example shows an omitted optional input, followed by another input
    y3 = op.Clip(x, None, 1.0)
    return y1 + y2 + y3

Specifying attribute-parameter values¶

The example below illustrates how to specify attribute-values in a call. In this example, we call the ONNX operator Shape and specify the attribute values for the attributes start and end.

# Copyright (c) Microsoft Corporation.
# Licensed under the MIT License.
from onnxscript import opset15 as op
from onnxscript import script


@script()
def FirstDim(X):
    return op.Shape(X, start=0, end=1)

In the translation of a call to an ONNX operator, the translator makes use of the OpSchema specification of the operator to map the actual parameters to appropriate input parameters and attribute parameters. Since the ONNX specification does not indicate any ordering for attribute parameters, it is recommended that attribute parameters be specified using keyword arguments (aka named arguments).

If the translator does not have an opschema for the called op, it uses the following strategy to map the actual parameters to appropriate input parameters and attribute parameters: Keyword arguments of Python are translated into attribute parameters (of ONNX), while positional arguments are translated into normal value-parameters. Thus, in the above example, X is treated as a normal value-parameter for this particular call, while start and end are treated as attribute-parameters (when an opschema is unavailable).

Specifying tensor constants¶

Tensor constants can be created using the ONNX utility make_tensor and these can be used as attribute values, as shown below. Further, they can be promoted to be used as tensor values using the ONNX Constant op, also as shown below.

# Copyright (c) Microsoft Corporation.
# Licensed under the MIT License.
from onnx import TensorProto, helper

from onnxscript import opset15 as op
from onnxscript import script


@script()
def tensor_attr(x):
    c = op.Constant(value=helper.make_tensor("scalar_half", TensorProto.FLOAT, (), [0.5]))
    return op.Mul(c, x)

The code shown above, while verbose, allows the users to explicitly specify what they want. The converter, as a convenience, allows users to use numeric constants, as in the example below, which is translated into the same ONNX representation as the one above.

# Copyright (c) Microsoft Corporation.
# Licensed under the MIT License.
from onnxscript import opset15 as op
from onnxscript import script


@script()
def tensor_attr(x):
    return op.Mul(0.5, x)

The direct usage of literals can be used to create scalars or one-dimensional tensors of type FLOAT or INT64 or STRING, as shown in the table below.

Python source

Generated ONNX constant

0

Scalar value 0 of type INT64

0.0

Scalar value 0.0 of type FLOAT

"x"

Scalar value "x" of type STRING

[0, 1]

One dimensional tensor of type INT64

[0.0, 1.0]

One dimensional tensor of type FLOAT

["x", "y"]

One dimensional tensor of type STRING

However, if the user wants to use tensor constants of other types or other rank, they need to do so more explicitly (as in the previous example).

Semantics: Script Constants¶

Attributes in ONNX are required to be constant values. In ONNX Script, the expression specified as an attribute is evaluated at script-time (when the script decorator is evaluated) in the context in which the script function is defined. The resulting python value is translated into an ONNX attribute, as long as it has a valid type.

This has several significant semantic implications. First, it allows the use of arbitrary python code in a context where an attribute-value is expected. However, the python code must be evaluatable using the global context in which the script-function is defined. For example, computation using the parameters of the function itself (even if they are attribute-parameters) is not permitted.

ONNX Script assumes that such python-code represents constants. If the values of the variables used in the expression are subsequently modified, this modification has no effect on the attribute-value or the ONNX function/model created. This may potentially cause the behavior of eager-mode execution to be inconsistent with the ONNX construct generated.

Thus, the example shown above is equivalent to the following:

# Copyright (c) Microsoft Corporation.
# Licensed under the MIT License.
from onnx import TensorProto, helper

from onnxscript import opset15 as op
from onnxscript import script

script_const = helper.make_tensor("scalar_half", TensorProto.FLOAT, (), [0.5])


@script()
def tensor_attr(x):
    c = op.Constant(value=script_const)
    return c * x


# The following assignment has no effect on the ONNX FunctionProto
# generated from tensor_attr:


script_const = helper.make_tensor("scalar_one", TensorProto.FLOAT, (), [1.0])

fp = tensor_attr.to_function_proto()

Specifying formal attribute parameters of functions¶

The (formal) input parameters of Python functions are treated by the converter as representing either attribute-parameters or input value parameters (of the generated ONNX function). However, the converter needs to know for each parameter whether it represents an attribute or input. The converter uses the type annotation on the formal input parameters to make this distinction. Thus, in the example below, alpha is treated as an attribute parameter (because of its float type annotation).

# Copyright (c) Microsoft Corporation.
# Licensed under the MIT License.
from onnxscript import opset15 as op
from onnxscript import script


@script()
def LeakyRelu(X, alpha: float):
    return op.Where(X < 0.0, alpha * X, X)

The (ONNX) types of attributes supported and their corresponding (Python) type annotations are shown in the table below. Other types of ONNX attributes are not yet supported.

ONNX Type

Python Type Annotation

AttributeProto.FLOAT

float

AttributeProto.INT

int, bool

AttributeProto.STRING

str

AttributeProto.FLOATS

Sequence[float]

AttributeProto.INTS

Sequence[int]

AttributeProto.STRINGS

Sequence[str]

Automatic promotion of attribute-parameters to values¶

As illustrated in the above example, when an attribute-parameter is used in a context requiring a value-parameter, the converter will automatically convert the attribute into a tensor-value. Specifically, in the sub-expression alpha * X, the attribute parameter alpha is used as a value-parameter of the call to the Mul op (denoted by the *) and is automatically converted. Thus,

# Copyright (c) Microsoft Corporation.
# Licensed under the MIT License.
from onnxscript import opset15 as op
from onnxscript import script


@script()
def LeakyRelu(X, alpha: float):
    return op.Where(X < 0.0, alpha * X, X)

is expanded to the following:

# Copyright (c) Microsoft Corporation.
# Licensed under the MIT License.
from onnxscript import opset15 as op
from onnxscript import script


@script()
def LeakyRelu(X, alpha: float):
    alpha_value = op.Constant(value_float=alpha)
    return op.Where(X < 0.0, alpha_value * X, X)

Automatic casts for constant values¶

The converter also automatically introduces casts (via the ONNX CastLike op) when constants are used in a context where they are constrained to be of the same type as some other (non-constant) operand. For example, the expression 2 * X is expanded to op.CastLike(2, X) * X, which allows the same code to work for different types of X.

Indexing and Slicing¶

ONNX Script supports the use of Python’s indexing and slicing operations on tensors, which are translated into ONNX’s Slice and Gather operations. The semantics of this operation is similar to that of Numpy’s.

In the expression e[i_1, i_2, ..., i_n], n is either the rank of the input tensor or any value less than that. Each index-value i_j may be a scalar value (a tensor of rank zero) or a higher-dimensional tensor or a slice-expression of the form start:end:step. Semantically, a slice-expression start:end:step is equivalent to a 1-dimensional tensor containing the corresponding sequence of values.

However, the translator maps indexing using slice-expressions to ONNX’s Slice operation which may be more efficient than the corresponding Gather operation. The more general case (where i_j is an arbitrary tensor) is translated using the Gather operation.

Note: The current implementation does not yet support the use of arbitrary tensors in the index-expressions. It does not support the use of ellipsis or newaxis in the index.

Control-Flow¶

The support for control-flow constructs in ONNX Script is limited by requirements of ONNX control-flow ops.

Conditional statements¶

The function definition below illustrates the use of conditionals.

# Copyright (c) Microsoft Corporation.
# Licensed under the MIT License.
from onnxscript import opset15 as op
from onnxscript import script


@script()
def Dropout(data, ratio, training_mode, seed: float):
    if training_mode:
        rand = op.RandomUniformLike(data, dtype=1, seed=seed)
        mask = rand >= ratio
        output = op.Where(mask, data, 0) / (1.0 - ratio)
    else:
        mask = op.ConstantOfShape(op.Shape(data), value=True)
        output = data
    return (output, mask)

The use of conditional statements requires that any variable that is used in the code has a definition of the same variable along all possible paths to the use.

Loops¶

ONNX implements a loop operator doing a fixed number of iterations and/or a loop breaking if a condition is not true anymore. First example below illustrates the use of the most simple case: a fixed number of iterations.

# Copyright (c) Microsoft Corporation.
# Licensed under the MIT License.
from onnxscript import opset15 as op
from onnxscript import script


@script()
def sumprod(x, N):
    sum = op.Identity(x)
    prod = op.Identity(x)
    for _ in range(N):
        sum = sum + x
        prod = prod * x
    return sum, prod

Second example shows a loop breaking if a condition is not true any more.

# Copyright (c) Microsoft Corporation.
# Licensed under the MIT License.
from onnx import TensorProto
from onnx.helper import make_tensor

from onnxscript import opset15 as op
from onnxscript import script


@script()
def sumprod(x, N):
    sum = op.Identity(x)
    prod = op.Identity(x)
    cond = op.Constant(value=make_tensor("true", TensorProto.BOOL, [1], [1]))
    i = op.Constant(value=make_tensor("i", TensorProto.INT64, [1], [0]))
    while cond:
        sum = sum + x
        prod = prod * x
        i = i + 1
        cond = i < 10
    return sum, prod

Third example mixes both types of loops.

# Copyright (c) Microsoft Corporation.
# Licensed under the MIT License.
from onnxscript import opset15 as op
from onnxscript import script


@script()
def sumprod_break(x, N):
    sum = op.Identity(x)
    prod = op.Identity(x)
    for _ in range(N):
        sum = sum + x
        prod = prod * x
        cond = op.ReduceSum(prod) > 1e7
        # ONNX does not support break instruction.
        # onnxscript can only convert example if the break
        # instruction is placed at the end of the loop body.
        if cond:
            break
    return sum, prod

Encoding Higher-Order Ops: Scan¶

ONNX allows graph-valued attributes. This is the mechanism used to define (quasi) higher-order ops, such as If, Loop, Scan, and SequenceMap. While we use Python control-flow to encode If and Loop, ONNX Script supports the use of nested Python functions to represent graph-valued attributes, as shown in the example below:

# Copyright (c) Microsoft Corporation.
# Licensed under the MIT License.
from onnxscript import graph, script
from onnxscript import opset15 as op


@script()
def CumulativeSum(X):
    @graph()
    def Sum(sum_in, next):
        sum_out = sum_in + next
        return sum_out, sum_out

    _all_sum, cumulative_sum = op.Scan(0, X, body=Sum, num_scan_inputs=1)
    return cumulative_sum

In this case, the function-definition of Sum is converted into a graph and used as the attribute-value when invoking the Scan op.

Function definitions used as graph-attributes must satisfy some constraints. They cannot update outer-scope variables, but may reference them. (Specifically, the functions cannot use global or nonlocal declarations.) They are also restricted from using local-variables with the same name as outer-scope variables (no shadowing).

There is also an interaction between SSA-renaming and the use of outer-scope variables inside a function-definition. The following code is invalid, since the function CumulativeSum references the global g, which is updated in between the function-definition and function-use. Note that, from an ONNX perspective, the two assignments to g represent two distinct tensors g1 and g2.

# Copyright (c) Microsoft Corporation.
# Licensed under the MIT License.
from onnxscript import graph, script
from onnxscript import opset15 as op

try:

    @script()
    def CumulativeSum(X):
        g = op.Constant(value=0)

        @graph()
        def Sum(sum_in, next):
            sum_out = sum_in + next + g
            return sum_out, sum_out

        g = op.Constant(value=1)
        _all_sum, cumulative_sum = op.Scan(0, X, body=Sum, num_scan_inputs=1)
        return cumulative_sum

except Exception as e:
    assert "Outer scope variable" in str(e)