From 1fcf19936bd83406e343f0dfcc2f39bd45dd7e0d Mon Sep 17 00:00:00 2001 From: Lutz Roeder Date: Sat, 19 Oct 2024 12:40:51 -0700 Subject: [PATCH] Update pytorch.js (#1061) --- source/python.js | 299 ++- source/pytorch-metadata.json | 3975 +++++++++++++++++++--------------- source/pytorch.js | 399 ++-- source/pytorch.py | 1 + test/models.json | 2 - tools/pytorch | 21 + tools/pytorch_script.py | 277 ++- 7 files changed, 2806 insertions(+), 2168 deletions(-) diff --git a/source/python.js b/source/python.js index 78ba55c213..befbcbb342 100644 --- a/source/python.js +++ b/source/python.js @@ -853,14 +853,17 @@ python.Parser = class { } _type() { - const type = this._node(); - type.type = 'type'; - type.name = this._expression(-1, ['[', '=']); - if (type.name) { + const target = this._expression(-1, ['[', '=']); + if (target) { if (this._tokenizer.peek().value === '[') { - type.arguments = this._typeArguments(); + const type = this._node(); + type.type = '[]'; + type.target = target; + type.arguments = this._expressions(); + // type.arguments = this._typeArguments(); + return type; } - return type; + return target; } return null; } @@ -6154,39 +6157,27 @@ python.Execution = class { this._name = name; } } + static get(kind, name) { + return new torch.Type(kind, name); + } kind() { return this._kind; } - __str__() { - if (this._kind === 'VarType' && this._name) { - return this._name; - } else if (this._kind === 'ScalarTypeType') { - return 'ScalarType'; - } else if (this._kind === 'Layout' || this._kind === 'MemoryFormat') { - return this._kind; - } - throw new python.Error(`Not implemented '${this.kind()}'.`); - } str() { if (this._kind === 'VarType' && this._name) { return this._name; } else if (this._kind === 'ScalarTypeType') { return 'ScalarType'; - } else if (this._kind === 'Layout' || this._kind === 'MemoryFormat') { - return this._kind; - } else if (this._kind === 'AnyType') { - return 'Any'; - } else if (this._kind === 'AnyEnumType') { - return 'AnyEnumType'; - } else if (this._kind === 'AnyClassType') { - return 'AnyClassType'; } else if (this._kind === 'QSchemeType') { return 'QScheme'; - } else if (this._kind === 'Storage') { - return 'Storage'; + } else if (this._kind) { + return this._kind; } throw new python.Error(`Not implemented '${this.kind()}'.`); } + __str__() { + return this.str(); + } toString() { return this.__str__(); } @@ -6240,195 +6231,255 @@ python.Execution = class { super('OptionalType'); this._elem = elem; } + static get(elem) { + return new torch.OptionalType(elem); + } getElementType() { return this._elem; } - __str__() { - return `Optional[${this.getElementType().toString()}]`; - } str() { return `${this.getElementType().str()}?`; } + __str__() { + return `Optional[${this.getElementType().__str__()}]`; + } }); this.registerType('torch.ListType', class extends torch.Type { constructor(elem) { super('ListType'); this._elem = elem; } + static get(elem) { + return new torch.ListType(elem); + } getElementType() { return this._elem; } - __str__() { - return `List[${this.getElementType().toString()}]`; - } str() { return `${this.getElementType().str()}[]`; } + __str__() { + return `List[${this.getElementType().__str__()}]`; + } }); this.registerType('torch.FutureType', class extends torch.Type { - constructor(elem, size) { + constructor(elem) { super('FutureType'); this._elem = elem; - this._size = size; + } + static get(elem) { + return new torch.FutureType(elem); } getElementType() { return this._elem; } - __str__() { - return `Future[${this.getElementType().toString()}]`; - } str() { return `Future(${this.getElementType().str()})`; } + __str__() { + return `Future[${this.getElementType().__str__()}]`; + } }); this.registerType('torch.RRefType', class extends torch.Type { constructor(elem) { super('RRefType'); this._elem = elem; } + get(elem) { + return new torch.RRefType(elem); + } getElementType() { return this._elem; } - __str__() { - return `RRef[${this.getElementType().toString()}]`; - } str() { return `RRef(${this.getElementType().str()})`; } + __str__() { + return `RRef[${this.getElementType().__str__()}]`; + } }); this.registerType('torch.AwaitType', class extends torch.Type { constructor(elem) { super('AwaitType'); this._elem = elem; } + static get(elem) { + return new torch.AwaitType(elem); + } getElementType() { return this._elem; } - __str__() { - return `Await[${this.getElementType().toString()}]`; - } str() { return `Await(${this.getElementType().str()})`; } + __str__() { + return `Await[${this.getElementType().__str__()}]`; + } }); this.registerType('torch.TupleType', class extends torch.Type { constructor(elements) { super('TupleType'); this._elements = elements; } + static get(elements) { + return new torch.TupleType(elements); + } elements() { return this._elements; } - __str__() { - return `Tuple[${this.elements().map((elem) => elem.toString()).join(', ')}]`; - } str() { return `(${this.elements().map((elem) => elem.str()).join(', ')})`; } - }); - this.registerType('torch.TensorType', class extends torch.Type { - constructor() { - super('TensorType'); - } __str__() { - return 'Tensor'; - } - str() { - return 'Tensor'; + return `Tuple[${this.elements().map((elem) => elem.__str__()).join(', ')}]`; } }); this.registerType('torch.AnyType', class extends torch.Type { constructor() { super('AnyType'); } + static get() { + torch.AnyType.value = torch.AnyType.value || new torch.AnyType(); + return torch.AnyType.value; + } + str() { + return 'AnyType'; + } }); this.registerType('torch.NoneType', class extends torch.Type { constructor() { super('NoneType'); } + static get() { + torch.NoneType.value = torch.NoneType.value || new torch.NoneType(); + return torch.NoneType.value; + } + str() { + return 'NoneType'; + } __str__() { - return this.kind(); + return 'NoneType'; + } + }); + this.registerType('torch.TensorType', class extends torch.Type { + constructor() { + super('TensorType'); + } + static get() { + torch.TensorType.value = torch.TensorType.value || new torch.TensorType(); + return torch.TensorType.value; } str() { - return this.kind(); + return 'Tensor'; + } + __str__() { + return 'Tensor'; } }); this.registerType('torch.NumberType', class extends torch.Type { constructor() { super('NumberType'); } - __str__() { - return 'number'; + static get() { + torch.NumberType.value = torch.NumberType.value || new torch.NumberType(); + return torch.NumberType.value; } str() { return 'Scalar'; } + __str__() { + return 'number'; + } }); this.registerType('torch.BoolType', class extends torch.Type { constructor() { super('BoolType'); } - __str__() { - return 'bool'; + static get() { + torch.BoolType.value = torch.BoolType.value || new torch.BoolType(); + return torch.BoolType.value; } str() { return 'bool'; } + __str__() { + return 'bool'; + } }); this.registerType('torch.IntType', class extends torch.Type { constructor() { super('IntType'); } - __str__() { - return 'int'; + static get() { + torch.IntType.value = torch.IntType.value || new torch.IntType(); + return torch.IntType.value; } str() { return 'int'; } + __str__() { + return 'int'; + } }); this.registerType('torch.SymIntType', class extends torch.Type { constructor() { super('SymIntType'); } - __str__() { - return 'int'; + static get() { + torch.SymIntType.value = torch.SymIntType.value || new torch.SymIntType(); + return torch.SymIntType.value; } str() { return 'SymInt'; } + __str__() { + return 'int'; + } }); this.registerType('torch.FloatType', class extends torch.Type { constructor() { super('FloatType'); } - __str__() { - return 'float'; + static get() { + torch.FloatType.value = torch.FloatType.value || new torch.FloatType(); + return torch.FloatType.value; } str() { return 'float'; } + __str__() { + return 'float'; + } }); this.registerType('torch.StringType', class extends torch.Type { constructor() { super('StringType'); } - __str__() { - return 'str'; + static get() { + torch.StringType.value = torch.StringType.value || new torch.StringType(); + return torch.StringType.value; } str() { return 'str'; } + __str__() { + return 'str'; + } }); this.registerType('torch.ComplexType', class extends torch.Type { constructor() { super('ComplexType'); } - __str__() { - return 'complex'; + static get() { + torch.ComplexType.value = torch.ComplexType.value || new torch.ComplexType(); + return torch.ComplexType.value; } str() { return 'complex'; } + __str__() { + return 'complex'; + } }); this.registerType('torch.DictType', class extends torch.Type { constructor(key, value) { @@ -6436,38 +6487,45 @@ python.Execution = class { this._key = key; this._value = value; } + static get(key, value) { + return new torch.DictType(key, value); + } getKeyType() { return this._key; } getValueType() { return this._value; } - __str__() { - return `Dict(${this.getKeyType().__str__()}, ${this.getValueType().__str__()})`; - } str() { return `Dict(${this.getKeyType().str()}, ${this.getValueType().str()})`; } + __str__() { + return `Dict(${this.getKeyType().__str__()}, ${this.getValueType().__str__()})`; + } }); this.registerType('torch.DeviceObjType', class extends torch.Type { constructor() { super('DeviceObjType'); } - __str__() { - return 'Device'; + static get() { + torch.DeviceObjType.value ||= new torch.DeviceObjType(); + return torch.DeviceObjType.value; } str() { return 'Device'; } + __str__() { + return 'Device'; + } }); this.registerType('torch.StreamObjType', class extends torch.Type { constructor() { super('StreamObjType'); } - __str__() { + str() { return 'Stream'; } - str() { + __str__() { return 'Stream'; } }); @@ -6475,12 +6533,16 @@ python.Execution = class { constructor() { super('GeneratorType'); } - __str__() { - return 'Generator'; + static get() { + torch._C._GeneratorType.value = torch._C._GeneratorType.value || new torch._C._GeneratorType(); + return torch._C._GeneratorType.value; } str() { return 'Generator'; } + __str__() { + return 'Generator'; + } }); this.registerType('torch._C.AliasInfo', class { constructor() { @@ -6512,7 +6574,7 @@ python.Execution = class { return list.join(''); } }); - this.registerType('torch._C.FunctionSchemaLexer', class { + this.registerType('torch._C.SchemaLexer', class { constructor(buffer) { this.buffer = buffer; this.position = 0; @@ -6610,27 +6672,27 @@ python.Execution = class { const value = L.value; L.next(); switch (value) { - case 'Tensor': return new torch.TensorType(); - case 'bool': return new torch.BoolType(); - case 'int': return new torch.IntType(); - case 'float': return new torch.FloatType(); - case 'complex': return new torch.ComplexType(); - case 'str': return new torch.StringType(); - case 'SymInt': return new torch.SymIntType(); - case 'Scalar': return new torch.NumberType(); - case 'ScalarType': return new torch.Type('ScalarTypeType'); - case 'Device': return new torch.DeviceObjType(); - case 'Layout': return new torch.Type('Layout'); - case 'MemoryFormat': return new torch.Type('MemoryFormat'); - case 'Generator': return new torch._C._GeneratorType(); - case 't': case 't1': case 't2': case 'tVal': return new torch.Type('VarType', value); - case 'Any': return new torch.AnyType(); - case 'AnyEnumType': return new torch.Type('AnyEnumType'); - case 'QScheme': return new torch.Type('QSchemeType'); - case 'Stream': return new torch.StreamObjType(); - case 'Storage': return new torch.Type('Storage'); - case 'AnyClassType': return new torch.Type('AnyClassType'); - case 'NoneType': return new torch.NoneType(); + case 'Tensor': return torch.TensorType.get(); + case 'bool': return torch.BoolType.get(); + case 'int': return torch.IntType.get(); + case 'float': return torch.FloatType.get(); + case 'complex': return torch.ComplexType.get(); + case 'str': return torch.StringType.get(); + case 'SymInt': return torch.SymIntType.get(); + case 'Scalar': return torch.NumberType.get(); + case 'ScalarType': return torch.Type.get('ScalarTypeType'); + case 'Device': return torch.DeviceObjType.get(); + case 'Layout': return torch.Type.get('Layout'); + case 'MemoryFormat': return torch.Type.get('MemoryFormat'); + case 'Generator': return torch._C._GeneratorType.get(); + case 't': case 't1': case 't2': case 'tVal': return torch.Type.get('VarType', value); + case 'Any': return torch.AnyType.get(); + case 'AnyEnumType': return torch.Type.get('AnyEnumType'); + case 'QScheme': return torch.Type.get('QSchemeType'); + case 'Stream': return torch.StreamObjType.get(); + case 'Storage': return torch.Type.get('Storage'); + case 'AnyClassType': return torch.Type.get('AnyClassType'); + case 'NoneType': return torch.NoneType.get(); default: throw new python.Error(`Unsupported type '${value}'.`); } } @@ -6652,7 +6714,7 @@ python.Execution = class { L.eat(','); L.whitespace(0); } - real_value = new torch.TupleType(types); + real_value = torch.TupleType.get(types); fake_value = real_value; } else if (L.value === 'Future') { L.next(); @@ -6661,7 +6723,7 @@ python.Execution = class { const subtype = p.first; // const subalias = p.second; L.expect(')'); - real_value = new torch.FutureType(subtype); + real_value = torch.FutureType.get(subtype); fake_value = real_value; } else if (L.value === 'Await') { L.next(); @@ -6670,7 +6732,7 @@ python.Execution = class { const subtype = p.first; // const subalias = p.second; L.expect(')'); - real_value = new torch.AwaitType(subtype); + real_value = torch.AwaitType.get(subtype); fake_value = real_value; } else if (L.value === 'RRef') { L.next(); @@ -6679,11 +6741,11 @@ python.Execution = class { const subtype = p.first; // const subalias = p.second; L.expect(')'); - real_value = new torch.RRefType(subtype); + real_value = torch.RRefType.get(subtype); fake_value = real_value; } else if (L.value === 'Tensor') { L.next(); - real_value = new torch.TensorType(); + real_value = torch.TensorType.get(); fake_value = real_value; alias_info = this.parseAliasAnnotation(); } else if (L.value === 'Dict') { @@ -6695,7 +6757,7 @@ python.Execution = class { const value_type = this.parseType().first; L.expect(')'); alias_info = this.parseAliasAnnotation(); - real_value = new torch.DictType(key_type, value_type); + real_value = torch.DictType.get(key_type, value_type); fake_value = real_value; } else if (L.eat('Union')) { L.next(); @@ -6727,15 +6789,15 @@ python.Execution = class { real_value.kind() === 'MemoryFormat' || real_value.kind() === 'Layout' || real_value.kind() === 'SymInt') { - fake_value = new torch.IntType(); + fake_value = torch.IntType.get(); } alias_info = this.parseAliasAnnotation(); } while (true) { if (L.kind === '[]') { L.expect('[]'); - fake_value = new torch.ListType(fake_value); - real_value = new torch.ListType(real_value); + fake_value = torch.ListType.get(fake_value); + real_value = torch.ListType.get(real_value); let container = this.parseAliasAnnotation(); if (alias_info) { if (!container) { @@ -6746,8 +6808,8 @@ python.Execution = class { } alias_info = container; } else if (L.eat('?')) { - fake_value = new torch.OptionalType(fake_value); - real_value = new torch.OptionalType(real_value); + fake_value = torch.OptionalType.get(fake_value); + real_value = torch.OptionalType.get(real_value); } else { break; } @@ -6804,8 +6866,8 @@ python.Execution = class { L.whitespace(0); let N = null; if (L.eat('[')) { - fake_type = new torch.ListType(fake_type); - real_type = new torch.ListType(real_type); + fake_type = torch.ListType.get(fake_type); + real_type = torch.ListType.get(real_type); if (L.kind === '#') { N = Number(L.value); L.next(); @@ -6822,9 +6884,9 @@ python.Execution = class { alias_info = container; if (L.eat('?')) { /* eslint-disable no-unused-vars */ - fake_type = new torch.OptionalType(fake_type); + fake_type = torch.OptionalType.get(fake_type); /* eslint-enable no-unused-vars */ - real_type = new torch.OptionalType(real_type); + real_type = torch.OptionalType.get(real_type); } } let name = null; @@ -6988,7 +7050,7 @@ python.Execution = class { } _parse() { if (this._buffer) { - const L = new torch._C.FunctionSchemaLexer(this._buffer); + const L = new torch._C.SchemaLexer(this._buffer); this._arguments = []; this._is_vararg = false; this._kwarg_only = false; @@ -7080,6 +7142,7 @@ python.Execution = class { const returns = this.returns; const braces = !this.is_varret && (returns.length !== 1 || + returns[0].name || returns[0].real_type instanceof torch.TupleType || returns[0].real_type instanceof torch.ListType && returns[0].real_type.getElementType() instanceof torch.TupleType); if (braces) { diff --git a/source/pytorch-metadata.json b/source/pytorch-metadata.json index c8538e1af1..0e78cf1c1a 100755 --- a/source/pytorch-metadata.json +++ b/source/pytorch-metadata.json @@ -32,7 +32,7 @@ ] }, { - "name": "_caffe2::BBoxTransform", + "name": "_caffe2::BBoxTransform(Tensor rois, Tensor deltas, Tensor im_info, float[] weights, bool apply_scale, bool rotated, bool angle_bound_on, int angle_bound_lo, int angle_bound_hi, float clip_angle_thresh, bool legacy_plus_one) -> (Tensor output_0, Tensor output_1)", "inputs": [ { "name": "rois", "type": "Tensor" }, { "name": "deltas", "type": "Tensor" }, @@ -52,7 +52,7 @@ ] }, { - "name": "_caffe2::BatchPermutation", + "name": "_caffe2::BatchPermutation(Tensor X, Tensor indices) -> Tensor", "inputs": [ { "name": "X", "type": "Tensor" }, { "name": "indices", "type": "Tensor" } @@ -62,7 +62,7 @@ ] }, { - "name": "_caffe2::BoxWithNMSLimit", + "name": "_caffe2::BoxWithNMSLimit(Tensor scores, Tensor boxes, Tensor batch_splits, float score_thresh, float nms, int detections_per_im, bool soft_nms_enabled, str soft_nms_method, float soft_nms_sigma, float soft_nms_min_score_thres, bool rotated, bool cls_agnostic_bbox_reg, bool input_boxes_include_bg_cls, bool output_classes_include_bg_cls, bool legacy_plus_one) -> (Tensor scores, Tensor boxes, Tensor classes, Tensor batch_splits, Tensor keeps, Tensor keeps_size)", "inputs": [ { "name": "scores", "type": "Tensor" }, { "name": "boxes", "type": "Tensor" }, @@ -90,7 +90,7 @@ ] }, { - "name": "_caffe2::CollectRpnProposals", + "name": "_caffe2::CollectRpnProposals(Tensor[] input_list, int rpn_max_level, int rpn_min_level, int rpn_post_nms_topN) -> (Tensor rois)", "inputs": [ { "name": "input_list", "type": "Tensor[]" }, { "name": "rpn_max_level", "type": "int64" }, @@ -102,7 +102,7 @@ ] }, { - "name": "_caffe2::CopyCPUToGPU", + "name": "_caffe2::CopyCPUToGPU(Tensor input) -> Tensor", "inputs": [ { "name": "input", "type": "Tensor" } ], @@ -111,7 +111,7 @@ ] }, { - "name": "_caffe2::CopyGPUToCPU", + "name": "_caffe2::CopyGPUToCPU(Tensor input) -> Tensor", "inputs": [ { "name": "input", "type": "Tensor" } ], @@ -120,7 +120,7 @@ ] }, { - "name": "_caffe2::DistributeFpnProposals", + "name": "_caffe2::DistributeFpnProposals(Tensor rois, int roi_canonical_scale, int roi_canonical_level, int roi_max_level, int roi_min_level, bool legacy_plus_one) -> (Tensor rois_fpn2, Tensor rois_fpn3, Tensor rois_fpn4, Tensor rois_fpn5, Tensor rois_idx_restore_int32)", "inputs": [ { "name": "rois", "type": "Tensor" }, { "name": "roi_canonical_scale", "type": "int64" }, @@ -138,7 +138,7 @@ ] }, { - "name": "_caffe2::GenerateProposals", + "name": "_caffe2::GenerateProposals(Tensor scores, Tensor bbox_deltas, Tensor im_info, Tensor anchors, float spatial_scale, int pre_nms_topN, int post_nms_topN, float nms_thresh, float min_size, bool angle_bound_on, int angle_bound_lo, int angle_bound_hi, float clip_angle_thresh, bool legacy_plus_one) -> (Tensor output_0, Tensor output_1)", "inputs": [ { "name": "scores", "type": "Tensor" }, { "name": "bbox_deltas", "type": "Tensor" }, @@ -161,7 +161,7 @@ ] }, { - "name": "_caffe2::RoIAlign", + "name": "_caffe2::RoIAlign(Tensor features, Tensor rois, str order, float spatial_scale, int pooled_h, int pooled_w, int sampling_ratio, bool aligned) -> Tensor", "inputs": [ { "name": "features", "type": "Tensor" }, { "name": "rois", "type": "Tensor" }, @@ -177,7 +177,7 @@ ] }, { - "name": "aten::Bool.Tensor", + "name": "aten::Bool.Tensor(Tensor a) -> bool", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -186,7 +186,7 @@ ] }, { - "name": "aten::Bool.float", + "name": "aten::Bool.float(float a) -> bool", "inputs": [ { "name": "a", "type": "float32" } ], @@ -195,7 +195,7 @@ ] }, { - "name": "aten::Bool.int", + "name": "aten::Bool.int(int a) -> bool", "inputs": [ { "name": "a", "type": "int64" } ], @@ -204,7 +204,7 @@ ] }, { - "name": "aten::Complex.Scalar", + "name": "aten::Complex.Scalar(Scalar a) -> complex", "inputs": [ { "name": "a", "type": "Scalar" } ], @@ -213,7 +213,7 @@ ] }, { - "name": "aten::Complex.Tensor_Tensor", + "name": "aten::Complex.Tensor_Tensor(Tensor a, Tensor b) -> complex", "inputs": [ { "name": "a", "type": "Tensor" }, { "name": "b", "type": "Tensor" } @@ -223,7 +223,7 @@ ] }, { - "name": "aten::Complex.Tensor_bool", + "name": "aten::Complex.Tensor_bool(Tensor x, bool y) -> complex", "inputs": [ { "name": "x", "type": "Tensor" }, { "name": "y", "type": "boolean" } @@ -233,7 +233,7 @@ ] }, { - "name": "aten::Complex.Tensor_float", + "name": "aten::Complex.Tensor_float(Tensor x, float y) -> complex", "inputs": [ { "name": "x", "type": "Tensor" }, { "name": "y", "type": "float32" } @@ -243,7 +243,7 @@ ] }, { - "name": "aten::Complex.Tensor_int", + "name": "aten::Complex.Tensor_int(Tensor x, int y) -> complex", "inputs": [ { "name": "x", "type": "Tensor" }, { "name": "y", "type": "int64" } @@ -253,7 +253,7 @@ ] }, { - "name": "aten::Complex.bool_Tensor", + "name": "aten::Complex.bool_Tensor(bool x, Tensor y) -> complex", "inputs": [ { "name": "x", "type": "boolean" }, { "name": "y", "type": "Tensor" } @@ -263,7 +263,7 @@ ] }, { - "name": "aten::Complex.bool_bool", + "name": "aten::Complex.bool_bool(bool x, bool y) -> complex", "inputs": [ { "name": "x", "type": "boolean" }, { "name": "y", "type": "boolean" } @@ -273,7 +273,7 @@ ] }, { - "name": "aten::Complex.bool_float", + "name": "aten::Complex.bool_float(bool x, float y) -> complex", "inputs": [ { "name": "x", "type": "boolean" }, { "name": "y", "type": "float32" } @@ -283,7 +283,7 @@ ] }, { - "name": "aten::Complex.bool_int", + "name": "aten::Complex.bool_int(bool x, int y) -> complex", "inputs": [ { "name": "x", "type": "boolean" }, { "name": "y", "type": "int64" } @@ -293,7 +293,7 @@ ] }, { - "name": "aten::Complex.float_Tensor", + "name": "aten::Complex.float_Tensor(float x, Tensor y) -> complex", "inputs": [ { "name": "x", "type": "float32" }, { "name": "y", "type": "Tensor" } @@ -303,7 +303,7 @@ ] }, { - "name": "aten::Complex.float_bool", + "name": "aten::Complex.float_bool(float x, bool y) -> complex", "inputs": [ { "name": "x", "type": "float32" }, { "name": "y", "type": "boolean" } @@ -313,7 +313,7 @@ ] }, { - "name": "aten::Complex.float_float", + "name": "aten::Complex.float_float(float x, float y) -> complex", "inputs": [ { "name": "x", "type": "float32" }, { "name": "y", "type": "float32" } @@ -323,7 +323,7 @@ ] }, { - "name": "aten::Complex.float_int", + "name": "aten::Complex.float_int(float x, int y) -> complex", "inputs": [ { "name": "x", "type": "float32" }, { "name": "y", "type": "int64" } @@ -333,7 +333,7 @@ ] }, { - "name": "aten::Complex.int_Tensor", + "name": "aten::Complex.int_Tensor(int x, Tensor y) -> complex", "inputs": [ { "name": "x", "type": "int64" }, { "name": "y", "type": "Tensor" } @@ -343,7 +343,7 @@ ] }, { - "name": "aten::Complex.int_bool", + "name": "aten::Complex.int_bool(int x, bool y) -> complex", "inputs": [ { "name": "x", "type": "int64" }, { "name": "y", "type": "boolean" } @@ -353,7 +353,7 @@ ] }, { - "name": "aten::Complex.int_float", + "name": "aten::Complex.int_float(int x, float y) -> complex", "inputs": [ { "name": "x", "type": "int64" }, { "name": "y", "type": "float32" } @@ -363,7 +363,7 @@ ] }, { - "name": "aten::Complex.int_int", + "name": "aten::Complex.int_int(int x, int y) -> complex", "inputs": [ { "name": "x", "type": "int64" }, { "name": "y", "type": "int64" } @@ -373,7 +373,7 @@ ] }, { - "name": "aten::ComplexImplicit", + "name": "aten::ComplexImplicit(Tensor a) -> complex", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -382,7 +382,7 @@ ] }, { - "name": "aten::Float.Scalar", + "name": "aten::Float.Scalar(Scalar a) -> float", "inputs": [ { "name": "a", "type": "Scalar" } ], @@ -391,7 +391,7 @@ ] }, { - "name": "aten::Float.Tensor", + "name": "aten::Float.Tensor(Tensor a) -> float", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -400,7 +400,7 @@ ] }, { - "name": "aten::Float.bool", + "name": "aten::Float.bool(bool a) -> float", "inputs": [ { "name": "a", "type": "boolean" } ], @@ -409,7 +409,7 @@ ] }, { - "name": "aten::Float.int", + "name": "aten::Float.int(int a) -> float", "inputs": [ { "name": "a", "type": "int64" } ], @@ -418,7 +418,7 @@ ] }, { - "name": "aten::Float.str", + "name": "aten::Float.str(str a) -> float", "inputs": [ { "name": "a", "type": "string" } ], @@ -427,7 +427,7 @@ ] }, { - "name": "aten::FloatImplicit", + "name": "aten::FloatImplicit(Tensor a) -> float", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -436,7 +436,7 @@ ] }, { - "name": "aten::Int.Scalar", + "name": "aten::Int.Scalar(Scalar a) -> int", "inputs": [ { "name": "a", "type": "Scalar" } ], @@ -445,7 +445,7 @@ ] }, { - "name": "aten::Int.Tensor", + "name": "aten::Int.Tensor(Tensor a) -> int", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -454,7 +454,7 @@ ] }, { - "name": "aten::Int.bool", + "name": "aten::Int.bool(bool a) -> int", "inputs": [ { "name": "a", "type": "boolean" } ], @@ -463,7 +463,7 @@ ] }, { - "name": "aten::Int.float", + "name": "aten::Int.float(float a) -> int", "inputs": [ { "name": "a", "type": "float32" } ], @@ -472,7 +472,7 @@ ] }, { - "name": "aten::Int.str", + "name": "aten::Int.str(str a) -> int", "inputs": [ { "name": "a", "type": "string" } ], @@ -481,7 +481,7 @@ ] }, { - "name": "aten::IntImplicit", + "name": "aten::IntImplicit(Tensor a) -> int", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -490,7 +490,7 @@ ] }, { - "name": "aten::ScalarImplicit", + "name": "aten::ScalarImplicit(Tensor a) -> Scalar", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -499,7 +499,7 @@ ] }, { - "name": "aten::__and__.Scalar", + "name": "aten::__and__.Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -509,7 +509,7 @@ ] }, { - "name": "aten::__and__.Tensor", + "name": "aten::__and__.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -519,7 +519,7 @@ ] }, { - "name": "aten::__and__.bool", + "name": "aten::__and__.bool(bool a, bool b) -> bool", "inputs": [ { "name": "a", "type": "boolean" }, { "name": "b", "type": "boolean" } @@ -529,7 +529,7 @@ ] }, { - "name": "aten::__and__.int", + "name": "aten::__and__.int(int a, int b) -> int", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "int64" } @@ -539,7 +539,7 @@ ] }, { - "name": "aten::__getitem__.Dict_Tensor", + "name": "aten::__getitem__.Dict_Tensor(Dict(Tensor, t) self, Tensor key) -> t(*)", "inputs": [ { "name": "self", "type": "Dict(Tensor, t)" }, { "name": "key", "type": "Tensor" } @@ -549,7 +549,7 @@ ] }, { - "name": "aten::__getitem__.Dict_bool", + "name": "aten::__getitem__.Dict_bool(Dict(bool, t) self, bool key) -> t(*)", "inputs": [ { "name": "self", "type": "Dict(boolean, t)" }, { "name": "key", "type": "boolean" } @@ -559,7 +559,7 @@ ] }, { - "name": "aten::__getitem__.Dict_complex", + "name": "aten::__getitem__.Dict_complex(Dict(complex, t) self, complex key) -> t(*)", "inputs": [ { "name": "self", "type": "Dict(complex, t)" }, { "name": "key", "type": "complex" } @@ -569,7 +569,7 @@ ] }, { - "name": "aten::__getitem__.Dict_float", + "name": "aten::__getitem__.Dict_float(Dict(float, t) self, float key) -> t(*)", "inputs": [ { "name": "self", "type": "Dict(float32, t)" }, { "name": "key", "type": "float32" } @@ -579,7 +579,7 @@ ] }, { - "name": "aten::__getitem__.Dict_int", + "name": "aten::__getitem__.Dict_int(Dict(int, t) self, int key) -> t(*)", "inputs": [ { "name": "self", "type": "Dict(int64, t)" }, { "name": "key", "type": "int64" } @@ -589,7 +589,7 @@ ] }, { - "name": "aten::__getitem__.Dict_str", + "name": "aten::__getitem__.Dict_str(Dict(str, t) self, str key) -> t(*)", "inputs": [ { "name": "self", "type": "Dict(string, t)" }, { "name": "key", "type": "string" } @@ -599,7 +599,7 @@ ] }, { - "name": "aten::__getitem__.str", + "name": "aten::__getitem__.str(str s, int index) -> str", "inputs": [ { "name": "s", "type": "string" }, { "name": "index", "type": "int64" } @@ -609,7 +609,7 @@ ] }, { - "name": "aten::__getitem__.t", + "name": "aten::__getitem__.t(t[](a) list, int idx) -> t(*)", "inputs": [ { "name": "list", "type": "t[]" }, { "name": "idx", "type": "int64" } @@ -619,7 +619,7 @@ ] }, { - "name": "aten::__iand__.Scalar", + "name": "aten::__iand__.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -629,7 +629,7 @@ ] }, { - "name": "aten::__iand__.Tensor", + "name": "aten::__iand__.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -639,7 +639,7 @@ ] }, { - "name": "aten::__ilshift__.Scalar", + "name": "aten::__ilshift__.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -649,7 +649,7 @@ ] }, { - "name": "aten::__ilshift__.Tensor", + "name": "aten::__ilshift__.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -659,7 +659,7 @@ ] }, { - "name": "aten::__interpolate", + "name": "aten::__interpolate(Tensor input, int? size = None, float? scale_factor = None, str mode = 'nearest', bool? align_corners = None, bool? recompute_scale_factor = None, bool antialias = False) -> Tensor", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "size", "type": "int64?", "default": null }, @@ -674,7 +674,7 @@ ] }, { - "name": "aten::__interpolate.scale_list", + "name": "aten::__interpolate.scale_list(Tensor input, int? size = None, float[]? scale_factor = None, str mode = 'nearest', bool? align_corners = None, bool? recompute_scale_factor = None, bool antialias = False) -> Tensor", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "size", "type": "int64?", "default": null }, @@ -689,7 +689,7 @@ ] }, { - "name": "aten::__interpolate.size_list", + "name": "aten::__interpolate.size_list(Tensor input, int[]? size = None, float? scale_factor = None, str mode = 'nearest', bool? align_corners = None, bool? recompute_scale_factor = None, bool antialias = False) -> Tensor", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "size", "type": "int64[]?", "default": null }, @@ -704,7 +704,7 @@ ] }, { - "name": "aten::__interpolate.size_list_scale_list", + "name": "aten::__interpolate.size_list_scale_list(Tensor input, int[]? size = None, float[]? scale_factor = None, str mode = 'nearest', bool? align_corners = None, bool? recompute_scale_factor = None, bool antialias = False) -> Tensor", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "size", "type": "int64[]?", "default": null }, @@ -719,7 +719,7 @@ ] }, { - "name": "aten::__irshift__.Scalar", + "name": "aten::__irshift__.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -729,7 +729,7 @@ ] }, { - "name": "aten::__irshift__.Tensor", + "name": "aten::__irshift__.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -739,7 +739,7 @@ ] }, { - "name": "aten::__isnot__", + "name": "aten::__isnot__(t1 self, t2 obj) -> bool", "inputs": [ { "name": "self", "type": "t1" }, { "name": "obj", "type": "t2" } @@ -749,7 +749,7 @@ ] }, { - "name": "aten::__ixor__.Scalar", + "name": "aten::__ixor__.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -759,7 +759,7 @@ ] }, { - "name": "aten::__ixor__.Tensor", + "name": "aten::__ixor__.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -769,7 +769,7 @@ ] }, { - "name": "aten::__lshift__.Scalar", + "name": "aten::__lshift__.Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -779,7 +779,7 @@ ] }, { - "name": "aten::__lshift__.Tensor", + "name": "aten::__lshift__.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -789,7 +789,7 @@ ] }, { - "name": "aten::__not__", + "name": "aten::__not__(bool self) -> bool", "inputs": [ { "name": "self", "type": "boolean" } ], @@ -798,7 +798,7 @@ ] }, { - "name": "aten::__or__.Scalar", + "name": "aten::__or__.Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -808,7 +808,7 @@ ] }, { - "name": "aten::__or__.Tensor", + "name": "aten::__or__.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -818,7 +818,7 @@ ] }, { - "name": "aten::__rshift__.Scalar", + "name": "aten::__rshift__.Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -828,7 +828,7 @@ ] }, { - "name": "aten::__rshift__.Tensor", + "name": "aten::__rshift__.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -838,7 +838,7 @@ ] }, { - "name": "aten::__upsample", + "name": "aten::__upsample(Tensor input, int? size = None, int? scale_factor = None, str mode = 'nearest', bool? align_corners = None) -> Tensor", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -852,7 +852,7 @@ ] }, { - "name": "aten::__upsample.size_list", + "name": "aten::__upsample.size_list(Tensor input, int[]? size = None, int? scale_factor = None, str mode = 'nearest', bool? align_corners = None) -> Tensor", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -866,7 +866,7 @@ ] }, { - "name": "aten::__upsample_bilinear", + "name": "aten::__upsample_bilinear(Tensor input, int? size = None, int? scale_factor = None) -> Tensor", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "size", "type": "int64?", "default": null }, @@ -877,7 +877,7 @@ ] }, { - "name": "aten::__upsample_bilinear.scale_list", + "name": "aten::__upsample_bilinear.scale_list(Tensor input, int? size = None, int[]? scale_factor = None) -> Tensor", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "size", "type": "int64?", "default": null }, @@ -888,7 +888,7 @@ ] }, { - "name": "aten::__upsample_bilinear.size_list", + "name": "aten::__upsample_bilinear.size_list(Tensor input, int[]? size = None, int? scale_factor = None) -> Tensor", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "size", "type": "int64[]?", "default": null }, @@ -899,7 +899,7 @@ ] }, { - "name": "aten::__upsample_bilinear.size_list_scale_list", + "name": "aten::__upsample_bilinear.size_list_scale_list(Tensor input, int[]? size = None, int[]? scale_factor = None) -> Tensor", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "size", "type": "int64[]?", "default": null }, @@ -910,7 +910,7 @@ ] }, { - "name": "aten::__xor__.Scalar", + "name": "aten::__xor__.Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -920,7 +920,7 @@ ] }, { - "name": "aten::__xor__.Tensor", + "name": "aten::__xor__.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -930,7 +930,7 @@ ] }, { - "name": "aten::_adaptive_avg_pool2d", + "name": "aten::_adaptive_avg_pool2d(Tensor self, SymInt[2] output_size) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[2]" } @@ -940,7 +940,7 @@ ] }, { - "name": "aten::_add_relu.Scalar", + "name": "aten::_add_relu.Scalar(Tensor self, Scalar other, Scalar alpha=1) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" }, @@ -951,7 +951,7 @@ ] }, { - "name": "aten::_add_relu.Tensor", + "name": "aten::_add_relu.Tensor(Tensor self, Tensor other, *, Scalar alpha=1) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -962,7 +962,7 @@ ] }, { - "name": "aten::_add_relu.out", + "name": "aten::_add_relu.out(Tensor self, Tensor other, *, Scalar alpha=1, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -973,7 +973,7 @@ ] }, { - "name": "aten::_add_relu_.Scalar", + "name": "aten::_add_relu_.Scalar(Tensor(a!) self, Scalar other, Scalar alpha=1) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" }, @@ -984,7 +984,7 @@ ] }, { - "name": "aten::_add_relu_.Tensor", + "name": "aten::_add_relu_.Tensor(Tensor(a!) self, Tensor other, *, Scalar alpha=1) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -995,7 +995,7 @@ ] }, { - "name": "aten::_aminmax", + "name": "aten::_aminmax(Tensor self) -> (Tensor, Tensor)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -1005,7 +1005,7 @@ ] }, { - "name": "aten::_aminmax.dim", + "name": "aten::_aminmax.dim(Tensor self, int dim, bool keepdim=False) -> (Tensor, Tensor)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -1017,7 +1017,7 @@ ] }, { - "name": "aten::_autocast_to_reduced_precision", + "name": "aten::_autocast_to_reduced_precision(Tensor(a) self, bool cuda_enabled, bool cpu_enabled, ScalarType cuda_dtype, ScalarType cpu_dtype) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "cuda_enabled", "type": "boolean" }, @@ -1030,7 +1030,7 @@ ] }, { - "name": "aten::_cast_Byte", + "name": "aten::_cast_Byte(Tensor self, bool non_blocking=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "non_blocking", "type": "boolean", "default": false } @@ -1040,7 +1040,7 @@ ] }, { - "name": "aten::_cast_Char", + "name": "aten::_cast_Char(Tensor self, bool non_blocking=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "non_blocking", "type": "boolean", "default": false } @@ -1050,7 +1050,7 @@ ] }, { - "name": "aten::_cast_Double", + "name": "aten::_cast_Double(Tensor self, bool non_blocking=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "non_blocking", "type": "boolean", "default": false } @@ -1060,7 +1060,7 @@ ] }, { - "name": "aten::_cast_Float", + "name": "aten::_cast_Float(Tensor self, bool non_blocking=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "non_blocking", "type": "boolean", "default": false } @@ -1070,7 +1070,7 @@ ] }, { - "name": "aten::_cast_Half", + "name": "aten::_cast_Half(Tensor self, bool non_blocking=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "non_blocking", "type": "boolean", "default": false } @@ -1080,7 +1080,7 @@ ] }, { - "name": "aten::_cast_Int", + "name": "aten::_cast_Int(Tensor self, bool non_blocking=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "non_blocking", "type": "boolean", "default": false } @@ -1090,7 +1090,7 @@ ] }, { - "name": "aten::_cast_Long", + "name": "aten::_cast_Long(Tensor self, bool non_blocking=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "non_blocking", "type": "boolean", "default": false } @@ -1100,7 +1100,7 @@ ] }, { - "name": "aten::_cast_Short", + "name": "aten::_cast_Short(Tensor self, bool non_blocking=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "non_blocking", "type": "boolean", "default": false } @@ -1110,7 +1110,7 @@ ] }, { - "name": "aten::_cdist_forward", + "name": "aten::_cdist_forward(Tensor x1, Tensor x2, float p, int? compute_mode) -> Tensor", "inputs": [ { "name": "x1", "type": "Tensor" }, { "name": "x2", "type": "Tensor" }, @@ -1122,7 +1122,7 @@ ] }, { - "name": "aten::_coalesce", + "name": "aten::_coalesce(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -1131,7 +1131,7 @@ ] }, { - "name": "aten::_conj", + "name": "aten::_conj(Tensor(a) self) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -1140,7 +1140,7 @@ ] }, { - "name": "aten::_convolution", + "name": "aten::_convolution(Tensor input, Tensor weight, Tensor? bias, SymInt[] stride, SymInt[] padding, SymInt[] dilation, bool transposed, SymInt[] output_padding, SymInt groups, bool benchmark, bool deterministic, bool cudnn_enabled, bool allow_tf32) -> Tensor", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -1162,7 +1162,7 @@ ] }, { - "name": "aten::_convolution.deprecated", + "name": "aten::_convolution.deprecated(Tensor input, Tensor weight, Tensor? bias, SymInt[] stride, SymInt[] padding, SymInt[] dilation, bool transposed, int[] output_padding, SymInt groups, bool benchmark, bool deterministic, bool cudnn_enabled) -> Tensor", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -1183,7 +1183,7 @@ ] }, { - "name": "aten::_convolution_mode", + "name": "aten::_convolution_mode(Tensor input, Tensor weight, Tensor? bias, SymInt[] stride, str padding, SymInt[] dilation, SymInt groups) -> Tensor", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "weight", "type": "Tensor" }, @@ -1198,7 +1198,7 @@ ] }, { - "name": "aten::_ctc_loss", + "name": "aten::_ctc_loss(Tensor log_probs, Tensor targets, int[] input_lengths, int[] target_lengths, int blank=0, bool zero_infinity=False) -> (Tensor, Tensor)", "inputs": [ { "name": "log_probs", "type": "Tensor" }, { "name": "targets", "type": "Tensor" }, @@ -1213,7 +1213,7 @@ ] }, { - "name": "aten::_ctc_loss.Tensor", + "name": "aten::_ctc_loss.Tensor(Tensor log_probs, Tensor targets, Tensor input_lengths, Tensor target_lengths, int blank=0, bool zero_infinity=False) -> (Tensor, Tensor)", "inputs": [ { "name": "log_probs", "type": "Tensor" }, { "name": "targets", "type": "Tensor" }, @@ -1228,7 +1228,7 @@ ] }, { - "name": "aten::_dim_arange", + "name": "aten::_dim_arange(Tensor like, int dim) -> Tensor", "inputs": [ { "name": "like", "type": "Tensor" }, { "name": "dim", "type": "int64" } @@ -1238,7 +1238,7 @@ ] }, { - "name": "aten::_fake_quantize_learnable_per_tensor_affine", + "name": "aten::_fake_quantize_learnable_per_tensor_affine(Tensor self, Tensor scale, Tensor zero_point, int quant_min, int quant_max, float grad_factor=1.0) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "scale", "type": "Tensor" }, @@ -1252,7 +1252,7 @@ ] }, { - "name": "aten::_fake_quantize_learnable_per_tensor_affine_backward", + "name": "aten::_fake_quantize_learnable_per_tensor_affine_backward(Tensor grad, Tensor self, Tensor scale, Tensor zero_point, int quant_min, int quant_max, float grad_factor=1.0) -> (Tensor, Tensor, Tensor)", "inputs": [ { "name": "grad", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -1269,7 +1269,7 @@ ] }, { - "name": "aten::_make_per_tensor_quantized_tensor", + "name": "aten::_make_per_tensor_quantized_tensor(Tensor self, float scale, int zero_point) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "scale", "type": "float32" }, @@ -1280,12 +1280,12 @@ ] }, { - "name": "aten::_native_batch_norm_legit_functional", + "name": "aten::_native_batch_norm_legit_functional(Tensor input, Tensor? weight, Tensor? bias, Tensor running_mean, Tensor running_var, bool training, float momentum, float eps) -> (Tensor, Tensor, Tensor, Tensor running_mean_out, Tensor running_var_out)", "category": "Normalization", "inputs": [ { "name": "input", "type": "Tensor" }, - { "name": "weight", "type": "Tensor", "optional": true }, - { "name": "bias", "type": "Tensor", "optional": true }, + { "name": "weight", "type": "Tensor?" }, + { "name": "bias", "type": "Tensor?" }, { "name": "running_mean", "type": "Tensor" }, { "name": "running_var", "type": "Tensor" }, { "name": "training", "type": "boolean" }, @@ -1295,11 +1295,13 @@ "outputs": [ { "name": "output", "type": "Tensor" }, { "name": "save_mean", "type": "Tensor" }, - { "name": "save_rstd", "type": "Tensor" } + { "name": "save_rstd", "type": "Tensor" }, + { "name": "running_mean_out", "type": "Tensor" }, + { "name": "running_var_out", "type": "Tensor" } ] }, { - "name": "aten::_native_batch_norm_legit_no_training", + "name": "aten::_native_batch_norm_legit_no_training(Tensor input, Tensor? weight, Tensor? bias, Tensor running_mean, Tensor running_var, float momentum, float eps) -> (Tensor, Tensor, Tensor)", "category": "Normalization", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -1317,7 +1319,24 @@ ] }, { - "name": "aten::_native_multi_head_attention", + "name": "aten::_native_batch_norm_legit_no_training.out(Tensor input, Tensor? weight, Tensor? bias, Tensor running_mean, Tensor running_var, float momentum, float eps, *, Tensor(a!) out0, Tensor(b!) out1, Tensor(c!) out2) -> (Tensor(a!), Tensor(b!), Tensor(c!))", + "inputs": [ + { "name": "input", "type": "Tensor" }, + { "name": "weight", "type": "Tensor?" }, + { "name": "bias", "type": "Tensor?" }, + { "name": "running_mean", "type": "Tensor" }, + { "name": "running_var", "type": "Tensor" }, + { "name": "momentum", "type": "float32" }, + { "name": "eps", "type": "float32" } + ], + "outputs": [ + { "type": "Tensor" }, + { "type": "Tensor" }, + { "type": "Tensor" } + ] + }, + { + "name": "aten::_native_multi_head_attention(Tensor query, Tensor key, Tensor value, int embed_dim, int num_head, Tensor qkv_weight, Tensor qkv_bias, Tensor proj_weight, Tensor proj_bias, Tensor? mask=None, bool need_weights=True, bool average_attn_weights=True, int? mask_type=None) -> (Tensor, Tensor)", "category": "Attention", "inputs": [ { "name": "query", "type": "Tensor" }, @@ -1340,7 +1359,29 @@ ] }, { - "name": "aten::_nested_tensor_from_mask", + "name": "aten::_native_multi_head_attention.out(Tensor query, Tensor key, Tensor value, int embed_dim, int num_head, Tensor qkv_weight, Tensor qkv_bias, Tensor proj_weight, Tensor proj_bias, Tensor? mask=None, bool need_weights=True, bool average_attn_weights=True, int? mask_type=None, *, Tensor(a!) out0, Tensor(b!) out1) -> (Tensor(a!), Tensor(b!))", + "inputs": [ + { "name": "query", "type": "Tensor" }, + { "name": "key", "type": "Tensor" }, + { "name": "value", "type": "Tensor" }, + { "name": "embed_dim", "type": "int64" }, + { "name": "num_head", "type": "int64" }, + { "name": "qkv_weight", "type": "Tensor" }, + { "name": "qkv_bias", "type": "Tensor" }, + { "name": "proj_weight", "type": "Tensor" }, + { "name": "proj_bias", "type": "Tensor" }, + { "name": "mask", "type": "Tensor?", "default": null }, + { "name": "need_weights", "type": "boolean", "default": true }, + { "name": "average_attn_weights", "type": "boolean", "default": true }, + { "name": "mask_type", "type": "int64?", "default": null } + ], + "outputs": [ + { "type": "Tensor" }, + { "type": "Tensor" } + ] + }, + { + "name": "aten::_nested_tensor_from_mask(Tensor t, Tensor mask, bool mask_check=True) -> Tensor", "inputs": [ { "name": "t", "type": "Tensor" }, { "name": "mask", "type": "Tensor" }, @@ -1351,7 +1392,7 @@ ] }, { - "name": "aten::_pack_padded_sequence", + "name": "aten::_pack_padded_sequence(Tensor input, Tensor lengths, bool batch_first) -> (Tensor, Tensor)", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "lengths", "type": "Tensor" }, @@ -1363,7 +1404,7 @@ ] }, { - "name": "aten::_pad_packed_sequence", + "name": "aten::_pad_packed_sequence(Tensor data, Tensor batch_sizes, bool batch_first, Scalar padding_value, int total_length) -> (Tensor, Tensor)", "inputs": [ { "name": "data", "type": "Tensor" }, { "name": "batch_sizes", "type": "Tensor" }, @@ -1377,7 +1418,7 @@ ] }, { - "name": "aten::_scaled_dot_product_efficient_attention", + "name": "aten::_scaled_dot_product_efficient_attention(Tensor query, Tensor key, Tensor value, Tensor? attn_bias, bool compute_log_sumexp, float dropout_p=0.0, bool is_causal=False, *, float? scale=None) -> (Tensor output, Tensor log_sumexp, Tensor philox_seed, Tensor philox_offset)", "inputs": [ { "name": "query", "type": "Tensor" }, { "name": "key", "type": "Tensor" }, @@ -1396,7 +1437,7 @@ ] }, { - "name": "aten::_scaled_dot_product_flash_attention_for_cpu", + "name": "aten::_scaled_dot_product_flash_attention_for_cpu(Tensor query, Tensor key, Tensor value, float dropout_p=0.0, bool is_causal=False, *, Tensor? attn_mask=None, float? scale=None) -> (Tensor output, Tensor logsumexp)", "inputs": [ { "name": "query", "type": "Tensor" }, { "name": "key", "type": "Tensor" }, @@ -1412,7 +1453,7 @@ ] }, { - "name": "aten::_shape_as_tensor", + "name": "aten::_shape_as_tensor(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -1421,7 +1462,7 @@ ] }, { - "name": "aten::_softmax", + "name": "aten::_softmax(Tensor self, int dim, bool half_to_float) -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -1433,7 +1474,7 @@ ] }, { - "name": "aten::_softmax.out", + "name": "aten::_softmax.out(Tensor self, int dim, bool half_to_float, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -1444,7 +1485,7 @@ ] }, { - "name": "aten::_sparse_coo_tensor_unsafe", + "name": "aten::_sparse_coo_tensor_unsafe(Tensor indices, Tensor values, SymInt[] size, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None, bool? is_coalesced=None) -> Tensor", "category": "Tensor", "inputs": [ { "name": "indices", "type": "Tensor" }, @@ -1461,7 +1502,7 @@ ] }, { - "name": "aten::_test_serialization_subcmul", + "name": "aten::_test_serialization_subcmul(Tensor self, Tensor other, Scalar alpha=1) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -1472,7 +1513,7 @@ ] }, { - "name": "aten::_thnn_fused_gru_cell", + "name": "aten::_thnn_fused_gru_cell(Tensor input_gates, Tensor hidden_gates, Tensor hx, Tensor? input_bias=None, Tensor? hidden_bias=None) -> (Tensor, Tensor)", "inputs": [ { "name": "input_gates", "type": "Tensor" }, { "name": "hidden_gates", "type": "Tensor" }, @@ -1486,7 +1527,7 @@ ] }, { - "name": "aten::_thnn_fused_lstm_cell", + "name": "aten::_thnn_fused_lstm_cell(Tensor input_gates, Tensor hidden_gates, Tensor cx, Tensor? input_bias=None, Tensor? hidden_bias=None) -> (Tensor, Tensor, Tensor)", "inputs": [ { "name": "input_gates", "type": "Tensor" }, { "name": "hidden_gates", "type": "Tensor" }, @@ -1501,7 +1542,7 @@ ] }, { - "name": "aten::_to_copy", + "name": "aten::_to_copy(Tensor self, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None, bool non_blocking=False, MemoryFormat? memory_format=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -1516,7 +1557,7 @@ ] }, { - "name": "aten::_transformer_encoder_layer_fwd", + "name": "aten::_transformer_encoder_layer_fwd(Tensor src, int embed_dim, int num_heads, Tensor qkv_weight, Tensor qkv_bias, Tensor proj_weight, Tensor proj_bias, bool use_gelu, bool norm_first, float eps, Tensor norm_weight_1, Tensor norm_bias_1, Tensor norm_weight_2, Tensor norm_bias_2, Tensor ffn_weight_1, Tensor ffn_bias_1, Tensor ffn_weight_2, Tensor ffn_bias_2, Tensor? mask=None, int? mask_type=None) -> Tensor", "inputs": [ { "name": "src", "type": "Tensor" }, { "name": "embed_dim", "type": "int64" }, @@ -1544,7 +1585,7 @@ ] }, { - "name": "aten::_unique2", + "name": "aten::_unique2(Tensor self, bool sorted=True, bool return_inverse=False, bool return_counts=False) -> (Tensor, Tensor, Tensor)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "sorted", "type": "boolean", "default": true }, @@ -1558,7 +1599,7 @@ ] }, { - "name": "aten::_unsafe_view", + "name": "aten::_unsafe_view(Tensor self, SymInt[] size) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "size", "type": "SymInt[]" } @@ -1568,7 +1609,7 @@ ] }, { - "name": "aten::_upsample_bilinear2d_aa", + "name": "aten::_upsample_bilinear2d_aa(Tensor self, SymInt[2] output_size, bool align_corners, float? scales_h=None, float? scales_w=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[2]" }, @@ -1581,7 +1622,7 @@ ] }, { - "name": "aten::_upsample_bilinear2d_aa.out", + "name": "aten::_upsample_bilinear2d_aa.out(Tensor self, SymInt[2] output_size, bool align_corners, float? scales_h=None, float? scales_w=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[2]" }, @@ -1594,7 +1635,7 @@ ] }, { - "name": "aten::_upsample_bilinear2d_aa.vec", + "name": "aten::_upsample_bilinear2d_aa.vec(Tensor input, SymInt[]? output_size, bool align_corners, float[]? scale_factors) -> Tensor", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[]?" }, @@ -1606,7 +1647,7 @@ ] }, { - "name": "aten::_weight_norm", + "name": "aten::_weight_norm(Tensor v, Tensor g, int dim=0) -> Tensor", "inputs": [ { "name": "v", "type": "Tensor" }, { "name": "g", "type": "Tensor" }, @@ -1617,7 +1658,7 @@ ] }, { - "name": "aten::_weight_norm_differentiable_backward", + "name": "aten::_weight_norm_differentiable_backward(Tensor grad_w, Tensor saved_v, Tensor saved_g, Tensor saved_norms, int dim) -> (Tensor, Tensor)", "inputs": [ { "name": "grad_w", "type": "Tensor" }, { "name": "saved_v", "type": "Tensor" }, @@ -1631,7 +1672,7 @@ ] }, { - "name": "aten::_weight_norm_interface", + "name": "aten::_weight_norm_interface(Tensor v, Tensor g, int dim=0) -> (Tensor, Tensor)", "inputs": [ { "name": "v", "type": "Tensor" }, { "name": "g", "type": "Tensor" }, @@ -1643,7 +1684,7 @@ ] }, { - "name": "aten::_weight_norm_interface_backward", + "name": "aten::_weight_norm_interface_backward(Tensor grad_w, Tensor saved_v, Tensor saved_g, Tensor saved_norms, int dim) -> (Tensor, Tensor)", "inputs": [ { "name": "grad_w", "type": "Tensor" }, { "name": "saved_v", "type": "Tensor" }, @@ -1657,7 +1698,7 @@ ] }, { - "name": "aten::abs", + "name": "aten::abs(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -1666,7 +1707,7 @@ ] }, { - "name": "aten::abs.out", + "name": "aten::abs.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -1675,7 +1716,7 @@ ] }, { - "name": "aten::abs_", + "name": "aten::abs_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -1684,7 +1725,7 @@ ] }, { - "name": "aten::acos", + "name": "aten::acos(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -1693,7 +1734,7 @@ ] }, { - "name": "aten::acos.out", + "name": "aten::acos.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -1702,7 +1743,7 @@ ] }, { - "name": "aten::acos_", + "name": "aten::acos_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -1711,7 +1752,7 @@ ] }, { - "name": "aten::acosh", + "name": "aten::acosh(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -1720,7 +1761,7 @@ ] }, { - "name": "aten::acosh.out", + "name": "aten::acosh.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -1729,7 +1770,7 @@ ] }, { - "name": "aten::acosh_", + "name": "aten::acosh_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -1738,7 +1779,7 @@ ] }, { - "name": "aten::adaptive_avg_pool1d", + "name": "aten::adaptive_avg_pool1d(Tensor self, int[1] output_size) -> Tensor", "category": "Pool", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -1749,7 +1790,7 @@ ] }, { - "name": "aten::adaptive_avg_pool2d", + "name": "aten::adaptive_avg_pool2d(Tensor self, SymInt[2] output_size) -> Tensor", "category": "Pool", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -1760,7 +1801,7 @@ ] }, { - "name": "aten::adaptive_avg_pool2d.out", + "name": "aten::adaptive_avg_pool2d.out(Tensor self, SymInt[2] output_size, *, Tensor(a!) out) -> Tensor(a!)", "category": "Pool", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -1771,7 +1812,7 @@ ] }, { - "name": "aten::adaptive_avg_pool3d", + "name": "aten::adaptive_avg_pool3d(Tensor self, SymInt[3] output_size) -> Tensor", "category": "Pool", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -1782,7 +1823,7 @@ ] }, { - "name": "aten::adaptive_avg_pool3d.out", + "name": "aten::adaptive_avg_pool3d.out(Tensor self, SymInt[3] output_size, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[3]" } @@ -1792,7 +1833,7 @@ ] }, { - "name": "aten::adaptive_max_pool1d", + "name": "aten::adaptive_max_pool1d(Tensor self, int[1] output_size) -> (Tensor, Tensor)", "category": "Pool", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -1804,7 +1845,7 @@ ] }, { - "name": "aten::adaptive_max_pool2d", + "name": "aten::adaptive_max_pool2d(Tensor self, int[2] output_size) -> (Tensor, Tensor)", "category": "Pool", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -1816,7 +1857,7 @@ ] }, { - "name": "aten::adaptive_max_pool2d.out", + "name": "aten::adaptive_max_pool2d.out(Tensor self, int[2] output_size, *, Tensor(a!) out, Tensor(b!) indices) -> (Tensor(a!), Tensor(b!))", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "output_size", "type": "int64[2]" } @@ -1827,7 +1868,7 @@ ] }, { - "name": "aten::adaptive_max_pool3d", + "name": "aten::adaptive_max_pool3d(Tensor self, int[3] output_size) -> (Tensor, Tensor)", "category": "Pool", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -1839,7 +1880,7 @@ ] }, { - "name": "aten::adaptive_max_pool3d.out", + "name": "aten::adaptive_max_pool3d.out(Tensor self, int[3] output_size, *, Tensor(a!) out, Tensor(b!) indices) -> (Tensor(a!), Tensor(b!))", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "output_size", "type": "int64[3]" } @@ -1850,7 +1891,28 @@ ] }, { - "name": "aten::add.Scalar", + "name": "aten::add(Scalar a, Scalar b) -> Scalar", + "inputs": [ + { "name": "a", "type": "Scalar" }, + { "name": "b", "type": "Scalar" } + ], + "outputs": [ + { "type": "Scalar" } + ] + }, + { + "name": "aten::add.Scalar(Tensor self, Scalar other, Scalar alpha=1) -> Tensor", + "inputs": [ + { "name": "self", "type": "Tensor" }, + { "name": "other", "type": "Scalar" }, + { "name": "alpha", "type": "Scalar", "default": 1 } + ], + "outputs": [ + { "type": "Tensor" } + ] + }, + { + "name": "aten::add.Scalar_out(Tensor self, Scalar other, Scalar alpha=1, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" }, @@ -1861,7 +1923,7 @@ ] }, { - "name": "aten::add.Tensor", + "name": "aten::add.Tensor(Tensor self, Tensor other, *, Scalar alpha=1) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -1872,7 +1934,37 @@ ] }, { - "name": "aten::add.float", + "name": "aten::add.complex(complex a, complex b) -> complex", + "inputs": [ + { "name": "a", "type": "complex" }, + { "name": "b", "type": "complex" } + ], + "outputs": [ + { "type": "complex" } + ] + }, + { + "name": "aten::add.complex_float(complex a, float b) -> complex", + "inputs": [ + { "name": "a", "type": "complex" }, + { "name": "b", "type": "float32" } + ], + "outputs": [ + { "type": "complex" } + ] + }, + { + "name": "aten::add.complex_int(complex a, int b) -> complex", + "inputs": [ + { "name": "a", "type": "complex" }, + { "name": "b", "type": "int64" } + ], + "outputs": [ + { "type": "complex" } + ] + }, + { + "name": "aten::add.float(float a, float b) -> float", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "float32" } @@ -1882,7 +1974,27 @@ ] }, { - "name": "aten::add.int", + "name": "aten::add.float_complex(float a, complex b) -> complex", + "inputs": [ + { "name": "a", "type": "float32" }, + { "name": "b", "type": "complex" } + ], + "outputs": [ + { "type": "complex" } + ] + }, + { + "name": "aten::add.float_int(float a, int b) -> float", + "inputs": [ + { "name": "a", "type": "float32" }, + { "name": "b", "type": "int64" } + ], + "outputs": [ + { "type": "float32" } + ] + }, + { + "name": "aten::add.int(int a, int b) -> int", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "int64" } @@ -1892,7 +2004,27 @@ ] }, { - "name": "aten::add.out", + "name": "aten::add.int_complex(int a, complex b) -> complex", + "inputs": [ + { "name": "a", "type": "int64" }, + { "name": "b", "type": "complex" } + ], + "outputs": [ + { "type": "complex" } + ] + }, + { + "name": "aten::add.int_float(int a, float b) -> float", + "inputs": [ + { "name": "a", "type": "int64" }, + { "name": "b", "type": "float32" } + ], + "outputs": [ + { "type": "float32" } + ] + }, + { + "name": "aten::add.out(Tensor self, Tensor other, *, Scalar alpha=1, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -1903,7 +2035,7 @@ ] }, { - "name": "aten::add.str", + "name": "aten::add.str(str a, str b) -> str", "inputs": [ { "name": "a", "type": "string" }, { "name": "b", "type": "string" } @@ -1913,7 +2045,7 @@ ] }, { - "name": "aten::add.t", + "name": "aten::add.t(t[] a, t[] b) -> t[]", "inputs": [ { "name": "a", "type": "t[]" }, { "name": "b", "type": "t[]" } @@ -1923,7 +2055,7 @@ ] }, { - "name": "aten::add_.Scalar", + "name": "aten::add_.Scalar(Tensor(a!) self, Scalar other, Scalar alpha=1) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" }, @@ -1934,7 +2066,7 @@ ] }, { - "name": "aten::add_.Tensor", + "name": "aten::add_.Tensor(Tensor(a!) self, Tensor other, *, Scalar alpha=1) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -1945,7 +2077,7 @@ ] }, { - "name": "aten::add_.t", + "name": "aten::add_.t(t[](a!) self, t[] b) -> t[]", "inputs": [ { "name": "self", "type": "t[]" }, { "name": "b", "type": "t[]" } @@ -1955,7 +2087,7 @@ ] }, { - "name": "aten::addcdiv", + "name": "aten::addcdiv(Tensor self, Tensor tensor1, Tensor tensor2, *, Scalar value=1) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "tensor1", "type": "Tensor" }, @@ -1967,7 +2099,7 @@ ] }, { - "name": "aten::addcdiv.out", + "name": "aten::addcdiv.out(Tensor self, Tensor tensor1, Tensor tensor2, *, Scalar value=1, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "tensor1", "type": "Tensor" }, @@ -1979,7 +2111,7 @@ ] }, { - "name": "aten::addcmul", + "name": "aten::addcmul(Tensor self, Tensor tensor1, Tensor tensor2, *, Scalar value=1) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "tensor1", "type": "Tensor" }, @@ -1991,7 +2123,7 @@ ] }, { - "name": "aten::addcmul.out", + "name": "aten::addcmul.out(Tensor self, Tensor tensor1, Tensor tensor2, *, Scalar value=1, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "tensor1", "type": "Tensor" }, @@ -2003,7 +2135,7 @@ ] }, { - "name": "aten::addcmul_", + "name": "aten::addcmul_(Tensor(a!) self, Tensor tensor1, Tensor tensor2, *, Scalar value=1) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "tensor1", "type": "Tensor" }, @@ -2015,7 +2147,7 @@ ] }, { - "name": "aten::addmm", + "name": "aten::addmm(Tensor self, Tensor mat1, Tensor mat2, *, Scalar beta=1, Scalar alpha=1) -> Tensor", "category": "Layer", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -2029,7 +2161,7 @@ ] }, { - "name": "aten::addmm.out", + "name": "aten::addmm.out(Tensor self, Tensor mat1, Tensor mat2, *, Scalar beta=1, Scalar alpha=1, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "mat1", "type": "Tensor" }, @@ -2042,7 +2174,7 @@ ] }, { - "name": "aten::addmm_", + "name": "aten::addmm_(Tensor(a!) self, Tensor mat1, Tensor mat2, *, Scalar beta=1, Scalar alpha=1) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "mat1", "type": "Tensor" }, @@ -2055,7 +2187,7 @@ ] }, { - "name": "aten::addmv", + "name": "aten::addmv(Tensor self, Tensor mat, Tensor vec, *, Scalar beta=1, Scalar alpha=1) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "mat", "type": "Tensor" }, @@ -2068,7 +2200,7 @@ ] }, { - "name": "aten::addmv.out", + "name": "aten::addmv.out(Tensor self, Tensor mat, Tensor vec, *, Scalar beta=1, Scalar alpha=1, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "mat", "type": "Tensor" }, @@ -2081,7 +2213,7 @@ ] }, { - "name": "aten::addmv_", + "name": "aten::addmv_(Tensor(a!) self, Tensor mat, Tensor vec, *, Scalar beta=1, Scalar alpha=1) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "mat", "type": "Tensor" }, @@ -2094,7 +2226,7 @@ ] }, { - "name": "aten::affine_grid_generator", + "name": "aten::affine_grid_generator(Tensor theta, SymInt[] size, bool align_corners) -> Tensor", "inputs": [ { "name": "theta", "type": "Tensor" }, { "name": "size", "type": "SymInt[]" }, @@ -2105,7 +2237,7 @@ ] }, { - "name": "aten::alias", + "name": "aten::alias(Tensor(a) self) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -2114,7 +2246,7 @@ ] }, { - "name": "aten::alias_copy", + "name": "aten::alias_copy(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -2123,7 +2255,7 @@ ] }, { - "name": "aten::all", + "name": "aten::all(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -2132,7 +2264,7 @@ ] }, { - "name": "aten::all.all_out", + "name": "aten::all.all_out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -2141,7 +2273,7 @@ ] }, { - "name": "aten::all.bool", + "name": "aten::all.bool(bool[] self) -> bool", "inputs": [ { "name": "self", "type": "boolean[]" } ], @@ -2150,7 +2282,7 @@ ] }, { - "name": "aten::all.dim", + "name": "aten::all.dim(Tensor self, int dim, bool keepdim=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -2161,7 +2293,7 @@ ] }, { - "name": "aten::all.dimname", + "name": "aten::all.dimname(Tensor self, Dimname dim, bool keepdim=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -2172,7 +2304,7 @@ ] }, { - "name": "aten::all.dimname_out", + "name": "aten::all.dimname_out(Tensor self, Dimname dim, bool keepdim=False, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -2183,7 +2315,7 @@ ] }, { - "name": "aten::all.dims", + "name": "aten::all.dims(Tensor self, int[]? dim=None, bool keepdim=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[]?", "default": null }, @@ -2194,7 +2326,7 @@ ] }, { - "name": "aten::all.dims_out", + "name": "aten::all.dims_out(Tensor self, int[]? dim=None, bool keepdim=False, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[]?", "default": null }, @@ -2205,7 +2337,7 @@ ] }, { - "name": "aten::all.float", + "name": "aten::all.float(float[] self) -> bool", "inputs": [ { "name": "self", "type": "float32[]" } ], @@ -2214,7 +2346,7 @@ ] }, { - "name": "aten::all.int", + "name": "aten::all.int(int[] self) -> bool", "inputs": [ { "name": "self", "type": "int64[]" } ], @@ -2223,7 +2355,7 @@ ] }, { - "name": "aten::all.out", + "name": "aten::all.out(Tensor self, int dim, bool keepdim=False, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -2234,7 +2366,7 @@ ] }, { - "name": "aten::alpha_dropout", + "name": "aten::alpha_dropout(Tensor input, float p, bool train) -> Tensor", "category": "Dropout", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -2246,7 +2378,7 @@ ] }, { - "name": "aten::alpha_dropout_", + "name": "aten::alpha_dropout_(Tensor(a!) self, float p, bool train) -> Tensor(a!)", "category": "Dropout", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -2258,7 +2390,7 @@ ] }, { - "name": "aten::amax", + "name": "aten::amax(Tensor self, int[1] dim=[], bool keepdim=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]", "default": [] }, @@ -2269,7 +2401,7 @@ ] }, { - "name": "aten::amax.out", + "name": "aten::amax.out(Tensor self, int[1] dim=[], bool keepdim=False, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]", "default": [] }, @@ -2280,7 +2412,7 @@ ] }, { - "name": "aten::amin", + "name": "aten::amin(Tensor self, int[1] dim=[], bool keepdim=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]", "default": [] }, @@ -2291,7 +2423,7 @@ ] }, { - "name": "aten::amin.out", + "name": "aten::amin.out(Tensor self, int[1] dim=[], bool keepdim=False, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]", "default": [] }, @@ -2302,7 +2434,7 @@ ] }, { - "name": "aten::aminmax", + "name": "aten::aminmax(Tensor self, *, int? dim=None, bool keepdim=False) -> (Tensor min, Tensor max)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64?", "default": null, "kwarg_only": true }, @@ -2314,7 +2446,7 @@ ] }, { - "name": "aten::aminmax.out", + "name": "aten::aminmax.out(Tensor self, *, int? dim=None, bool keepdim=False, Tensor(a!) min, Tensor(b!) max) -> (Tensor(a!) min, Tensor(b!) max)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64?", "default": null, "kwarg_only": true }, @@ -2326,7 +2458,7 @@ ] }, { - "name": "aten::angle", + "name": "aten::angle(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -2335,7 +2467,7 @@ ] }, { - "name": "aten::angle.out", + "name": "aten::angle.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -2344,7 +2476,7 @@ ] }, { - "name": "aten::any", + "name": "aten::any(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -2353,7 +2485,7 @@ ] }, { - "name": "aten::any.all_out", + "name": "aten::any.all_out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -2362,7 +2494,7 @@ ] }, { - "name": "aten::any.bool", + "name": "aten::any.bool(bool[] self) -> bool", "inputs": [ { "name": "self", "type": "boolean[]" } ], @@ -2371,7 +2503,7 @@ ] }, { - "name": "aten::any.dim", + "name": "aten::any.dim(Tensor self, int dim, bool keepdim=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -2382,7 +2514,7 @@ ] }, { - "name": "aten::any.dimname", + "name": "aten::any.dimname(Tensor self, str dim, bool keepdim=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "string" }, @@ -2393,7 +2525,7 @@ ] }, { - "name": "aten::any.dimname_out", + "name": "aten::any.dimname_out(Tensor self, str dim, bool keepdim=False, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "string" }, @@ -2404,7 +2536,7 @@ ] }, { - "name": "aten::any.dims", + "name": "aten::any.dims(Tensor self, int[]? dim=None, bool keepdim=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[]?", "default": null }, @@ -2415,7 +2547,7 @@ ] }, { - "name": "aten::any.dims_out", + "name": "aten::any.dims_out(Tensor self, int[]? dim=None, bool keepdim=False, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[]?", "default": null }, @@ -2426,7 +2558,7 @@ ] }, { - "name": "aten::any.float", + "name": "aten::any.float(float[] self) -> bool", "inputs": [ { "name": "self", "type": "float32[]" } ], @@ -2435,7 +2567,7 @@ ] }, { - "name": "aten::any.int", + "name": "aten::any.int(int[] self) -> bool", "inputs": [ { "name": "self", "type": "int64[]" } ], @@ -2444,7 +2576,7 @@ ] }, { - "name": "aten::any.out", + "name": "aten::any.out(Tensor self, int dim, bool keepdim=False, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -2455,7 +2587,7 @@ ] }, { - "name": "aten::any.str", + "name": "aten::any.str(str[] self) -> bool", "inputs": [ { "name": "self", "type": "string[]" } ], @@ -2464,7 +2596,7 @@ ] }, { - "name": "aten::append.t", + "name": "aten::append.t(t[](a!) self, t(c -> *) el) -> t[](a!)", "inputs": [ { "name": "self", "type": "t[]" }, { "name": "el", "type": "t" } @@ -2474,7 +2606,7 @@ ] }, { - "name": "aten::arange", + "name": "aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "end", "type": "Scalar" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -2487,7 +2619,7 @@ ] }, { - "name": "aten::arange.out", + "name": "aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "end", "type": "Scalar" } ], @@ -2496,7 +2628,7 @@ ] }, { - "name": "aten::arange.start", + "name": "aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "start", "type": "Scalar" }, { "name": "end", "type": "Scalar" }, @@ -2510,7 +2642,7 @@ ] }, { - "name": "aten::arange.start_out", + "name": "aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "start", "type": "Scalar" }, { "name": "end", "type": "Scalar" }, @@ -2521,7 +2653,7 @@ ] }, { - "name": "aten::arange.start_out_", + "name": "aten::arange.start_out_(Scalar start, Scalar end) -> Tensor", "inputs": [ { "name": "start", "type": "Scalar" }, { "name": "end", "type": "Scalar" } @@ -2531,7 +2663,7 @@ ] }, { - "name": "aten::arange.start_step", + "name": "aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "start", "type": "Scalar" }, { "name": "end", "type": "Scalar" }, @@ -2546,7 +2678,7 @@ ] }, { - "name": "aten::arctan", + "name": "aten::arctan(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -2555,7 +2687,7 @@ ] }, { - "name": "aten::arctan.out", + "name": "aten::arctan.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -2564,7 +2696,7 @@ ] }, { - "name": "aten::arctan_", + "name": "aten::arctan_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -2573,7 +2705,7 @@ ] }, { - "name": "aten::argmax", + "name": "aten::argmax(Tensor self, int? dim=None, bool keepdim=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64?", "default": null }, @@ -2584,7 +2716,7 @@ ] }, { - "name": "aten::argmax.out", + "name": "aten::argmax.out(Tensor self, int? dim=None, bool keepdim=False, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64?", "default": null }, @@ -2595,7 +2727,7 @@ ] }, { - "name": "aten::argmin", + "name": "aten::argmin(Tensor self, int? dim=None, bool keepdim=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64?", "default": null }, @@ -2606,7 +2738,7 @@ ] }, { - "name": "aten::argmin.out", + "name": "aten::argmin.out(Tensor self, int? dim=None, bool keepdim=False, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64?", "default": null }, @@ -2617,7 +2749,7 @@ ] }, { - "name": "aten::argsort", + "name": "aten::argsort(Tensor self, int dim=-1, bool descending=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64", "default": -1 }, @@ -2628,7 +2760,7 @@ ] }, { - "name": "aten::argsort.dimname", + "name": "aten::argsort.dimname(Tensor self, Dimname dim, bool descending=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -2639,7 +2771,7 @@ ] }, { - "name": "aten::argsort.stable", + "name": "aten::argsort.stable(Tensor self, *, bool stable, int dim=-1, bool descending=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "stable", "type": "boolean", "kwarg_only": true }, @@ -2651,7 +2783,7 @@ ] }, { - "name": "aten::argsort.stable_out", + "name": "aten::argsort.stable_out(Tensor self, *, bool stable, int dim=-1, bool descending=False, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "stable", "type": "boolean", "kwarg_only": true }, @@ -2663,7 +2795,7 @@ ] }, { - "name": "aten::argwhere", + "name": "aten::argwhere(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -2672,7 +2804,7 @@ ] }, { - "name": "aten::as_strided", + "name": "aten::as_strided(Tensor(a) self, SymInt[] size, SymInt[] stride, SymInt? storage_offset=None) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "size", "type": "SymInt[]" }, @@ -2684,7 +2816,7 @@ ] }, { - "name": "aten::as_strided_", + "name": "aten::as_strided_(Tensor(a!) self, SymInt[] size, SymInt[] stride, SymInt? storage_offset=None) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "size", "type": "SymInt[]" }, @@ -2696,7 +2828,7 @@ ] }, { - "name": "aten::as_strided_copy", + "name": "aten::as_strided_copy(Tensor self, SymInt[] size, SymInt[] stride, SymInt? storage_offset=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "size", "type": "SymInt[]" }, @@ -2708,7 +2840,7 @@ ] }, { - "name": "aten::as_strided_scatter", + "name": "aten::as_strided_scatter(Tensor self, Tensor src, SymInt[] size, SymInt[] stride, SymInt? storage_offset=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "src", "type": "Tensor" }, @@ -2721,7 +2853,7 @@ ] }, { - "name": "aten::as_tensor", + "name": "aten::as_tensor(Tensor(a) data, *, ScalarType? dtype=None, Device? device=None) -> Tensor(b|a)", "inputs": [ { "name": "data", "type": "Tensor" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -2732,7 +2864,7 @@ ] }, { - "name": "aten::as_tensor.bool", + "name": "aten::as_tensor.bool(bool t, *, ScalarType? dtype=None, Device? device=None) -> Tensor", "inputs": [ { "name": "t", "type": "boolean" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -2743,7 +2875,7 @@ ] }, { - "name": "aten::as_tensor.complex", + "name": "aten::as_tensor.complex(complex t, *, ScalarType? dtype=None, Device? device=None) -> Tensor", "inputs": [ { "name": "t", "type": "complex" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -2754,7 +2886,7 @@ ] }, { - "name": "aten::as_tensor.float", + "name": "aten::as_tensor.float(float t, *, ScalarType? dtype=None, Device? device=None) -> Tensor", "inputs": [ { "name": "t", "type": "float32" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -2765,7 +2897,7 @@ ] }, { - "name": "aten::as_tensor.int", + "name": "aten::as_tensor.int(int t, *, ScalarType? dtype=None, Device? device=None) -> Tensor", "inputs": [ { "name": "t", "type": "int64" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -2776,7 +2908,7 @@ ] }, { - "name": "aten::as_tensor.list", + "name": "aten::as_tensor.list(t[] data, *, ScalarType? dtype=None, Device? device=None) -> Tensor", "inputs": [ { "name": "data", "type": "t[]" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -2787,7 +2919,7 @@ ] }, { - "name": "aten::asin", + "name": "aten::asin(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -2796,7 +2928,7 @@ ] }, { - "name": "aten::asin.out", + "name": "aten::asin.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -2805,7 +2937,7 @@ ] }, { - "name": "aten::asinh", + "name": "aten::asinh(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -2814,7 +2946,7 @@ ] }, { - "name": "aten::asinh.out", + "name": "aten::asinh.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -2823,7 +2955,7 @@ ] }, { - "name": "aten::atan", + "name": "aten::atan(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -2832,7 +2964,7 @@ ] }, { - "name": "aten::atan.out", + "name": "aten::atan.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -2841,7 +2973,7 @@ ] }, { - "name": "aten::atan2", + "name": "aten::atan2(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -2851,7 +2983,7 @@ ] }, { - "name": "aten::atan2.out", + "name": "aten::atan2.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -2861,7 +2993,7 @@ ] }, { - "name": "aten::atan2_", + "name": "aten::atan2_(Tensor(a!) self, Tensor other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -2871,7 +3003,7 @@ ] }, { - "name": "aten::atan_", + "name": "aten::atan_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -2880,7 +3012,7 @@ ] }, { - "name": "aten::atanh", + "name": "aten::atanh(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -2889,7 +3021,7 @@ ] }, { - "name": "aten::atanh.out", + "name": "aten::atanh.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -2898,7 +3030,7 @@ ] }, { - "name": "aten::atanh_", + "name": "aten::atanh_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -2907,7 +3039,7 @@ ] }, { - "name": "aten::avg_pool1d", + "name": "aten::avg_pool1d(Tensor self, int[1] kernel_size, int[1] stride=[], int[1] padding=0, bool ceil_mode=False, bool count_include_pad=True) -> Tensor", "category": "Pool", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -2922,7 +3054,7 @@ ] }, { - "name": "aten::avg_pool2d", + "name": "aten::avg_pool2d(Tensor self, int[2] kernel_size, int[2] stride=[], int[2] padding=0, bool ceil_mode=False, bool count_include_pad=True, int? divisor_override=None) -> Tensor", "category": "Pool", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -2938,7 +3070,7 @@ ] }, { - "name": "aten::avg_pool2d.out", + "name": "aten::avg_pool2d.out(Tensor self, int[2] kernel_size, int[2] stride=[], int[2] padding=0, bool ceil_mode=False, bool count_include_pad=True, int? divisor_override=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "kernel_size", "type": "int64[2]" }, @@ -2953,7 +3085,7 @@ ] }, { - "name": "aten::avg_pool3d", + "name": "aten::avg_pool3d(Tensor self, int[3] kernel_size, int[3] stride=[], int[3] padding=0, bool ceil_mode=False, bool count_include_pad=True, int? divisor_override=None) -> Tensor", "category": "Pool", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -2969,7 +3101,7 @@ ] }, { - "name": "aten::avg_pool3d.out", + "name": "aten::avg_pool3d.out(Tensor self, int[3] kernel_size, int[3] stride=[], int[3] padding=0, bool ceil_mode=False, bool count_include_pad=True, int? divisor_override=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "kernel_size", "type": "int64[3]" }, @@ -2984,7 +3116,7 @@ ] }, { - "name": "aten::baddbmm", + "name": "aten::baddbmm(Tensor self, Tensor batch1, Tensor batch2, *, Scalar beta=1, Scalar alpha=1) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "batch1", "type": "Tensor" }, @@ -2997,7 +3129,7 @@ ] }, { - "name": "aten::baddbmm.out", + "name": "aten::baddbmm.out(Tensor self, Tensor batch1, Tensor batch2, *, Scalar beta=1, Scalar alpha=1, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "batch1", "type": "Tensor" }, @@ -3010,7 +3142,7 @@ ] }, { - "name": "aten::baddbmm_", + "name": "aten::baddbmm_(Tensor(a!) self, Tensor batch1, Tensor batch2, *, Scalar beta=1, Scalar alpha=1) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "batch1", "type": "Tensor" }, @@ -3023,7 +3155,7 @@ ] }, { - "name": "aten::batch_norm", + "name": "aten::batch_norm(Tensor input, Tensor? weight, Tensor? bias, Tensor? running_mean, Tensor? running_var, bool training, float momentum, float eps, bool cudnn_enabled) -> Tensor", "category": "Normalization", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -3041,7 +3173,7 @@ ] }, { - "name": "aten::bernoulli", + "name": "aten::bernoulli(Tensor self, *, Generator? generator=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "generator", "type": "Generator?", "default": null, "kwarg_only": true } @@ -3051,7 +3183,7 @@ ] }, { - "name": "aten::bernoulli.out", + "name": "aten::bernoulli.out(Tensor self, *, Generator? generator=None, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "generator", "type": "Generator?", "default": null, "kwarg_only": true } @@ -3061,7 +3193,7 @@ ] }, { - "name": "aten::bernoulli.p", + "name": "aten::bernoulli.p(Tensor self, float p, *, Generator? generator=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "p", "type": "float32", "default": 0.5 }, @@ -3072,7 +3204,7 @@ ] }, { - "name": "aten::bernoulli_.Tensor", + "name": "aten::bernoulli_.Tensor(Tensor(a!) self, Tensor p, *, Generator? generator=None) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "p", "type": "Tensor" }, @@ -3083,7 +3215,7 @@ ] }, { - "name": "aten::bernoulli_.float", + "name": "aten::bernoulli_.float(Tensor(a!) self, float p=0.5, *, Generator? generator=None) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "p", "type": "float32", "default": 0.5 }, @@ -3094,7 +3226,7 @@ ] }, { - "name": "aten::bilinear", + "name": "aten::bilinear(Tensor input1, Tensor input2, Tensor weight, Tensor? bias=None) -> Tensor", "inputs": [ { "name": "input1", "type": "Tensor" }, { "name": "input2", "type": "Tensor" }, @@ -3106,7 +3238,7 @@ ] }, { - "name": "aten::bin", + "name": "aten::bin(int i) -> str", "inputs": [ { "name": "i", "type": "int64" } ], @@ -3115,7 +3247,7 @@ ] }, { - "name": "aten::binary_cross_entropy", + "name": "aten::binary_cross_entropy(Tensor self, Tensor target, Tensor? weight=None, int reduction=Mean) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "target", "type": "Tensor" }, @@ -3127,7 +3259,7 @@ ] }, { - "name": "aten::binary_cross_entropy.out", + "name": "aten::binary_cross_entropy.out(Tensor self, Tensor target, Tensor? weight=None, int reduction=Mean, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "target", "type": "Tensor" }, @@ -3139,7 +3271,7 @@ ] }, { - "name": "aten::binary_cross_entropy_backward", + "name": "aten::binary_cross_entropy_backward(Tensor grad_output, Tensor self, Tensor target, Tensor? weight=None, int reduction=Mean) -> Tensor", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -3152,7 +3284,7 @@ ] }, { - "name": "aten::binary_cross_entropy_backward.grad_input", + "name": "aten::binary_cross_entropy_backward.grad_input(Tensor grad_output, Tensor self, Tensor target, Tensor? weight=None, int reduction=Mean, *, Tensor(a!) grad_input) -> Tensor(a!)", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -3165,7 +3297,7 @@ ] }, { - "name": "aten::binary_cross_entropy_with_logits", + "name": "aten::binary_cross_entropy_with_logits(Tensor self, Tensor target, Tensor? weight=None, Tensor? pos_weight=None, int reduction=Mean) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "target", "type": "Tensor" }, @@ -3178,7 +3310,7 @@ ] }, { - "name": "aten::bincount", + "name": "aten::bincount(Tensor self, Tensor? weights=None, int minlength=0) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "weights", "type": "Tensor?", "default": null }, @@ -3189,7 +3321,7 @@ ] }, { - "name": "aten::binomial", + "name": "aten::binomial(Tensor count, Tensor prob, Generator? generator=None) -> Tensor", "inputs": [ { "name": "count", "type": "Tensor" }, { "name": "prob", "type": "Tensor" }, @@ -3200,7 +3332,7 @@ ] }, { - "name": "aten::bitwise_and.Scalar", + "name": "aten::bitwise_and.Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -3210,7 +3342,7 @@ ] }, { - "name": "aten::bitwise_and.Scalar_Tensor", + "name": "aten::bitwise_and.Scalar_Tensor(Scalar self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Scalar" }, { "name": "other", "type": "Tensor" } @@ -3220,7 +3352,7 @@ ] }, { - "name": "aten::bitwise_and.Scalar_Tensor_out", + "name": "aten::bitwise_and.Scalar_Tensor_out(Scalar self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Scalar" }, { "name": "other", "type": "Tensor" } @@ -3230,7 +3362,7 @@ ] }, { - "name": "aten::bitwise_and.Scalar_out", + "name": "aten::bitwise_and.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -3240,7 +3372,7 @@ ] }, { - "name": "aten::bitwise_and.Tensor", + "name": "aten::bitwise_and.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -3250,7 +3382,7 @@ ] }, { - "name": "aten::bitwise_and.Tensor_out", + "name": "aten::bitwise_and.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -3260,7 +3392,7 @@ ] }, { - "name": "aten::bitwise_and_.Scalar", + "name": "aten::bitwise_and_.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -3270,7 +3402,7 @@ ] }, { - "name": "aten::bitwise_and_.Tensor", + "name": "aten::bitwise_and_.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -3280,7 +3412,7 @@ ] }, { - "name": "aten::bitwise_left_shift.Scalar_Tensor", + "name": "aten::bitwise_left_shift.Scalar_Tensor(Scalar self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Scalar" }, { "name": "other", "type": "Tensor" } @@ -3290,7 +3422,7 @@ ] }, { - "name": "aten::bitwise_left_shift.Scalar_Tensor_out", + "name": "aten::bitwise_left_shift.Scalar_Tensor_out(Scalar self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Scalar" }, { "name": "other", "type": "Tensor" } @@ -3300,7 +3432,7 @@ ] }, { - "name": "aten::bitwise_left_shift.Tensor", + "name": "aten::bitwise_left_shift.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -3310,7 +3442,7 @@ ] }, { - "name": "aten::bitwise_left_shift.Tensor_Scalar", + "name": "aten::bitwise_left_shift.Tensor_Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -3320,7 +3452,7 @@ ] }, { - "name": "aten::bitwise_left_shift.Tensor_Scalar_out", + "name": "aten::bitwise_left_shift.Tensor_Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -3330,7 +3462,7 @@ ] }, { - "name": "aten::bitwise_left_shift.Tensor_out", + "name": "aten::bitwise_left_shift.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -3340,7 +3472,7 @@ ] }, { - "name": "aten::bitwise_left_shift_.Tensor", + "name": "aten::bitwise_left_shift_.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -3350,7 +3482,7 @@ ] }, { - "name": "aten::bitwise_left_shift_.Tensor_Scalar", + "name": "aten::bitwise_left_shift_.Tensor_Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -3360,7 +3492,7 @@ ] }, { - "name": "aten::bitwise_not", + "name": "aten::bitwise_not(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -3369,7 +3501,7 @@ ] }, { - "name": "aten::bitwise_not.out", + "name": "aten::bitwise_not.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -3378,7 +3510,7 @@ ] }, { - "name": "aten::bitwise_not_", + "name": "aten::bitwise_not_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -3387,7 +3519,7 @@ ] }, { - "name": "aten::bitwise_or.Scalar", + "name": "aten::bitwise_or.Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -3397,7 +3529,7 @@ ] }, { - "name": "aten::bitwise_or.Scalar_Tensor", + "name": "aten::bitwise_or.Scalar_Tensor(Scalar self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Scalar" }, { "name": "other", "type": "Tensor" } @@ -3407,7 +3539,7 @@ ] }, { - "name": "aten::bitwise_or.Scalar_Tensor_out", + "name": "aten::bitwise_or.Scalar_Tensor_out(Scalar self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Scalar" }, { "name": "other", "type": "Tensor" } @@ -3417,7 +3549,7 @@ ] }, { - "name": "aten::bitwise_or.Scalar_out", + "name": "aten::bitwise_or.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -3427,7 +3559,7 @@ ] }, { - "name": "aten::bitwise_or.Tensor", + "name": "aten::bitwise_or.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -3437,7 +3569,7 @@ ] }, { - "name": "aten::bitwise_or.Tensor_out", + "name": "aten::bitwise_or.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -3447,7 +3579,7 @@ ] }, { - "name": "aten::bitwise_or_.Scalar", + "name": "aten::bitwise_or_.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -3457,7 +3589,7 @@ ] }, { - "name": "aten::bitwise_or_.Tensor", + "name": "aten::bitwise_or_.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -3467,7 +3599,7 @@ ] }, { - "name": "aten::bitwise_right_shift.Scalar_Tensor", + "name": "aten::bitwise_right_shift.Scalar_Tensor(Scalar self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Scalar" }, { "name": "other", "type": "Tensor" } @@ -3477,7 +3609,7 @@ ] }, { - "name": "aten::bitwise_right_shift.Scalar_Tensor_out", + "name": "aten::bitwise_right_shift.Scalar_Tensor_out(Scalar self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Scalar" }, { "name": "other", "type": "Tensor" } @@ -3487,7 +3619,7 @@ ] }, { - "name": "aten::bitwise_right_shift.Tensor", + "name": "aten::bitwise_right_shift.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -3497,7 +3629,7 @@ ] }, { - "name": "aten::bitwise_right_shift.Tensor_Scalar", + "name": "aten::bitwise_right_shift.Tensor_Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -3507,7 +3639,7 @@ ] }, { - "name": "aten::bitwise_right_shift.Tensor_Scalar_out", + "name": "aten::bitwise_right_shift.Tensor_Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -3517,7 +3649,7 @@ ] }, { - "name": "aten::bitwise_right_shift.Tensor_out", + "name": "aten::bitwise_right_shift.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -3527,7 +3659,7 @@ ] }, { - "name": "aten::bitwise_right_shift_.Tensor", + "name": "aten::bitwise_right_shift_.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -3537,7 +3669,7 @@ ] }, { - "name": "aten::bitwise_right_shift_.Tensor_Scalar", + "name": "aten::bitwise_right_shift_.Tensor_Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -3547,7 +3679,7 @@ ] }, { - "name": "aten::bitwise_xor.Scalar", + "name": "aten::bitwise_xor.Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -3557,7 +3689,7 @@ ] }, { - "name": "aten::bitwise_xor.Scalar_Tensor", + "name": "aten::bitwise_xor.Scalar_Tensor(Scalar self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Scalar" }, { "name": "other", "type": "Tensor" } @@ -3567,7 +3699,7 @@ ] }, { - "name": "aten::bitwise_xor.Scalar_Tensor_out", + "name": "aten::bitwise_xor.Scalar_Tensor_out(Scalar self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Scalar" }, { "name": "other", "type": "Tensor" } @@ -3577,7 +3709,7 @@ ] }, { - "name": "aten::bitwise_xor.Scalar_out", + "name": "aten::bitwise_xor.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -3587,7 +3719,7 @@ ] }, { - "name": "aten::bitwise_xor.Tensor", + "name": "aten::bitwise_xor.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -3597,7 +3729,7 @@ ] }, { - "name": "aten::bitwise_xor.Tensor_out", + "name": "aten::bitwise_xor.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -3607,7 +3739,7 @@ ] }, { - "name": "aten::bitwise_xor_.Scalar", + "name": "aten::bitwise_xor_.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -3617,7 +3749,7 @@ ] }, { - "name": "aten::bitwise_xor_.Tensor", + "name": "aten::bitwise_xor_.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -3627,7 +3759,7 @@ ] }, { - "name": "aten::block_diag", + "name": "aten::block_diag(Tensor[] tensors) -> Tensor", "inputs": [ { "name": "tensors", "type": "Tensor[]" } ], @@ -3636,7 +3768,7 @@ ] }, { - "name": "aten::bmm", + "name": "aten::bmm(Tensor self, Tensor mat2) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "mat2", "type": "Tensor" } @@ -3646,7 +3778,7 @@ ] }, { - "name": "aten::bmm.out", + "name": "aten::bmm.out(Tensor self, Tensor mat2, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "mat2", "type": "Tensor" } @@ -3656,7 +3788,7 @@ ] }, { - "name": "aten::broadcast_tensors", + "name": "aten::broadcast_tensors(Tensor[] tensors) -> Tensor[]", "inputs": [ { "name": "tensors", "type": "Tensor[]" } ], @@ -3665,7 +3797,7 @@ ] }, { - "name": "aten::broadcast_to", + "name": "aten::broadcast_to(Tensor(a) self, SymInt[] size) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "size", "type": "SymInt[]" } @@ -3675,7 +3807,7 @@ ] }, { - "name": "aten::bucketize.Scalar", + "name": "aten::bucketize.Scalar(Scalar self, Tensor boundaries, *, bool out_int32=False, bool right=False) -> Tensor", "inputs": [ { "name": "self", "type": "Scalar" }, { "name": "boundaries", "type": "Tensor" }, @@ -3687,7 +3819,7 @@ ] }, { - "name": "aten::bucketize.Tensor", + "name": "aten::bucketize.Tensor(Tensor self, Tensor boundaries, *, bool out_int32=False, bool right=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "boundaries", "type": "Tensor" }, @@ -3699,7 +3831,7 @@ ] }, { - "name": "aten::bucketize.Tensor_out", + "name": "aten::bucketize.Tensor_out(Tensor self, Tensor boundaries, *, bool out_int32=False, bool right=False, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "boundaries", "type": "Tensor" }, @@ -3711,7 +3843,7 @@ ] }, { - "name": "aten::cartesian_prod", + "name": "aten::cartesian_prod(Tensor[] tensors) -> Tensor", "inputs": [ { "name": "tensors", "type": "Tensor[]" } ], @@ -3720,7 +3852,7 @@ ] }, { - "name": "aten::cat", + "name": "aten::cat(Tensor[] tensors, int dim=0) -> Tensor", "category": "Tensor", "inputs": [ { "name": "tensors", "type": "Tensor[]" }, @@ -3731,7 +3863,7 @@ ] }, { - "name": "aten::cat.names", + "name": "aten::cat.names(Tensor[] tensors, Dimname dim) -> Tensor", "category": "Tensor", "inputs": [ { "name": "tensors", "type": "Tensor[]" }, @@ -3742,7 +3874,7 @@ ] }, { - "name": "aten::cat.names_out", + "name": "aten::cat.names_out(Tensor[] tensors, Dimname dim, *, Tensor(a!) out) -> Tensor(a!)", "category": "Tensor", "inputs": [ { "name": "tensors", "type": "Tensor[]" }, @@ -3753,7 +3885,7 @@ ] }, { - "name": "aten::cat.out", + "name": "aten::cat.out(Tensor[] tensors, int dim=0, *, Tensor(a!) out) -> Tensor(a!)", "category": "Tensor", "inputs": [ { "name": "tensors", "type": "Tensor[]" }, @@ -3764,7 +3896,7 @@ ] }, { - "name": "aten::cauchy_", + "name": "aten::cauchy_(Tensor(a!) self, float median=0, float sigma=1, *, Generator? generator=None) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "median", "type": "float32", "default": 0 }, @@ -3776,7 +3908,7 @@ ] }, { - "name": "aten::cdist", + "name": "aten::cdist(Tensor x1, Tensor x2, float p=2, int? compute_mode=None) -> Tensor", "inputs": [ { "name": "x1", "type": "Tensor" }, { "name": "x2", "type": "Tensor" }, @@ -3788,7 +3920,7 @@ ] }, { - "name": "aten::ceil", + "name": "aten::ceil(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -3797,7 +3929,7 @@ ] }, { - "name": "aten::ceil.Scalar", + "name": "aten::ceil.Scalar(Scalar a) -> Scalar", "inputs": [ { "name": "a", "type": "Scalar" } ], @@ -3806,7 +3938,7 @@ ] }, { - "name": "aten::ceil.float", + "name": "aten::ceil.float(float a) -> int", "inputs": [ { "name": "a", "type": "float32" } ], @@ -3815,7 +3947,7 @@ ] }, { - "name": "aten::ceil.int", + "name": "aten::ceil.int(int a) -> int", "inputs": [ { "name": "a", "type": "int64" } ], @@ -3824,7 +3956,7 @@ ] }, { - "name": "aten::ceil.out", + "name": "aten::ceil.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -3833,7 +3965,7 @@ ] }, { - "name": "aten::ceil_", + "name": "aten::ceil_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -3842,7 +3974,7 @@ ] }, { - "name": "aten::celu", + "name": "aten::celu(Tensor self, Scalar alpha=1.0) -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -3853,7 +3985,7 @@ ] }, { - "name": "aten::celu_", + "name": "aten::celu_(Tensor(a!) self, Scalar alpha=1.0) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "alpha", "type": "Scalar", "default": 1.0 } @@ -3863,7 +3995,7 @@ ] }, { - "name": "aten::channel_shuffle", + "name": "aten::channel_shuffle(Tensor self, SymInt groups) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "groups", "type": "SymInt" } @@ -3873,7 +4005,7 @@ ] }, { - "name": "aten::chunk", + "name": "aten::chunk(Tensor(a -> *) self, int chunks, int dim=0) -> Tensor(a)[]", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "chunks", "type": "int64" }, @@ -3884,7 +4016,7 @@ ] }, { - "name": "aten::clamp", + "name": "aten::clamp(Tensor self, Scalar? min=None, Scalar? max=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "min", "type": "Scalar?", "default": null }, @@ -3895,7 +4027,7 @@ ] }, { - "name": "aten::clamp.Tensor", + "name": "aten::clamp.Tensor(Tensor self, Tensor? min=None, Tensor? max=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "min", "type": "Tensor?", "default": null }, @@ -3906,7 +4038,7 @@ ] }, { - "name": "aten::clamp.Tensor_out", + "name": "aten::clamp.Tensor_out(Tensor self, Tensor? min=None, Tensor? max=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "min", "type": "Tensor?", "default": null }, @@ -3917,7 +4049,7 @@ ] }, { - "name": "aten::clamp.out", + "name": "aten::clamp.out(Tensor self, Scalar? min=None, Scalar? max=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "min", "type": "Scalar?", "default": null }, @@ -3928,7 +4060,7 @@ ] }, { - "name": "aten::clamp_", + "name": "aten::clamp_(Tensor(a!) self, Scalar? min=None, Scalar? max=None) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "min", "type": "Scalar?", "default": null }, @@ -3939,7 +4071,7 @@ ] }, { - "name": "aten::clamp_.Tensor", + "name": "aten::clamp_.Tensor(Tensor(a!) self, Tensor? min=None, Tensor? max=None) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "min", "type": "Tensor?", "default": null }, @@ -3950,7 +4082,7 @@ ] }, { - "name": "aten::clamp_max", + "name": "aten::clamp_max(Tensor self, Scalar max) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "max", "type": "Scalar" } @@ -3960,7 +4092,7 @@ ] }, { - "name": "aten::clamp_max.Tensor", + "name": "aten::clamp_max.Tensor(Tensor self, Tensor max) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "max", "type": "Tensor" } @@ -3970,7 +4102,7 @@ ] }, { - "name": "aten::clamp_max.Tensor_out", + "name": "aten::clamp_max.Tensor_out(Tensor self, Tensor max, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "max", "type": "Tensor" } @@ -3980,7 +4112,7 @@ ] }, { - "name": "aten::clamp_max.out", + "name": "aten::clamp_max.out(Tensor self, Scalar max, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "max", "type": "Scalar" } @@ -3990,7 +4122,7 @@ ] }, { - "name": "aten::clamp_max_", + "name": "aten::clamp_max_(Tensor(a!) self, Scalar max) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "max", "type": "Scalar" } @@ -4000,7 +4132,7 @@ ] }, { - "name": "aten::clamp_max_.Tensor", + "name": "aten::clamp_max_.Tensor(Tensor(a!) self, Tensor max) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "max", "type": "Tensor" } @@ -4010,7 +4142,7 @@ ] }, { - "name": "aten::clamp_min", + "name": "aten::clamp_min(Tensor self, Scalar min) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "min", "type": "Scalar" } @@ -4020,7 +4152,7 @@ ] }, { - "name": "aten::clamp_min.Tensor", + "name": "aten::clamp_min.Tensor(Tensor self, Tensor min) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "min", "type": "Tensor" } @@ -4030,7 +4162,7 @@ ] }, { - "name": "aten::clamp_min.Tensor_out", + "name": "aten::clamp_min.Tensor_out(Tensor self, Tensor min, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "min", "type": "Tensor" } @@ -4040,7 +4172,7 @@ ] }, { - "name": "aten::clamp_min.out", + "name": "aten::clamp_min.out(Tensor self, Scalar min, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "min", "type": "Scalar" } @@ -4050,7 +4182,7 @@ ] }, { - "name": "aten::clamp_min_", + "name": "aten::clamp_min_(Tensor(a!) self, Scalar min) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "min", "type": "Scalar" } @@ -4060,7 +4192,7 @@ ] }, { - "name": "aten::clamp_min_.Tensor", + "name": "aten::clamp_min_.Tensor(Tensor(a!) self, Tensor min) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "min", "type": "Tensor" } @@ -4081,7 +4213,7 @@ ] }, { - "name": "aten::clip", + "name": "aten::clip(Tensor self, Scalar? min=None, Scalar? max=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "min", "type": "Scalar?", "default": null }, @@ -4092,7 +4224,7 @@ ] }, { - "name": "aten::clip.Tensor", + "name": "aten::clip.Tensor(Tensor self, Tensor? min=None, Tensor? max=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "min", "type": "Tensor?", "default": null }, @@ -4103,7 +4235,7 @@ ] }, { - "name": "aten::clip.Tensor_out", + "name": "aten::clip.Tensor_out(Tensor self, Tensor? min=None, Tensor? max=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "min", "type": "Tensor?", "default": null }, @@ -4114,7 +4246,7 @@ ] }, { - "name": "aten::clip.out", + "name": "aten::clip.out(Tensor self, Scalar? min=None, Scalar? max=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "min", "type": "Scalar?", "default": null }, @@ -4125,7 +4257,7 @@ ] }, { - "name": "aten::clip_", + "name": "aten::clip_(Tensor(a!) self, Scalar? min=None, Scalar? max=None) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "min", "type": "Scalar?", "default": null }, @@ -4136,7 +4268,7 @@ ] }, { - "name": "aten::clip_.Tensor", + "name": "aten::clip_.Tensor(Tensor(a!) self, Tensor? min=None, Tensor? max=None) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "min", "type": "Tensor?", "default": null }, @@ -4147,7 +4279,7 @@ ] }, { - "name": "aten::clone", + "name": "aten::clone(Tensor self, *, MemoryFormat? memory_format=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "memory_format", "type": "MemoryFormat?", "default": null, "kwarg_only": true } @@ -4157,7 +4289,7 @@ ] }, { - "name": "aten::coalesce", + "name": "aten::coalesce(Tensor(a) self) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -4166,7 +4298,7 @@ ] }, { - "name": "aten::col2im", + "name": "aten::col2im(Tensor self, SymInt[2] output_size, int[2] kernel_size, int[2] dilation, int[2] padding, int[2] stride) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[2]" }, @@ -4180,7 +4312,7 @@ ] }, { - "name": "aten::col2im.out", + "name": "aten::col2im.out(Tensor self, SymInt[2] output_size, int[2] kernel_size, int[2] dilation, int[2] padding, int[2] stride, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[2]" }, @@ -4194,7 +4326,7 @@ ] }, { - "name": "aten::column_stack", + "name": "aten::column_stack(Tensor[] tensors) -> Tensor", "inputs": [ { "name": "tensors", "type": "Tensor[]" } ], @@ -4203,7 +4335,7 @@ ] }, { - "name": "aten::column_stack.out", + "name": "aten::column_stack.out(Tensor[] tensors, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "tensors", "type": "Tensor[]" } ], @@ -4212,7 +4344,7 @@ ] }, { - "name": "aten::complex", + "name": "aten::complex(Tensor real, Tensor imag) -> Tensor", "inputs": [ { "name": "real", "type": "Tensor" }, { "name": "imag", "type": "Tensor" } @@ -4222,7 +4354,7 @@ ] }, { - "name": "aten::complex.out", + "name": "aten::complex.out(Tensor real, Tensor imag, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "real", "type": "Tensor" }, { "name": "imag", "type": "Tensor" } @@ -4232,7 +4364,7 @@ ] }, { - "name": "aten::concat", + "name": "aten::concat(Tensor[] tensors, int dim=0) -> Tensor", "category": "Tensor", "inputs": [ { "name": "tensors", "type": "Tensor[]" }, @@ -4243,7 +4375,7 @@ ] }, { - "name": "aten::concat.names", + "name": "aten::concat.names(Tensor[] tensors, Dimname dim) -> Tensor", "category": "Tensor", "inputs": [ { "name": "tensors", "type": "Tensor[]" }, @@ -4254,7 +4386,7 @@ ] }, { - "name": "aten::concat.names_out", + "name": "aten::concat.names_out(Tensor[] tensors, Dimname dim, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "tensors", "type": "Tensor[]" }, { "name": "dim", "type": "Dimname" } @@ -4264,7 +4396,7 @@ ] }, { - "name": "aten::concat.out", + "name": "aten::concat.out(Tensor[] tensors, int dim=0, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "tensors", "type": "Tensor[]" }, { "name": "dim", "type": "int64", "default": 0 } @@ -4274,7 +4406,7 @@ ] }, { - "name": "aten::concatenate", + "name": "aten::concatenate(Tensor[] tensors, int dim=0) -> Tensor", "inputs": [ { "name": "tensors", "type": "Tensor[]" }, { "name": "dim", "type": "int64", "default": 0 } @@ -4284,7 +4416,7 @@ ] }, { - "name": "aten::concatenate.names", + "name": "aten::concatenate.names(Tensor[] tensors, Dimname dim) -> Tensor", "inputs": [ { "name": "tensors", "type": "Tensor[]" }, { "name": "dim", "type": "Dimname" } @@ -4294,7 +4426,7 @@ ] }, { - "name": "aten::concatenate.names_out", + "name": "aten::concatenate.names_out(Tensor[] tensors, Dimname dim, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "tensors", "type": "Tensor[]" }, { "name": "dim", "type": "Dimname" } @@ -4304,7 +4436,7 @@ ] }, { - "name": "aten::concatenate.out", + "name": "aten::concatenate.out(Tensor[] tensors, int dim=0, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "tensors", "type": "Tensor[]" }, { "name": "dim", "type": "int64", "default": 0 } @@ -4314,7 +4446,7 @@ ] }, { - "name": "aten::conj", + "name": "aten::conj(Tensor(a) self) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -4323,7 +4455,7 @@ ] }, { - "name": "aten::constant_pad_nd", + "name": "aten::constant_pad_nd(Tensor self, SymInt[] pad, Scalar value=0) -> Tensor", "category": "Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -4335,7 +4467,7 @@ ] }, { - "name": "aten::contiguous", + "name": "aten::contiguous(Tensor(a) self, *, MemoryFormat memory_format=contiguous_format) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "memory_format", "type": "MemoryFormat", "default": "contiguous_format", "kwarg_only": true } @@ -4345,7 +4477,7 @@ ] }, { - "name": "aten::conv1d", + "name": "aten::conv1d(Tensor input, Tensor weight, Tensor? bias=None, SymInt[1] stride=1, SymInt[1] padding=0, SymInt[1] dilation=1, SymInt groups=1) -> Tensor", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -4361,7 +4493,7 @@ ] }, { - "name": "aten::conv1d.padding", + "name": "aten::conv1d.padding(Tensor input, Tensor weight, Tensor? bias=None, SymInt[1] stride=1, str padding=\"valid\", SymInt[1] dilation=1, SymInt groups=1) -> Tensor", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -4377,7 +4509,7 @@ ] }, { - "name": "aten::conv2d", + "name": "aten::conv2d(Tensor input, Tensor weight, Tensor? bias=None, SymInt[2] stride=1, SymInt[2] padding=0, SymInt[2] dilation=1, SymInt groups=1) -> Tensor", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -4393,7 +4525,7 @@ ] }, { - "name": "aten::conv2d.padding", + "name": "aten::conv2d.padding(Tensor input, Tensor weight, Tensor? bias=None, SymInt[2] stride=1, str padding=\"valid\", SymInt[2] dilation=1, SymInt groups=1) -> Tensor", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -4409,7 +4541,7 @@ ] }, { - "name": "aten::conv3d", + "name": "aten::conv3d(Tensor input, Tensor weight, Tensor? bias=None, SymInt[3] stride=1, SymInt[3] padding=0, SymInt[3] dilation=1, SymInt groups=1) -> Tensor", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -4425,7 +4557,7 @@ ] }, { - "name": "aten::conv3d.padding", + "name": "aten::conv3d.padding(Tensor input, Tensor weight, Tensor? bias=None, SymInt[3] stride=1, str padding=\"valid\", SymInt[3] dilation=1, SymInt groups=1) -> Tensor", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -4441,7 +4573,7 @@ ] }, { - "name": "aten::conv_transpose1d", + "name": "aten::conv_transpose1d(Tensor input, Tensor weight, Tensor? bias=None, SymInt[1] stride=1, SymInt[1] padding=0, SymInt[1] output_padding=0, SymInt groups=1, SymInt[1] dilation=1) -> Tensor", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -4458,7 +4590,7 @@ ] }, { - "name": "aten::conv_transpose2d.input", + "name": "aten::conv_transpose2d.input(Tensor input, Tensor weight, Tensor? bias=None, SymInt[2] stride=1, SymInt[2] padding=0, SymInt[2] output_padding=0, SymInt groups=1, SymInt[2] dilation=1) -> Tensor", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -4475,7 +4607,7 @@ ] }, { - "name": "aten::conv_transpose3d.input", + "name": "aten::conv_transpose3d.input(Tensor input, Tensor weight, Tensor? bias=None, SymInt[3] stride=1, SymInt[3] padding=0, SymInt[3] output_padding=0, SymInt groups=1, SymInt[3] dilation=1) -> Tensor", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -4492,7 +4624,7 @@ ] }, { - "name": "aten::convolution", + "name": "aten::convolution(Tensor input, Tensor weight, Tensor? bias, SymInt[] stride, SymInt[] padding, SymInt[] dilation, bool transposed, SymInt[] output_padding, SymInt groups) -> Tensor", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -4510,7 +4642,7 @@ ] }, { - "name": "aten::convolution_backward", + "name": "aten::convolution_backward(Tensor grad_output, Tensor input, Tensor weight, SymInt[]? bias_sizes, SymInt[] stride, SymInt[] padding, SymInt[] dilation, bool transposed, SymInt[] output_padding, SymInt groups, bool[3] output_mask) -> (Tensor, Tensor, Tensor)", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "input", "type": "Tensor" }, @@ -4531,7 +4663,7 @@ ] }, { - "name": "aten::convolution_backward_overrideable", + "name": "aten::convolution_backward_overrideable(Tensor grad_output, Tensor input, Tensor weight, SymInt[] stride, SymInt[] padding, SymInt[] dilation, bool transposed, SymInt[] output_padding, SymInt groups, bool[3] output_mask) -> (Tensor grad_input, Tensor grad_weight, Tensor grad_bias)", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "input", "type": "Tensor" }, @@ -4551,7 +4683,7 @@ ] }, { - "name": "aten::convolution_overrideable", + "name": "aten::convolution_overrideable(Tensor input, Tensor weight, Tensor? bias, SymInt[] stride, SymInt[] padding, SymInt[] dilation, bool transposed, SymInt[] output_padding, SymInt groups) -> Tensor", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "weight", "type": "Tensor" }, @@ -4568,7 +4700,7 @@ ] }, { - "name": "aten::copy_", + "name": "aten::copy_(Tensor(a!) self, Tensor src, bool non_blocking=False) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "src", "type": "Tensor" }, @@ -4579,7 +4711,7 @@ ] }, { - "name": "aten::cos", + "name": "aten::cos(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -4588,7 +4720,7 @@ ] }, { - "name": "aten::cos.out", + "name": "aten::cos.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -4597,7 +4729,7 @@ ] }, { - "name": "aten::cosh", + "name": "aten::cosh(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -4606,7 +4738,7 @@ ] }, { - "name": "aten::cosh.out", + "name": "aten::cosh.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -4615,7 +4747,7 @@ ] }, { - "name": "aten::cosine_similarity", + "name": "aten::cosine_similarity(Tensor x1, Tensor x2, int dim=1, float eps=1e-08) -> Tensor", "inputs": [ { "name": "x1", "type": "Tensor" }, { "name": "x2", "type": "Tensor" }, @@ -4627,7 +4759,7 @@ ] }, { - "name": "aten::count_nonzero", + "name": "aten::count_nonzero(Tensor self, int? dim=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64?", "default": null } @@ -4637,7 +4769,7 @@ ] }, { - "name": "aten::count_nonzero.dim_IntList", + "name": "aten::count_nonzero.dim_IntList(Tensor self, int[] dim) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[]" } @@ -4647,7 +4779,7 @@ ] }, { - "name": "aten::cpu", + "name": "aten::cpu(Tensor(a) self) -> Tensor(a|b)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -4656,7 +4788,7 @@ ] }, { - "name": "aten::cross", + "name": "aten::cross(Tensor self, Tensor other, int? dim=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -4667,7 +4799,7 @@ ] }, { - "name": "aten::cross.out", + "name": "aten::cross.out(Tensor self, Tensor other, int? dim=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -4678,7 +4810,7 @@ ] }, { - "name": "aten::cross_entropy_loss", + "name": "aten::cross_entropy_loss(Tensor self, Tensor target, Tensor? weight=None, int reduction=Mean, SymInt ignore_index=-100, float label_smoothing=0.0) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "target", "type": "Tensor" }, @@ -4692,7 +4824,7 @@ ] }, { - "name": "aten::ctc_loss.IntList", + "name": "aten::ctc_loss.IntList(Tensor log_probs, Tensor targets, int[] input_lengths, int[] target_lengths, int blank=0, int reduction=Mean, bool zero_infinity=False) -> Tensor", "inputs": [ { "name": "log_probs", "type": "Tensor" }, { "name": "targets", "type": "Tensor" }, @@ -4707,7 +4839,7 @@ ] }, { - "name": "aten::ctc_loss.Tensor", + "name": "aten::ctc_loss.Tensor(Tensor log_probs, Tensor targets, Tensor input_lengths, Tensor target_lengths, int blank=0, int reduction=Mean, bool zero_infinity=False) -> Tensor", "inputs": [ { "name": "log_probs", "type": "Tensor" }, { "name": "targets", "type": "Tensor" }, @@ -4722,7 +4854,7 @@ ] }, { - "name": "aten::cudnn_convolution_add_relu", + "name": "aten::cudnn_convolution_add_relu(Tensor self, Tensor weight, Tensor z, Scalar? alpha, Tensor? bias, SymInt[] stride, SymInt[] padding, SymInt[] dilation, SymInt groups) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "weight", "type": "Tensor" }, @@ -4739,7 +4871,7 @@ ] }, { - "name": "aten::cudnn_convolution_relu", + "name": "aten::cudnn_convolution_relu(Tensor self, Tensor weight, Tensor? bias, SymInt[] stride, SymInt[] padding, SymInt[] dilation, SymInt groups) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "weight", "type": "Tensor" }, @@ -4754,7 +4886,7 @@ ] }, { - "name": "aten::cummax", + "name": "aten::cummax(Tensor self, int dim) -> (Tensor values, Tensor indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" } @@ -4765,7 +4897,7 @@ ] }, { - "name": "aten::cummax.dimname", + "name": "aten::cummax.dimname(Tensor self, Dimname dim) -> (Tensor values, Tensor indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" } @@ -4776,7 +4908,7 @@ ] }, { - "name": "aten::cummax.dimname_out", + "name": "aten::cummax.dimname_out(Tensor self, Dimname dim, *, Tensor(a!) values, Tensor(b!) indices) -> (Tensor(a!) values, Tensor(b!) indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" } @@ -4787,7 +4919,7 @@ ] }, { - "name": "aten::cummax.out", + "name": "aten::cummax.out(Tensor self, int dim, *, Tensor(a!) values, Tensor(b!) indices) -> (Tensor(a!) values, Tensor(b!) indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" } @@ -4798,7 +4930,7 @@ ] }, { - "name": "aten::cummaxmin_backward", + "name": "aten::cummaxmin_backward(Tensor grad, Tensor input, Tensor indices, int dim) -> Tensor", "inputs": [ { "name": "grad", "type": "Tensor" }, { "name": "input", "type": "Tensor" }, @@ -4810,7 +4942,7 @@ ] }, { - "name": "aten::cumsum", + "name": "aten::cumsum(Tensor self, int dim, *, ScalarType? dtype=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -4821,7 +4953,7 @@ ] }, { - "name": "aten::cumsum.dimname", + "name": "aten::cumsum.dimname(Tensor self, Dimname dim, *, ScalarType? dtype=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -4832,7 +4964,7 @@ ] }, { - "name": "aten::cumsum.dimname_out", + "name": "aten::cumsum.dimname_out(Tensor self, Dimname dim, *, ScalarType? dtype=None, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -4843,7 +4975,7 @@ ] }, { - "name": "aten::cumsum.out", + "name": "aten::cumsum.out(Tensor self, int dim, *, ScalarType? dtype=None, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -4854,7 +4986,7 @@ ] }, { - "name": "aten::cumsum_", + "name": "aten::cumsum_(Tensor(a!) self, int dim, *, ScalarType? dtype=None) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -4865,7 +4997,7 @@ ] }, { - "name": "aten::cumsum_.dimname", + "name": "aten::cumsum_.dimname(Tensor(a!) self, Dimname dim, *, ScalarType? dtype=None) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -4876,7 +5008,7 @@ ] }, { - "name": "aten::dequantize.any", + "name": "aten::dequantize.any(Any tensors) -> Any", "category": "Quantization", "inputs": [ { "name": "tensors", "type": "Any" } @@ -4886,7 +5018,7 @@ ] }, { - "name": "aten::dequantize.list", + "name": "aten::dequantize.list(Tensor[] qtensors) -> Tensor[]", "category": "Quantization", "inputs": [ { "name": "qtensors", "type": "Tensor[]" } @@ -4896,7 +5028,7 @@ ] }, { - "name": "aten::dequantize.self", + "name": "aten::dequantize.self(Tensor self) -> Tensor", "category": "Quantization", "inputs": [ { "name": "self", "type": "Tensor" } @@ -4906,7 +5038,7 @@ ] }, { - "name": "aten::dequantize.tensor", + "name": "aten::dequantize.tensor(Tensor qtensor) -> Tensor", "category": "Quantization", "inputs": [ { "name": "qtensor", "type": "Tensor" } @@ -4916,7 +5048,7 @@ ] }, { - "name": "aten::dequantize.tensors", + "name": "aten::dequantize.tensors(Tensor[] tensors) -> Tensor[]", "category": "Quantization", "inputs": [ { "name": "tensors", "type": "Tensor[]" } @@ -4926,7 +5058,7 @@ ] }, { - "name": "aten::detach", + "name": "aten::detach(Tensor(a) self) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -4935,7 +5067,7 @@ ] }, { - "name": "aten::detach_", + "name": "aten::detach_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -4944,7 +5076,7 @@ ] }, { - "name": "aten::detach_copy", + "name": "aten::detach_copy(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -4953,7 +5085,7 @@ ] }, { - "name": "aten::diag", + "name": "aten::diag(Tensor self, int diagonal=0) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "diagonal", "type": "int64", "default": 0 } @@ -4963,7 +5095,7 @@ ] }, { - "name": "aten::diag.out", + "name": "aten::diag.out(Tensor self, int diagonal=0, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "diagonal", "type": "int64", "default": 0 } @@ -4973,7 +5105,7 @@ ] }, { - "name": "aten::diag_embed", + "name": "aten::diag_embed(Tensor self, int offset=0, int dim1=-2, int dim2=-1) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "offset", "type": "int64", "default": 0 }, @@ -4985,7 +5117,7 @@ ] }, { - "name": "aten::diagflat", + "name": "aten::diagflat(Tensor self, int offset=0) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "offset", "type": "int64", "default": 0 } @@ -4995,7 +5127,7 @@ ] }, { - "name": "aten::diagonal", + "name": "aten::diagonal(Tensor(a) self, int offset=0, int dim1=0, int dim2=1) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "offset", "type": "int64", "default": 0 }, @@ -5007,7 +5139,7 @@ ] }, { - "name": "aten::diagonal.Dimname", + "name": "aten::diagonal.Dimname(Tensor(a) self, *, Dimname outdim, Dimname dim1, Dimname dim2, int offset=0) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "outdim", "type": "Dimname", "kwarg_only": true }, @@ -5020,7 +5152,7 @@ ] }, { - "name": "aten::diagonal_backward", + "name": "aten::diagonal_backward(Tensor grad_output, SymInt[] input_sizes, int offset, int dim1, int dim2) -> Tensor", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "input_sizes", "type": "SymInt[]" }, @@ -5033,7 +5165,7 @@ ] }, { - "name": "aten::diagonal_copy", + "name": "aten::diagonal_copy(Tensor self, int offset=0, int dim1=0, int dim2=1) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "offset", "type": "int64", "default": 0 }, @@ -5045,7 +5177,7 @@ ] }, { - "name": "aten::diagonal_scatter", + "name": "aten::diagonal_scatter(Tensor self, Tensor src, int offset=0, int dim1=0, int dim2=1) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "src", "type": "Tensor" }, @@ -5058,14 +5190,14 @@ ] }, { - "name": "aten::dict", + "name": "aten::dict() -> Dict(str, Tensor)", "inputs": [], "outputs": [ { "type": "Dict(string, Tensor)" } ] }, { - "name": "aten::dict.Dict_Tensor", + "name": "aten::dict.Dict_Tensor(Dict(Tensor, t)(a) self) -> Dict(Tensor, t)", "inputs": [ { "name": "self", "type": "Dict(Tensor, t)" } ], @@ -5074,7 +5206,7 @@ ] }, { - "name": "aten::dict.Dict_bool", + "name": "aten::dict.Dict_bool(Dict(bool, t)(a) self) -> Dict(bool, t)", "inputs": [ { "name": "self", "type": "Dict(boolean, t)" } ], @@ -5083,7 +5215,7 @@ ] }, { - "name": "aten::dict.Dict_complex", + "name": "aten::dict.Dict_complex(Dict(complex, t)(a) self) -> Dict(complex, t)", "inputs": [ { "name": "self", "type": "Dict(complex, t)" } ], @@ -5092,7 +5224,7 @@ ] }, { - "name": "aten::dict.Dict_float", + "name": "aten::dict.Dict_float(Dict(float, t)(a) self) -> Dict(float, t)", "inputs": [ { "name": "self", "type": "Dict(float32, t)" } ], @@ -5101,7 +5233,7 @@ ] }, { - "name": "aten::dict.Dict_int", + "name": "aten::dict.Dict_int(Dict(int, t)(a) self) -> Dict(int, t)", "inputs": [ { "name": "self", "type": "Dict(int64, t)" } ], @@ -5110,7 +5242,7 @@ ] }, { - "name": "aten::dict.Dict_str", + "name": "aten::dict.Dict_str(Dict(str, t)(a) self) -> Dict(str, t)", "inputs": [ { "name": "self", "type": "Dict(string, t)" } ], @@ -5119,7 +5251,7 @@ ] }, { - "name": "aten::dict.Tensor", + "name": "aten::dict.Tensor((Tensor, tVal)[] inputs) -> Dict(Tensor, tVal)", "inputs": [ { "name": "inputs", "type": "(Tensor, tVal)[]" } ], @@ -5128,7 +5260,7 @@ ] }, { - "name": "aten::dict.bool", + "name": "aten::dict.bool((bool, tVal)[] inputs) -> Dict(bool, tVal)", "inputs": [ { "name": "inputs", "type": "(boolean, tVal)[]" } ], @@ -5137,7 +5269,7 @@ ] }, { - "name": "aten::dict.complex", + "name": "aten::dict.complex((complex, tVal)[] inputs) -> Dict(complex, tVal)", "inputs": [ { "name": "inputs", "type": "(complex, tVal)[]" } ], @@ -5146,7 +5278,7 @@ ] }, { - "name": "aten::dict.float", + "name": "aten::dict.float((float, tVal)[] inputs) -> Dict(float, tVal)", "inputs": [ { "name": "inputs", "type": "(float32, tVal)[]" } ], @@ -5155,7 +5287,7 @@ ] }, { - "name": "aten::dict.int", + "name": "aten::dict.int((int, tVal)[] inputs) -> Dict(int, tVal)", "inputs": [ { "name": "inputs", "type": "(int64, tVal)[]" } ], @@ -5164,7 +5296,7 @@ ] }, { - "name": "aten::dict.str", + "name": "aten::dict.str((str, tVal)[] inputs) -> Dict(str, tVal)", "inputs": [ { "name": "inputs", "type": "(string, tVal)[]" } ], @@ -5173,7 +5305,7 @@ ] }, { - "name": "aten::diff", + "name": "aten::diff(Tensor self, int n=1, int dim=-1, Tensor? prepend=None, Tensor? append=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "n", "type": "int64", "default": 1 }, @@ -5186,7 +5318,7 @@ ] }, { - "name": "aten::diff.out", + "name": "aten::diff.out(Tensor self, int n=1, int dim=-1, Tensor? prepend=None, Tensor? append=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "n", "type": "int64", "default": 1 }, @@ -5199,7 +5331,7 @@ ] }, { - "name": "aten::dim", + "name": "aten::dim(Tensor self) -> int", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -5208,7 +5340,7 @@ ] }, { - "name": "aten::dist", + "name": "aten::dist(Tensor self, Tensor other, Scalar p=2) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -5219,7 +5351,7 @@ ] }, { - "name": "aten::div", + "name": "aten::div(Scalar a, Scalar b) -> float", "inputs": [ { "name": "a", "type": "Scalar" }, { "name": "b", "type": "Scalar" } @@ -5229,7 +5361,7 @@ ] }, { - "name": "aten::div.Scalar", + "name": "aten::div.Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -5239,7 +5371,7 @@ ] }, { - "name": "aten::div.Scalar_mode", + "name": "aten::div.Scalar_mode(Tensor self, Scalar other, *, str? rounding_mode) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" }, @@ -5250,7 +5382,7 @@ ] }, { - "name": "aten::div.Scalar_mode_out", + "name": "aten::div.Scalar_mode_out(Tensor self, Scalar other, *, str? rounding_mode, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" }, @@ -5261,7 +5393,7 @@ ] }, { - "name": "aten::div.Scalar_out", + "name": "aten::div.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -5271,7 +5403,7 @@ ] }, { - "name": "aten::div.Tensor", + "name": "aten::div.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -5281,7 +5413,7 @@ ] }, { - "name": "aten::div.Tensor_mode", + "name": "aten::div.Tensor_mode(Tensor self, Tensor other, *, str? rounding_mode) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -5292,7 +5424,7 @@ ] }, { - "name": "aten::div.complex", + "name": "aten::div.complex(complex a, complex b) -> complex", "inputs": [ { "name": "a", "type": "complex" }, { "name": "b", "type": "complex" } @@ -5302,7 +5434,7 @@ ] }, { - "name": "aten::div.float", + "name": "aten::div.float(float a, float b) -> float", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "float32" } @@ -5312,7 +5444,7 @@ ] }, { - "name": "aten::div.int", + "name": "aten::div.int(int a, int b) -> float", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "int64" } @@ -5322,7 +5454,7 @@ ] }, { - "name": "aten::div.out", + "name": "aten::div.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -5332,7 +5464,7 @@ ] }, { - "name": "aten::div.out_mode", + "name": "aten::div.out_mode(Tensor self, Tensor other, *, str? rounding_mode, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -5343,7 +5475,7 @@ ] }, { - "name": "aten::div_.Scalar", + "name": "aten::div_.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -5353,7 +5485,7 @@ ] }, { - "name": "aten::div_.Scalar_mode", + "name": "aten::div_.Scalar_mode(Tensor(a!) self, Scalar other, *, str? rounding_mode) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" }, @@ -5364,7 +5496,7 @@ ] }, { - "name": "aten::div_.Tensor", + "name": "aten::div_.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -5374,7 +5506,7 @@ ] }, { - "name": "aten::div_.Tensor_mode", + "name": "aten::div_.Tensor_mode(Tensor(a!) self, Tensor other, *, str? rounding_mode) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -5385,7 +5517,7 @@ ] }, { - "name": "aten::divide.Scalar", + "name": "aten::divide.Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -5395,7 +5527,7 @@ ] }, { - "name": "aten::divide.Scalar_mode", + "name": "aten::divide.Scalar_mode(Tensor self, Scalar other, *, str? rounding_mode) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" }, @@ -5406,7 +5538,7 @@ ] }, { - "name": "aten::divide.Tensor", + "name": "aten::divide.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -5416,7 +5548,7 @@ ] }, { - "name": "aten::divide.Tensor_mode", + "name": "aten::divide.Tensor_mode(Tensor self, Tensor other, *, str? rounding_mode) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -5427,7 +5559,7 @@ ] }, { - "name": "aten::divide.out", + "name": "aten::divide.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -5437,7 +5569,7 @@ ] }, { - "name": "aten::divide.out_mode", + "name": "aten::divide.out_mode(Tensor self, Tensor other, *, str? rounding_mode, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -5448,7 +5580,7 @@ ] }, { - "name": "aten::divide_.Scalar", + "name": "aten::divide_.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -5458,7 +5590,7 @@ ] }, { - "name": "aten::divide_.Scalar_mode", + "name": "aten::divide_.Scalar_mode(Tensor(a!) self, Scalar other, *, str? rounding_mode) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" }, @@ -5469,7 +5601,7 @@ ] }, { - "name": "aten::divide_.Tensor", + "name": "aten::divide_.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -5479,7 +5611,7 @@ ] }, { - "name": "aten::divide_.Tensor_mode", + "name": "aten::divide_.Tensor_mode(Tensor(a!) self, Tensor other, *, str? rounding_mode) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -5490,7 +5622,7 @@ ] }, { - "name": "aten::divmod.float", + "name": "aten::divmod.float(float x, float y) -> (float, float)", "inputs": [ { "name": "x", "type": "float32" }, { "name": "y", "type": "float32" } @@ -5501,7 +5633,7 @@ ] }, { - "name": "aten::divmod.int", + "name": "aten::divmod.int(int x, int y) -> (int, int)", "inputs": [ { "name": "x", "type": "int64" }, { "name": "y", "type": "int64" } @@ -5512,7 +5644,7 @@ ] }, { - "name": "aten::dot", + "name": "aten::dot(Tensor self, Tensor tensor) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "tensor", "type": "Tensor" } @@ -5522,7 +5654,7 @@ ] }, { - "name": "aten::dot.out", + "name": "aten::dot.out(Tensor self, Tensor tensor, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "tensor", "type": "Tensor" } @@ -5532,7 +5664,7 @@ ] }, { - "name": "aten::dropout", + "name": "aten::dropout(Tensor input, float p, bool train) -> Tensor", "category": "Dropout", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -5544,7 +5676,7 @@ ] }, { - "name": "aten::dropout_", + "name": "aten::dropout_(Tensor(a!) self, float p, bool train) -> Tensor(a!)", "category": "Dropout", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -5556,7 +5688,7 @@ ] }, { - "name": "aten::einsum", + "name": "aten::einsum(str equation, Tensor[] tensors, *, int[]? path=None) -> Tensor", "inputs": [ { "name": "equation", "type": "string" }, { "name": "tensors", "type": "Tensor[]" }, @@ -5567,7 +5699,7 @@ ] }, { - "name": "aten::einsum.sublist", + "name": "aten::einsum.sublist(Tensor a, ...) -> Tensor", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -5576,7 +5708,7 @@ ] }, { - "name": "aten::elu", + "name": "aten::elu(Tensor self, Scalar alpha=1, Scalar scale=1, Scalar input_scale=1) -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -5589,7 +5721,7 @@ ] }, { - "name": "aten::elu.out", + "name": "aten::elu.out(Tensor self, Scalar alpha=1, Scalar scale=1, Scalar input_scale=1, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "alpha", "type": "Scalar", "default": 1 }, @@ -5601,7 +5733,7 @@ ] }, { - "name": "aten::elu_", + "name": "aten::elu_(Tensor(a!) self, Scalar alpha=1, Scalar scale=1, Scalar input_scale=1) -> Tensor(a!)", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -5614,7 +5746,7 @@ ] }, { - "name": "aten::embedding", + "name": "aten::embedding(Tensor weight, Tensor indices, SymInt padding_idx=-1, bool scale_grad_by_freq=False, bool sparse=False) -> Tensor", "category": "Transform", "inputs": [ { "name": "weight", "type": "Tensor" }, @@ -5628,7 +5760,7 @@ ] }, { - "name": "aten::embedding_bag", + "name": "aten::embedding_bag(Tensor weight, Tensor indices, Tensor offsets, bool scale_grad_by_freq=False, int mode=0, bool sparse=False, Tensor? per_sample_weights=None, bool include_last_offset=False) -> (Tensor, Tensor, Tensor, Tensor)", "category": "Transform", "inputs": [ { "name": "weight", "type": "Tensor" }, @@ -5648,7 +5780,7 @@ ] }, { - "name": "aten::embedding_bag.padding_idx", + "name": "aten::embedding_bag.padding_idx(Tensor weight, Tensor indices, Tensor offsets, bool scale_grad_by_freq, int mode, bool sparse, Tensor? per_sample_weights, bool include_last_offset, int? padding_idx) -> (Tensor, Tensor, Tensor, Tensor)", "inputs": [ { "name": "weight", "type": "Tensor" }, { "name": "indices", "type": "Tensor" }, @@ -5668,7 +5800,7 @@ ] }, { - "name": "aten::embedding_renorm_", + "name": "aten::embedding_renorm_(Tensor(a!) self, Tensor indices, float max_norm, float norm_type) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "indices", "type": "Tensor" }, @@ -5680,7 +5812,7 @@ ] }, { - "name": "aten::empty.memory_format", + "name": "aten::empty.memory_format(SymInt[] size, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None, MemoryFormat? memory_format=None) -> Tensor", "inputs": [ { "name": "size", "type": "SymInt[]" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -5694,7 +5826,7 @@ ] }, { - "name": "aten::empty.names", + "name": "aten::empty.names(int[] size, *, Dimname[]? names, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None, MemoryFormat? memory_format=None) -> Tensor", "inputs": [ { "name": "size", "type": "int64[]" }, { "name": "names", "type": "Dimname[]?", "default": null, "kwarg_only": true }, @@ -5709,7 +5841,7 @@ ] }, { - "name": "aten::empty.out", + "name": "aten::empty.out(SymInt[] size, *, MemoryFormat? memory_format=None, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "size", "type": "SymInt[]" }, { "name": "memory_format", "type": "MemoryFormat?", "default": null, "kwarg_only": true } @@ -5719,7 +5851,7 @@ ] }, { - "name": "aten::empty_like", + "name": "aten::empty_like(Tensor self, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None, MemoryFormat? memory_format=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -5733,7 +5865,7 @@ ] }, { - "name": "aten::eq", + "name": "aten::eq(Scalar a, Scalar b) -> bool", "inputs": [ { "name": "a", "type": "Scalar" }, { "name": "b", "type": "Scalar" } @@ -5743,7 +5875,7 @@ ] }, { - "name": "aten::eq.Scalar", + "name": "aten::eq.Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -5753,7 +5885,7 @@ ] }, { - "name": "aten::eq.Scalar_out", + "name": "aten::eq.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -5763,7 +5895,7 @@ ] }, { - "name": "aten::eq.Tensor", + "name": "aten::eq.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -5773,7 +5905,7 @@ ] }, { - "name": "aten::eq.Tensor_list", + "name": "aten::eq.Tensor_list(Tensor[] a, Tensor[] b) -> bool", "inputs": [ { "name": "a", "type": "Tensor[]" }, { "name": "b", "type": "Tensor[]" } @@ -5783,7 +5915,7 @@ ] }, { - "name": "aten::eq.Tensor_out", + "name": "aten::eq.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -5793,7 +5925,7 @@ ] }, { - "name": "aten::eq.bool", + "name": "aten::eq.bool(bool a, bool b) -> bool", "inputs": [ { "name": "a", "type": "boolean" }, { "name": "b", "type": "boolean" } @@ -5803,7 +5935,7 @@ ] }, { - "name": "aten::eq.bool_list", + "name": "aten::eq.bool_list(bool[] a, bool[] b) -> bool", "inputs": [ { "name": "a", "type": "boolean[]" }, { "name": "b", "type": "boolean[]" } @@ -5813,7 +5945,7 @@ ] }, { - "name": "aten::eq.complex", + "name": "aten::eq.complex(complex a, complex b) -> bool", "inputs": [ { "name": "a", "type": "complex" }, { "name": "b", "type": "complex" } @@ -5823,7 +5955,7 @@ ] }, { - "name": "aten::eq.complex_float", + "name": "aten::eq.complex_float(complex a, float b) -> bool", "inputs": [ { "name": "a", "type": "complex" }, { "name": "b", "type": "float32" } @@ -5833,7 +5965,7 @@ ] }, { - "name": "aten::eq.device", + "name": "aten::eq.device(Device a, Device b) -> bool", "inputs": [ { "name": "a", "type": "Device" }, { "name": "b", "type": "Device" } @@ -5843,7 +5975,7 @@ ] }, { - "name": "aten::eq.enum", + "name": "aten::eq.enum(AnyEnumType a, AnyEnumType b) -> bool", "inputs": [ { "name": "a", "type": "AnyEnumType" }, { "name": "b", "type": "AnyEnumType" } @@ -5853,7 +5985,7 @@ ] }, { - "name": "aten::eq.float", + "name": "aten::eq.float(float a, float b) -> bool", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "float32" } @@ -5863,7 +5995,7 @@ ] }, { - "name": "aten::eq.float_complex", + "name": "aten::eq.float_complex(float a, complex b) -> bool", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "complex" } @@ -5873,7 +6005,7 @@ ] }, { - "name": "aten::eq.float_int", + "name": "aten::eq.float_int(float a, int b) -> bool", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "int64" } @@ -5883,7 +6015,7 @@ ] }, { - "name": "aten::eq.float_list", + "name": "aten::eq.float_list(float[] a, float[] b) -> bool", "inputs": [ { "name": "a", "type": "float32[]" }, { "name": "b", "type": "float32[]" } @@ -5893,7 +6025,7 @@ ] }, { - "name": "aten::eq.int", + "name": "aten::eq.int(int a, int b) -> bool", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "int64" } @@ -5903,7 +6035,7 @@ ] }, { - "name": "aten::eq.int_float", + "name": "aten::eq.int_float(int a, float b) -> bool", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "float32" } @@ -5913,7 +6045,7 @@ ] }, { - "name": "aten::eq.int_list", + "name": "aten::eq.int_list(int[] a, int[] b) -> bool", "inputs": [ { "name": "a", "type": "int64[]" }, { "name": "b", "type": "int64[]" } @@ -5923,7 +6055,7 @@ ] }, { - "name": "aten::eq.str", + "name": "aten::eq.str(str a, str b) -> bool", "inputs": [ { "name": "a", "type": "string" }, { "name": "b", "type": "string" } @@ -5933,7 +6065,7 @@ ] }, { - "name": "aten::eq.str_list", + "name": "aten::eq.str_list(str[] a, str[] b) -> bool", "inputs": [ { "name": "a", "type": "string[]" }, { "name": "b", "type": "string[]" } @@ -5943,7 +6075,7 @@ ] }, { - "name": "aten::equal", + "name": "aten::equal(Tensor self, Tensor other) -> bool", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -5953,7 +6085,7 @@ ] }, { - "name": "aten::erf", + "name": "aten::erf(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -5962,7 +6094,7 @@ ] }, { - "name": "aten::erf.out", + "name": "aten::erf.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -5971,7 +6103,7 @@ ] }, { - "name": "aten::erfc", + "name": "aten::erfc(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -5980,7 +6112,7 @@ ] }, { - "name": "aten::erfc.out", + "name": "aten::erfc.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -5989,7 +6121,7 @@ ] }, { - "name": "aten::exp", + "name": "aten::exp(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -5998,7 +6130,7 @@ ] }, { - "name": "aten::exp.out", + "name": "aten::exp.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -6007,7 +6139,7 @@ ] }, { - "name": "aten::exp_", + "name": "aten::exp_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -6016,7 +6148,7 @@ ] }, { - "name": "aten::expand", + "name": "aten::expand(Tensor(a) self, SymInt[] size, *, bool implicit=False) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "size", "type": "SymInt[]" }, @@ -6027,7 +6159,7 @@ ] }, { - "name": "aten::expand_as", + "name": "aten::expand_as(Tensor(a) self, Tensor other) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -6037,7 +6169,7 @@ ] }, { - "name": "aten::expm1", + "name": "aten::expm1(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -6046,7 +6178,7 @@ ] }, { - "name": "aten::expm1.out", + "name": "aten::expm1.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -6055,7 +6187,7 @@ ] }, { - "name": "aten::expm1_", + "name": "aten::expm1_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -6064,7 +6196,7 @@ ] }, { - "name": "aten::exponential_", + "name": "aten::exponential_(Tensor(a!) self, float lambd=1, *, Generator? generator=None) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "lambd", "type": "float32", "default": 1 }, @@ -6075,7 +6207,7 @@ ] }, { - "name": "aten::extend.t", + "name": "aten::extend.t(t[](a!) self, t[] other) -> ()", "inputs": [ { "name": "self", "type": "t[]" }, { "name": "other", "type": "t[]" } @@ -6083,7 +6215,7 @@ "outputs": [] }, { - "name": "aten::eye", + "name": "aten::eye(SymInt n, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "n", "type": "SymInt" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -6096,7 +6228,7 @@ ] }, { - "name": "aten::eye.m", + "name": "aten::eye.m(SymInt n, SymInt m, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "n", "type": "SymInt" }, { "name": "m", "type": "SymInt" }, @@ -6110,7 +6242,7 @@ ] }, { - "name": "aten::eye.m_out", + "name": "aten::eye.m_out(SymInt n, SymInt m, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "n", "type": "SymInt" }, { "name": "m", "type": "SymInt" } @@ -6120,7 +6252,7 @@ ] }, { - "name": "aten::eye.out", + "name": "aten::eye.out(SymInt n, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "n", "type": "SymInt" } ], @@ -6129,7 +6261,7 @@ ] }, { - "name": "aten::fake_quantize_per_channel_affine", + "name": "aten::fake_quantize_per_channel_affine(Tensor self, Tensor scale, Tensor zero_point, int axis, int quant_min, int quant_max) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "scale", "type": "Tensor" }, @@ -6143,7 +6275,7 @@ ] }, { - "name": "aten::fake_quantize_per_tensor_affine", + "name": "aten::fake_quantize_per_tensor_affine(Tensor self, float scale, int zero_point, int quant_min, int quant_max) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "scale", "type": "float32" }, @@ -6156,7 +6288,7 @@ ] }, { - "name": "aten::fake_quantize_per_tensor_affine.tensor_qparams", + "name": "aten::fake_quantize_per_tensor_affine.tensor_qparams(Tensor self, Tensor scale, Tensor zero_point, int quant_min, int quant_max) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "scale", "type": "Tensor" }, @@ -6169,7 +6301,7 @@ ] }, { - "name": "aten::fake_quantize_per_tensor_affine_cachemask", + "name": "aten::fake_quantize_per_tensor_affine_cachemask(Tensor self, float scale, int zero_point, int quant_min, int quant_max) -> (Tensor output, Tensor mask)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "scale", "type": "float32" }, @@ -6183,7 +6315,7 @@ ] }, { - "name": "aten::fake_quantize_per_tensor_affine_cachemask_backward", + "name": "aten::fake_quantize_per_tensor_affine_cachemask_backward(Tensor grad, Tensor mask) -> Tensor", "inputs": [ { "name": "grad", "type": "Tensor" }, { "name": "mask", "type": "Tensor" } @@ -6193,7 +6325,7 @@ ] }, { - "name": "aten::feature_alpha_dropout", + "name": "aten::feature_alpha_dropout(Tensor input, float p, bool train) -> Tensor", "category": "Dropout", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -6205,7 +6337,7 @@ ] }, { - "name": "aten::feature_alpha_dropout_", + "name": "aten::feature_alpha_dropout_(Tensor(a!) self, float p, bool train) -> Tensor(a!)", "category": "Dropout", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -6217,7 +6349,7 @@ ] }, { - "name": "aten::feature_dropout", + "name": "aten::feature_dropout(Tensor input, float p, bool train) -> Tensor", "category": "Dropout", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -6229,7 +6361,7 @@ ] }, { - "name": "aten::feature_dropout_", + "name": "aten::feature_dropout_(Tensor(a!) self, float p, bool train) -> Tensor(a!)", "category": "Dropout", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -6241,7 +6373,7 @@ ] }, { - "name": "aten::fft", + "name": "aten::fft(Tensor self, int signal_ndim, bool normalized=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "signal_ndim", "type": "int64" }, @@ -6252,7 +6384,7 @@ ] }, { - "name": "aten::fft_fft", + "name": "aten::fft_fft(Tensor self, SymInt? n=None, int dim=-1, str? norm=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "n", "type": "SymInt?", "default": null }, @@ -6264,7 +6396,7 @@ ] }, { - "name": "aten::fft_fft.out", + "name": "aten::fft_fft.out(Tensor self, SymInt? n=None, int dim=-1, str? norm=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "n", "type": "SymInt?", "default": null }, @@ -6276,7 +6408,7 @@ ] }, { - "name": "aten::fft_fft2", + "name": "aten::fft_fft2(Tensor self, SymInt[1]? s=None, int[1] dim=[-2,-1], str? norm=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6288,7 +6420,7 @@ ] }, { - "name": "aten::fft_fft2.out", + "name": "aten::fft_fft2.out(Tensor self, SymInt[1]? s=None, int[1] dim=[-2,-1], str? norm=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6300,7 +6432,7 @@ ] }, { - "name": "aten::fft_fftn", + "name": "aten::fft_fftn(Tensor self, SymInt[1]? s=None, int[1]? dim=None, str? norm=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6312,7 +6444,7 @@ ] }, { - "name": "aten::fft_fftn.out", + "name": "aten::fft_fftn.out(Tensor self, SymInt[1]? s=None, int[1]? dim=None, str? norm=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6324,7 +6456,7 @@ ] }, { - "name": "aten::fft_fftshift", + "name": "aten::fft_fftshift(Tensor self, int[1]? dim=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]?", "default": null } @@ -6334,7 +6466,7 @@ ] }, { - "name": "aten::fft_hfft2", + "name": "aten::fft_hfft2(Tensor self, SymInt[1]? s=None, int[1] dim=[-2,-1], str? norm=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6346,7 +6478,7 @@ ] }, { - "name": "aten::fft_hfft2.out", + "name": "aten::fft_hfft2.out(Tensor self, SymInt[1]? s=None, int[1] dim=[-2,-1], str? norm=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6358,7 +6490,7 @@ ] }, { - "name": "aten::fft_hfftn", + "name": "aten::fft_hfftn(Tensor self, SymInt[1]? s=None, int[1]? dim=None, str? norm=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6370,7 +6502,7 @@ ] }, { - "name": "aten::fft_hfftn.out", + "name": "aten::fft_hfftn.out(Tensor self, SymInt[1]? s=None, int[1]? dim=None, str? norm=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6382,7 +6514,7 @@ ] }, { - "name": "aten::fft_ifft", + "name": "aten::fft_ifft(Tensor self, SymInt? n=None, int dim=-1, str? norm=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "n", "type": "SymInt?", "default": null }, @@ -6394,7 +6526,7 @@ ] }, { - "name": "aten::fft_ifft.out", + "name": "aten::fft_ifft.out(Tensor self, SymInt? n=None, int dim=-1, str? norm=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "n", "type": "SymInt?", "default": null }, @@ -6406,7 +6538,7 @@ ] }, { - "name": "aten::fft_ifft2", + "name": "aten::fft_ifft2(Tensor self, SymInt[1]? s=None, int[1] dim=[-2,-1], str? norm=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6418,7 +6550,7 @@ ] }, { - "name": "aten::fft_ifft2.out", + "name": "aten::fft_ifft2.out(Tensor self, SymInt[1]? s=None, int[1] dim=[-2,-1], str? norm=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6430,7 +6562,7 @@ ] }, { - "name": "aten::fft_ifftn", + "name": "aten::fft_ifftn(Tensor self, SymInt[1]? s=None, int[1]? dim=None, str? norm=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6442,7 +6574,7 @@ ] }, { - "name": "aten::fft_ifftn.out", + "name": "aten::fft_ifftn.out(Tensor self, SymInt[1]? s=None, int[1]? dim=None, str? norm=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6454,7 +6586,7 @@ ] }, { - "name": "aten::fft_ifftshift", + "name": "aten::fft_ifftshift(Tensor self, int[1]? dim=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]?", "default": null } @@ -6464,7 +6596,7 @@ ] }, { - "name": "aten::fft_ihfft2", + "name": "aten::fft_ihfft2(Tensor self, SymInt[1]? s=None, int[1] dim=[-2,-1], str? norm=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6476,7 +6608,7 @@ ] }, { - "name": "aten::fft_ihfft2.out", + "name": "aten::fft_ihfft2.out(Tensor self, SymInt[1]? s=None, int[1] dim=[-2,-1], str? norm=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6488,7 +6620,7 @@ ] }, { - "name": "aten::fft_ihfftn", + "name": "aten::fft_ihfftn(Tensor self, SymInt[1]? s=None, int[1]? dim=None, str? norm=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6500,7 +6632,7 @@ ] }, { - "name": "aten::fft_ihfftn.out", + "name": "aten::fft_ihfftn.out(Tensor self, SymInt[1]? s=None, int[1]? dim=None, str? norm=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6512,7 +6644,7 @@ ] }, { - "name": "aten::fft_irfft", + "name": "aten::fft_irfft(Tensor self, SymInt? n=None, int dim=-1, str? norm=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "n", "type": "SymInt?", "default": null }, @@ -6524,7 +6656,7 @@ ] }, { - "name": "aten::fft_irfft.out", + "name": "aten::fft_irfft.out(Tensor self, SymInt? n=None, int dim=-1, str? norm=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "n", "type": "SymInt?", "default": null }, @@ -6536,7 +6668,7 @@ ] }, { - "name": "aten::fft_irfft2", + "name": "aten::fft_irfft2(Tensor self, SymInt[1]? s=None, int[1] dim=[-2,-1], str? norm=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6548,7 +6680,7 @@ ] }, { - "name": "aten::fft_irfft2.out", + "name": "aten::fft_irfft2.out(Tensor self, SymInt[1]? s=None, int[1] dim=[-2,-1], str? norm=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6560,7 +6692,7 @@ ] }, { - "name": "aten::fft_irfftn", + "name": "aten::fft_irfftn(Tensor self, SymInt[1]? s=None, int[1]? dim=None, str? norm=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6572,7 +6704,7 @@ ] }, { - "name": "aten::fft_irfftn.out", + "name": "aten::fft_irfftn.out(Tensor self, SymInt[1]? s=None, int[1]? dim=None, str? norm=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6584,7 +6716,7 @@ ] }, { - "name": "aten::fft_rfft", + "name": "aten::fft_rfft(Tensor self, SymInt? n=None, int dim=-1, str? norm=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "n", "type": "SymInt?", "default": null }, @@ -6596,7 +6728,7 @@ ] }, { - "name": "aten::fft_rfft.out", + "name": "aten::fft_rfft.out(Tensor self, SymInt? n=None, int dim=-1, str? norm=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "n", "type": "SymInt?", "default": null }, @@ -6608,7 +6740,7 @@ ] }, { - "name": "aten::fft_rfft2", + "name": "aten::fft_rfft2(Tensor self, SymInt[1]? s=None, int[1] dim=[-2,-1], str? norm=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6620,7 +6752,7 @@ ] }, { - "name": "aten::fft_rfft2.out", + "name": "aten::fft_rfft2.out(Tensor self, SymInt[1]? s=None, int[1] dim=[-2,-1], str? norm=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6632,7 +6764,7 @@ ] }, { - "name": "aten::fft_rfftn", + "name": "aten::fft_rfftn(Tensor self, SymInt[1]? s=None, int[1]? dim=None, str? norm=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6644,7 +6776,7 @@ ] }, { - "name": "aten::fft_rfftn.out", + "name": "aten::fft_rfftn.out(Tensor self, SymInt[1]? s=None, int[1]? dim=None, str? norm=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "s", "type": "SymInt[1]?", "default": null }, @@ -6656,7 +6788,7 @@ ] }, { - "name": "aten::fill_.Scalar", + "name": "aten::fill_.Scalar(Tensor(a!) self, Scalar value) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "value", "type": "Scalar" } @@ -6666,7 +6798,7 @@ ] }, { - "name": "aten::fill_.Tensor", + "name": "aten::fill_.Tensor(Tensor(a!) self, Tensor value) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "value", "type": "Tensor" } @@ -6676,7 +6808,7 @@ ] }, { - "name": "aten::find", + "name": "aten::find(str self, str substr, int start=0, int end=-1) -> int", "inputs": [ { "name": "self", "type": "string" }, { "name": "substr", "type": "string" }, @@ -6688,7 +6820,7 @@ ] }, { - "name": "aten::flatten.DimnameList", + "name": "aten::flatten.DimnameList(Tensor(a) self, Dimname[] dims, Dimname out_dim) -> Tensor(a)", "category": "Shape", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -6700,7 +6832,7 @@ ] }, { - "name": "aten::flatten.named_out_dim", + "name": "aten::flatten.named_out_dim(Tensor(a) self, int start_dim, int end_dim, Dimname out_dim) -> Tensor(a)", "category": "Shape", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -6713,7 +6845,7 @@ ] }, { - "name": "aten::flatten.using_ints", + "name": "aten::flatten.using_ints(Tensor(a) self, int start_dim=0, int end_dim=-1) -> Tensor(a)", "category": "Shape", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -6725,7 +6857,7 @@ ] }, { - "name": "aten::flatten.using_names", + "name": "aten::flatten.using_names(Tensor(a) self, Dimname start_dim, Dimname end_dim, Dimname out_dim) -> Tensor(a)", "category": "Shape", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -6738,7 +6870,7 @@ ] }, { - "name": "aten::flip", + "name": "aten::flip(Tensor self, int[] dims) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dims", "type": "int64[]" } @@ -6748,7 +6880,7 @@ ] }, { - "name": "aten::floor", + "name": "aten::floor(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -6757,7 +6889,7 @@ ] }, { - "name": "aten::floor.Scalar", + "name": "aten::floor.Scalar(Scalar a) -> Scalar", "inputs": [ { "name": "a", "type": "Scalar" } ], @@ -6766,16 +6898,16 @@ ] }, { - "name": "aten::floor.float", + "name": "aten::floor.float(float a) -> int", "inputs": [ { "name": "a", "type": "float32" } ], "outputs": [ - { "type": "float32" } + { "type": "int64" } ] }, { - "name": "aten::floor.int", + "name": "aten::floor.int(int a) -> int", "inputs": [ { "name": "a", "type": "int64" } ], @@ -6784,7 +6916,7 @@ ] }, { - "name": "aten::floor.out", + "name": "aten::floor.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -6793,7 +6925,7 @@ ] }, { - "name": "aten::floor_", + "name": "aten::floor_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -6802,7 +6934,7 @@ ] }, { - "name": "aten::floor_divide", + "name": "aten::floor_divide(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -6812,7 +6944,7 @@ ] }, { - "name": "aten::floor_divide.Scalar", + "name": "aten::floor_divide.Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -6822,7 +6954,17 @@ ] }, { - "name": "aten::floor_divide.out", + "name": "aten::floor_divide.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)", + "inputs": [ + { "name": "self", "type": "Tensor" }, + { "name": "other", "type": "Scalar" } + ], + "outputs": [ + { "type": "Tensor" } + ] + }, + { + "name": "aten::floor_divide.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -6832,7 +6974,7 @@ ] }, { - "name": "aten::floor_divide_.Scalar", + "name": "aten::floor_divide_.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -6842,7 +6984,7 @@ ] }, { - "name": "aten::floor_divide_.Tensor", + "name": "aten::floor_divide_.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -6852,7 +6994,7 @@ ] }, { - "name": "aten::floordiv.Scalar", + "name": "aten::floordiv(Scalar a, Scalar b) -> Scalar", "inputs": [ { "name": "a", "type": "Scalar" }, { "name": "b", "type": "Scalar" } @@ -6862,7 +7004,7 @@ ] }, { - "name": "aten::floordiv.float", + "name": "aten::floordiv.float(float a, float b) -> float", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "float32" } @@ -6872,7 +7014,7 @@ ] }, { - "name": "aten::floordiv.float_int", + "name": "aten::floordiv.float_int(float a, int b) -> float", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "int64" } @@ -6882,7 +7024,7 @@ ] }, { - "name": "aten::floordiv.int", + "name": "aten::floordiv.int(int a, int b) -> int", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "int64" } @@ -6892,7 +7034,7 @@ ] }, { - "name": "aten::floordiv.int_float", + "name": "aten::floordiv.int_float(int a, float b) -> float", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "float32" } @@ -6902,7 +7044,17 @@ ] }, { - "name": "aten::fmod.Scalar", + "name": "aten::fmod(Scalar a, Scalar b) -> float", + "inputs": [ + { "name": "a", "type": "Scalar" }, + { "name": "b", "type": "Scalar" } + ], + "outputs": [ + { "type": "float32" } + ] + }, + { + "name": "aten::fmod.Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -6912,7 +7064,7 @@ ] }, { - "name": "aten::fmod.Scalar_out", + "name": "aten::fmod.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -6922,7 +7074,7 @@ ] }, { - "name": "aten::fmod.Tensor", + "name": "aten::fmod.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -6932,7 +7084,7 @@ ] }, { - "name": "aten::fmod.Tensor_out", + "name": "aten::fmod.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -6942,7 +7094,47 @@ ] }, { - "name": "aten::format", + "name": "aten::fmod.float(float a, float b) -> float", + "inputs": [ + { "name": "a", "type": "float32" }, + { "name": "b", "type": "float32" } + ], + "outputs": [ + { "type": "float32" } + ] + }, + { + "name": "aten::fmod.float_int(float a, int b) -> float", + "inputs": [ + { "name": "a", "type": "float32" }, + { "name": "b", "type": "int64" } + ], + "outputs": [ + { "type": "float32" } + ] + }, + { + "name": "aten::fmod.int(int a, int b) -> float", + "inputs": [ + { "name": "a", "type": "int64" }, + { "name": "b", "type": "int64" } + ], + "outputs": [ + { "type": "float32" } + ] + }, + { + "name": "aten::fmod.int_float(int a, float b) -> float", + "inputs": [ + { "name": "a", "type": "int64" }, + { "name": "b", "type": "float32" } + ], + "outputs": [ + { "type": "float32" } + ] + }, + { + "name": "aten::format(str self, ...) -> str", "is_vararg": true, "inputs": [ { "name": "self", "type": "string" } @@ -6952,7 +7144,7 @@ ] }, { - "name": "aten::frobenius_norm.dim", + "name": "aten::frobenius_norm.dim(Tensor self, int[1] dim, bool keepdim=False) -> Tensor", "category": "Normalization", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -6964,7 +7156,7 @@ ] }, { - "name": "aten::frobenius_norm.out", + "name": "aten::frobenius_norm.out(Tensor self, int[1] dim, bool keepdim=False, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]" }, @@ -6975,7 +7167,7 @@ ] }, { - "name": "aten::full", + "name": "aten::full(SymInt[] size, Scalar fill_value, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "size", "type": "SymInt[]" }, { "name": "fill_value", "type": "Scalar" }, @@ -6989,7 +7181,7 @@ ] }, { - "name": "aten::full.names", + "name": "aten::full.names(int[] size, Scalar fill_value, *, Dimname[]? names, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "size", "type": "int64[]" }, { "name": "fill_value", "type": "Scalar" }, @@ -7004,7 +7196,7 @@ ] }, { - "name": "aten::full.out", + "name": "aten::full.out(SymInt[] size, Scalar fill_value, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "size", "type": "SymInt[]" }, { "name": "fill_value", "type": "Scalar" } @@ -7014,7 +7206,7 @@ ] }, { - "name": "aten::full_like", + "name": "aten::full_like(Tensor self, Scalar fill_value, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None, MemoryFormat? memory_format=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "fill_value", "type": "Scalar" }, @@ -7029,7 +7221,7 @@ ] }, { - "name": "aten::fused_moving_avg_obs_fake_quant", + "name": "aten::fused_moving_avg_obs_fake_quant(Tensor self, Tensor observer_on, Tensor fake_quant_on, Tensor(a!) running_min, Tensor(b!) running_max, Tensor(c!) scale, Tensor(d!) zero_point, float averaging_const, int quant_min, int quant_max, int ch_axis, bool per_row_fake_quant=False, bool symmetric_quant=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "observer_on", "type": "Tensor" }, @@ -7050,7 +7242,7 @@ ] }, { - "name": "aten::gather", + "name": "aten::gather(Tensor self, int dim, Tensor index, *, bool sparse_grad=False) -> Tensor", "category": "Transform", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -7063,7 +7255,7 @@ ] }, { - "name": "aten::gather.dimname", + "name": "aten::gather.dimname(Tensor self, Dimname dim, Tensor index, *, bool sparse_grad=False) -> Tensor", "category": "Transform", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -7076,7 +7268,7 @@ ] }, { - "name": "aten::gather.dimname_out", + "name": "aten::gather.dimname_out(Tensor self, Dimname dim, Tensor index, *, bool sparse_grad=False, Tensor(a!) out) -> Tensor(a!)", "category": "Transform", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -7089,7 +7281,7 @@ ] }, { - "name": "aten::gather.out", + "name": "aten::gather.out(Tensor self, int dim, Tensor index, *, bool sparse_grad=False, Tensor(a!) out) -> Tensor(a!)", "category": "Transform", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -7102,7 +7294,7 @@ ] }, { - "name": "aten::gather_backward", + "name": "aten::gather_backward(Tensor grad, Tensor self, int dim, Tensor index, bool sparse_grad) -> Tensor", "inputs": [ { "name": "grad", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -7115,7 +7307,7 @@ ] }, { - "name": "aten::gcd", + "name": "aten::gcd(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -7125,7 +7317,7 @@ ] }, { - "name": "aten::gcd.out", + "name": "aten::gcd.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -7135,7 +7327,7 @@ ] }, { - "name": "aten::gcd_", + "name": "aten::gcd_(Tensor(a!) self, Tensor other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -7145,7 +7337,7 @@ ] }, { - "name": "aten::ge.Scalar", + "name": "aten::ge.Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -7155,7 +7347,7 @@ ] }, { - "name": "aten::ge.Scalar_out", + "name": "aten::ge.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -7165,7 +7357,7 @@ ] }, { - "name": "aten::ge.Tensor", + "name": "aten::ge.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -7175,7 +7367,7 @@ ] }, { - "name": "aten::ge.Tensor_out", + "name": "aten::ge.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -7185,7 +7377,7 @@ ] }, { - "name": "aten::ge_.Scalar", + "name": "aten::ge_.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -7195,7 +7387,7 @@ ] }, { - "name": "aten::ge_.Tensor", + "name": "aten::ge_.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -7205,7 +7397,7 @@ ] }, { - "name": "aten::gelu", + "name": "aten::gelu(Tensor self, *, str approximate='none') -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -7216,7 +7408,7 @@ ] }, { - "name": "aten::gelu.out", + "name": "aten::gelu.out(Tensor self, *, str approximate='none', Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "approximate", "type": "string", "default": "none", "kwarg_only": true } @@ -7226,7 +7418,7 @@ ] }, { - "name": "aten::gelu_", + "name": "aten::gelu_(Tensor(a!) self, *, str approximate='none') -> Tensor(a!)", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -7237,7 +7429,7 @@ ] }, { - "name": "aten::gelu_backward", + "name": "aten::gelu_backward(Tensor grad_output, Tensor self, *, str approximate='none') -> Tensor", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -7248,7 +7440,7 @@ ] }, { - "name": "aten::gelu_backward.grad_input", + "name": "aten::gelu_backward.grad_input(Tensor grad_output, Tensor self, *, str approximate='none', Tensor(a!) grad_input) -> Tensor(a!)", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -7259,7 +7451,7 @@ ] }, { - "name": "aten::geometric_", + "name": "aten::geometric_(Tensor(a!) self, float p, *, Generator? generator=None) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "p", "type": "float32" }, @@ -7270,7 +7462,7 @@ ] }, { - "name": "aten::geqrf", + "name": "aten::geqrf(Tensor self) -> (Tensor a, Tensor tau)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -7280,7 +7472,7 @@ ] }, { - "name": "aten::geqrf.a", + "name": "aten::geqrf.a(Tensor self, *, Tensor(a!) a, Tensor(b!) tau) -> (Tensor(a!) a, Tensor(b!) tau)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -7290,7 +7482,7 @@ ] }, { - "name": "aten::ger", + "name": "aten::ger(Tensor self, Tensor vec2) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "vec2", "type": "Tensor" } @@ -7300,7 +7492,7 @@ ] }, { - "name": "aten::ger.out", + "name": "aten::ger.out(Tensor self, Tensor vec2, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "vec2", "type": "Tensor" } @@ -7310,7 +7502,7 @@ ] }, { - "name": "aten::get.Tensor", + "name": "aten::get.Tensor(Dict(Tensor, t) self, Tensor key) -> t(*)?", "inputs": [ { "name": "self", "type": "Dict(Tensor, t)" }, { "name": "key", "type": "Tensor" } @@ -7320,7 +7512,7 @@ ] }, { - "name": "aten::get.bool", + "name": "aten::get.bool(Dict(bool, t) self, bool key) -> t(*)?", "inputs": [ { "name": "self", "type": "Dict(boolean, t)" }, { "name": "key", "type": "boolean" } @@ -7330,7 +7522,7 @@ ] }, { - "name": "aten::get.complex", + "name": "aten::get.complex(Dict(complex, t) self, complex key) -> t(*)?", "inputs": [ { "name": "self", "type": "Dict(complex, t)" }, { "name": "key", "type": "complex" } @@ -7340,7 +7532,7 @@ ] }, { - "name": "aten::get.default_Tensor", + "name": "aten::get.default_Tensor(Dict(Tensor, t) self, Tensor key, t default_value) -> t(*)", "inputs": [ { "name": "self", "type": "Dict(Tensor, t)" }, { "name": "key", "type": "Tensor" }, @@ -7351,7 +7543,7 @@ ] }, { - "name": "aten::get.default_bool", + "name": "aten::get.default_bool(Dict(bool, t) self, bool key, t default_value) -> t(*)", "inputs": [ { "name": "self", "type": "Dict(boolean, t)" }, { "name": "key", "type": "boolean" }, @@ -7362,7 +7554,7 @@ ] }, { - "name": "aten::get.default_complex", + "name": "aten::get.default_complex(Dict(complex, t) self, complex key, t default_value) -> t(*)", "inputs": [ { "name": "self", "type": "Dict(complex, t)" }, { "name": "key", "type": "complex" }, @@ -7373,7 +7565,7 @@ ] }, { - "name": "aten::get.default_float", + "name": "aten::get.default_float(Dict(float, t) self, float key, t default_value) -> t(*)", "inputs": [ { "name": "self", "type": "Dict(float32, t)" }, { "name": "key", "type": "float32" }, @@ -7384,7 +7576,7 @@ ] }, { - "name": "aten::get.default_int", + "name": "aten::get.default_int(Dict(int, t) self, int key, t default_value) -> t(*)", "inputs": [ { "name": "self", "type": "Dict(int64, t)" }, { "name": "key", "type": "int64" }, @@ -7395,7 +7587,7 @@ ] }, { - "name": "aten::get.default_str", + "name": "aten::get.default_str(Dict(str, t) self, str key, t default_value) -> t(*)", "inputs": [ { "name": "self", "type": "Dict(string, t)" }, { "name": "key", "type": "string" }, @@ -7406,7 +7598,7 @@ ] }, { - "name": "aten::get.float", + "name": "aten::get.float(Dict(float, t) self, float key) -> t(*)?", "inputs": [ { "name": "self", "type": "Dict(float32, t)" }, { "name": "key", "type": "float32" } @@ -7416,7 +7608,7 @@ ] }, { - "name": "aten::get.int", + "name": "aten::get.int(Dict(int, t) self, int key) -> t(*)?", "inputs": [ { "name": "self", "type": "Dict(int64, t)" }, { "name": "key", "type": "int64" } @@ -7426,7 +7618,7 @@ ] }, { - "name": "aten::get.str", + "name": "aten::get.str(Dict(str, t) self, str key) -> t(*)?", "inputs": [ { "name": "self", "type": "Dict(string, t)" }, { "name": "key", "type": "string" } @@ -7436,7 +7628,7 @@ ] }, { - "name": "aten::get_device", + "name": "aten::get_device(Tensor self) -> int", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -7445,7 +7637,7 @@ ] }, { - "name": "aten::glu", + "name": "aten::glu(Tensor self, int dim=-1) -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -7456,7 +7648,7 @@ ] }, { - "name": "aten::glu.out", + "name": "aten::glu.out(Tensor self, int dim=-1, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64", "default": -1 } @@ -7466,7 +7658,7 @@ ] }, { - "name": "aten::greater_equal.Scalar", + "name": "aten::greater_equal.Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -7476,7 +7668,7 @@ ] }, { - "name": "aten::greater_equal.Scalar_out", + "name": "aten::greater_equal.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -7486,7 +7678,7 @@ ] }, { - "name": "aten::greater_equal.Tensor", + "name": "aten::greater_equal.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -7496,7 +7688,7 @@ ] }, { - "name": "aten::greater_equal.Tensor_out", + "name": "aten::greater_equal.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -7506,7 +7698,7 @@ ] }, { - "name": "aten::greater_equal_.Scalar", + "name": "aten::greater_equal_.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -7516,7 +7708,7 @@ ] }, { - "name": "aten::greater_equal_.Tensor", + "name": "aten::greater_equal_.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -7526,7 +7718,7 @@ ] }, { - "name": "aten::grid_sampler", + "name": "aten::grid_sampler(Tensor input, Tensor grid, int interpolation_mode, int padding_mode, bool align_corners) -> Tensor", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "grid", "type": "Tensor" }, @@ -7539,7 +7731,7 @@ ] }, { - "name": "aten::grid_sampler.legacy", + "name": "aten::grid_sampler.legacy(Tensor input, Tensor grid, int interpolation_mode, int padding_mode) -> Tensor", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "grid", "type": "Tensor" }, @@ -7551,7 +7743,7 @@ ] }, { - "name": "aten::group_norm", + "name": "aten::group_norm(Tensor input, int num_groups, Tensor? weight=None, Tensor? bias=None, float eps=1e-05, bool cudnn_enabled=True) -> Tensor", "category": "Normalization", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -7566,7 +7758,7 @@ ] }, { - "name": "aten::gru.data", + "name": "aten::gru.data(Tensor data, Tensor batch_sizes, Tensor hx, Tensor[] params, bool has_biases, int num_layers, float dropout, bool train, bool bidirectional) -> (Tensor, Tensor)", "category": "Layer", "inputs": [ { "name": "data", "type": "Tensor" }, @@ -7585,7 +7777,7 @@ ] }, { - "name": "aten::gru.input", + "name": "aten::gru.input(Tensor input, Tensor hx, Tensor[] params, bool has_biases, int num_layers, float dropout, bool train, bool bidirectional, bool batch_first) -> (Tensor, Tensor)", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -7604,7 +7796,7 @@ ] }, { - "name": "aten::gt", + "name": "aten::gt(Scalar a, Scalar b) -> bool", "inputs": [ { "name": "a", "type": "Scalar" }, { "name": "b", "type": "Scalar" } @@ -7614,7 +7806,7 @@ ] }, { - "name": "aten::gt.Scalar", + "name": "aten::gt.Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -7624,7 +7816,7 @@ ] }, { - "name": "aten::gt.Scalar_out", + "name": "aten::gt.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -7634,7 +7826,7 @@ ] }, { - "name": "aten::gt.Tensor", + "name": "aten::gt.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -7644,7 +7836,7 @@ ] }, { - "name": "aten::gt.Tensor_out", + "name": "aten::gt.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -7654,7 +7846,7 @@ ] }, { - "name": "aten::gt.float", + "name": "aten::gt.float(float a, float b) -> bool", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "float32" } @@ -7664,7 +7856,7 @@ ] }, { - "name": "aten::gt.float_int", + "name": "aten::gt.float_int(float a, int b) -> bool", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "int64" } @@ -7674,7 +7866,7 @@ ] }, { - "name": "aten::gt.int", + "name": "aten::gt.int(int a, int b) -> bool", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "int64" } @@ -7684,7 +7876,7 @@ ] }, { - "name": "aten::gt.int_float", + "name": "aten::gt.int_float(int a, float b) -> bool", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "float32" } @@ -7694,7 +7886,7 @@ ] }, { - "name": "aten::gt.str", + "name": "aten::gt.str(str a, str b) -> bool", "inputs": [ { "name": "a", "type": "string" }, { "name": "b", "type": "string" } @@ -7704,7 +7896,7 @@ ] }, { - "name": "aten::hamming_window", + "name": "aten::hamming_window(int window_length, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "window_length", "type": "int64" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -7717,7 +7909,7 @@ ] }, { - "name": "aten::hamming_window.periodic", + "name": "aten::hamming_window.periodic(int window_length, bool periodic, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "window_length", "type": "int64" }, { "name": "periodic", "type": "boolean" }, @@ -7731,7 +7923,7 @@ ] }, { - "name": "aten::hamming_window.periodic_alpha", + "name": "aten::hamming_window.periodic_alpha(int window_length, bool periodic, float alpha, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "window_length", "type": "int64" }, { "name": "periodic", "type": "boolean" }, @@ -7746,7 +7938,7 @@ ] }, { - "name": "aten::hamming_window.periodic_alpha_beta", + "name": "aten::hamming_window.periodic_alpha_beta(int window_length, bool periodic, float alpha, float beta, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "window_length", "type": "int64" }, { "name": "periodic", "type": "boolean" }, @@ -7762,7 +7954,7 @@ ] }, { - "name": "aten::hann_window", + "name": "aten::hann_window(int window_length, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "window_length", "type": "int64" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -7775,7 +7967,7 @@ ] }, { - "name": "aten::hann_window.periodic", + "name": "aten::hann_window.periodic(int window_length, bool periodic, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "window_length", "type": "int64" }, { "name": "periodic", "type": "boolean" }, @@ -7789,7 +7981,7 @@ ] }, { - "name": "aten::hardshrink", + "name": "aten::hardshrink(Tensor self, Scalar lambd=0.5) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "lambd", "type": "Scalar", "default": 0.5 } @@ -7799,7 +7991,7 @@ ] }, { - "name": "aten::hardshrink.out", + "name": "aten::hardshrink.out(Tensor self, Scalar lambd=0.5, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "lambd", "type": "Scalar", "default": 0.5 } @@ -7809,7 +8001,7 @@ ] }, { - "name": "aten::hardsigmoid", + "name": "aten::hardsigmoid(Tensor self) -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" } @@ -7819,7 +8011,7 @@ ] }, { - "name": "aten::hardsigmoid.out", + "name": "aten::hardsigmoid.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -7828,7 +8020,7 @@ ] }, { - "name": "aten::hardsigmoid_", + "name": "aten::hardsigmoid_(Tensor(a!) self) -> Tensor(a!)", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" } @@ -7838,7 +8030,7 @@ ] }, { - "name": "aten::hardswish", + "name": "aten::hardswish(Tensor self) -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" } @@ -7848,7 +8040,7 @@ ] }, { - "name": "aten::hardswish.out", + "name": "aten::hardswish.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -7857,7 +8049,7 @@ ] }, { - "name": "aten::hardswish_", + "name": "aten::hardswish_(Tensor(a!) self) -> Tensor(a!)", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" } @@ -7867,7 +8059,7 @@ ] }, { - "name": "aten::hardswish_backward", + "name": "aten::hardswish_backward(Tensor grad_output, Tensor self) -> Tensor", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "self", "type": "Tensor" } @@ -7877,7 +8069,7 @@ ] }, { - "name": "aten::hardtanh", + "name": "aten::hardtanh(Tensor self, Scalar min_val=-1, Scalar max_val=1) -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -7889,7 +8081,7 @@ ] }, { - "name": "aten::hardtanh.out", + "name": "aten::hardtanh.out(Tensor self, Scalar min_val=-1, Scalar max_val=1, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "min_val", "type": "Scalar", "default": -1 }, @@ -7900,7 +8092,7 @@ ] }, { - "name": "aten::hardtanh_", + "name": "aten::hardtanh_(Tensor(a!) self, Scalar min_val=-1, Scalar max_val=1) -> Tensor(a!)", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -7912,7 +8104,7 @@ ] }, { - "name": "aten::hardtanh_backward", + "name": "aten::hardtanh_backward(Tensor grad_output, Tensor self, Scalar min_val, Scalar max_val) -> Tensor", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -7924,7 +8116,7 @@ ] }, { - "name": "aten::hardtanh_backward.grad_input", + "name": "aten::hardtanh_backward.grad_input(Tensor grad_output, Tensor self, Scalar min_val, Scalar max_val, *, Tensor(a!) grad_input) -> Tensor(a!)", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -7936,7 +8128,7 @@ ] }, { - "name": "aten::histc", + "name": "aten::histc(Tensor self, int bins=100, Scalar min=0, Scalar max=0) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "bins", "type": "int64", "default": 100 }, @@ -7948,7 +8140,7 @@ ] }, { - "name": "aten::histc.out", + "name": "aten::histc.out(Tensor self, int bins=100, Scalar min=0, Scalar max=0, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "bins", "type": "int64", "default": 100 }, @@ -7960,7 +8152,7 @@ ] }, { - "name": "aten::hstack", + "name": "aten::hstack(Tensor[] tensors) -> Tensor", "inputs": [ { "name": "tensors", "type": "Tensor[]" } ], @@ -7969,7 +8161,7 @@ ] }, { - "name": "aten::hstack.out", + "name": "aten::hstack.out(Tensor[] tensors, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "tensors", "type": "Tensor[]" } ], @@ -7978,7 +8170,7 @@ ] }, { - "name": "aten::huber_loss", + "name": "aten::huber_loss(Tensor self, Tensor target, int reduction=Mean, float delta=1.0) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "target", "type": "Tensor" }, @@ -7990,7 +8182,7 @@ ] }, { - "name": "aten::huber_loss.out", + "name": "aten::huber_loss.out(Tensor self, Tensor target, int reduction=Mean, float delta=1.0, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "target", "type": "Tensor" }, @@ -8002,7 +8194,7 @@ ] }, { - "name": "aten::huber_loss_backward", + "name": "aten::huber_loss_backward(Tensor grad_output, Tensor self, Tensor target, int reduction, float delta) -> Tensor", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -8015,7 +8207,7 @@ ] }, { - "name": "aten::huber_loss_backward.out", + "name": "aten::huber_loss_backward.out(Tensor grad_output, Tensor self, Tensor target, int reduction, float delta, *, Tensor(a!) grad_input) -> Tensor(a!)", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -8028,7 +8220,7 @@ ] }, { - "name": "aten::im2col", + "name": "aten::im2col(Tensor self, int[2] kernel_size, int[2] dilation, int[2] padding, int[2] stride) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "kernel_size", "type": "int64[2]" }, @@ -8041,7 +8233,7 @@ ] }, { - "name": "aten::im2col.out", + "name": "aten::im2col.out(Tensor self, int[2] kernel_size, int[2] dilation, int[2] padding, int[2] stride, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "kernel_size", "type": "int64[2]" }, @@ -8054,7 +8246,7 @@ ] }, { - "name": "aten::imag", + "name": "aten::imag(Tensor(a) self) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -8063,7 +8255,7 @@ ] }, { - "name": "aten::index.Tensor", + "name": "aten::index.Tensor(Tensor self, Tensor?[] indices) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "indices", "type": "Tensor?[]" } @@ -8073,7 +8265,7 @@ ] }, { - "name": "aten::index.Tensor_hacked_twin", + "name": "aten::index.Tensor_hacked_twin(Tensor self, Tensor[] indices) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "indices", "type": "Tensor[]" } @@ -8083,7 +8275,7 @@ ] }, { - "name": "aten::index.Tensor_out", + "name": "aten::index.Tensor_out(Tensor self, Tensor?[] indices, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "indices", "type": "Tensor?[]" } @@ -8093,7 +8285,7 @@ ] }, { - "name": "aten::index.str", + "name": "aten::index.str(str self, str substr, int start=0, int end=-1) -> int", "inputs": [ { "name": "self", "type": "string" }, { "name": "substr", "type": "string" }, @@ -8105,7 +8297,7 @@ ] }, { - "name": "aten::index_add", + "name": "aten::index_add(Tensor self, int dim, Tensor index, Tensor source, *, Scalar alpha=1) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -8118,7 +8310,7 @@ ] }, { - "name": "aten::index_add.dimname", + "name": "aten::index_add.dimname(Tensor self, Dimname dim, Tensor index, Tensor source, *, Scalar alpha=1) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -8131,7 +8323,7 @@ ] }, { - "name": "aten::index_add.out", + "name": "aten::index_add.out(Tensor self, int dim, Tensor index, Tensor source, *, Scalar alpha=1, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -8144,7 +8336,7 @@ ] }, { - "name": "aten::index_add_", + "name": "aten::index_add_(Tensor(a!) self, int dim, Tensor index, Tensor source, *, Scalar alpha=1) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -8157,7 +8349,7 @@ ] }, { - "name": "aten::index_copy", + "name": "aten::index_copy(Tensor self, int dim, Tensor index, Tensor source) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -8169,7 +8361,7 @@ ] }, { - "name": "aten::index_copy.dimname", + "name": "aten::index_copy.dimname(Tensor self, Dimname dim, Tensor index, Tensor source) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -8181,7 +8373,7 @@ ] }, { - "name": "aten::index_copy.out", + "name": "aten::index_copy.out(Tensor self, int dim, Tensor index, Tensor source, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -8193,7 +8385,7 @@ ] }, { - "name": "aten::index_copy_", + "name": "aten::index_copy_(Tensor(a!) self, int dim, Tensor index, Tensor source) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -8205,7 +8397,7 @@ ] }, { - "name": "aten::index_copy_.dimname", + "name": "aten::index_copy_.dimname(Tensor(a!) self, Dimname dim, Tensor index, Tensor source) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -8217,7 +8409,7 @@ ] }, { - "name": "aten::index_fill.Dimname_Scalar", + "name": "aten::index_fill.Dimname_Scalar(Tensor self, Dimname dim, Tensor index, Scalar value) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -8229,7 +8421,7 @@ ] }, { - "name": "aten::index_fill.Dimname_Tensor", + "name": "aten::index_fill.Dimname_Tensor(Tensor self, Dimname dim, Tensor index, Tensor value) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -8241,7 +8433,7 @@ ] }, { - "name": "aten::index_fill.int_Scalar", + "name": "aten::index_fill.int_Scalar(Tensor self, int dim, Tensor index, Scalar value) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -8253,7 +8445,7 @@ ] }, { - "name": "aten::index_fill.int_Tensor", + "name": "aten::index_fill.int_Tensor(Tensor self, int dim, Tensor index, Tensor value) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -8265,7 +8457,7 @@ ] }, { - "name": "aten::index_fill_.Dimname_Scalar", + "name": "aten::index_fill_.Dimname_Scalar(Tensor(a!) self, Dimname dim, Tensor index, Scalar value) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -8277,7 +8469,7 @@ ] }, { - "name": "aten::index_fill_.Dimname_Tensor", + "name": "aten::index_fill_.Dimname_Tensor(Tensor(a!) self, Dimname dim, Tensor index, Tensor value) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -8289,7 +8481,7 @@ ] }, { - "name": "aten::index_fill_.int_Scalar", + "name": "aten::index_fill_.int_Scalar(Tensor(a!) self, int dim, Tensor index, Scalar value) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -8301,7 +8493,7 @@ ] }, { - "name": "aten::index_fill_.int_Tensor", + "name": "aten::index_fill_.int_Tensor(Tensor(a!) self, int dim, Tensor index, Tensor value) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -8313,7 +8505,7 @@ ] }, { - "name": "aten::index_put", + "name": "aten::index_put(Tensor self, Tensor?[] indices, Tensor values, bool accumulate=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "indices", "type": "Tensor?[]" }, @@ -8325,7 +8517,7 @@ ] }, { - "name": "aten::index_put.hacked_twin", + "name": "aten::index_put.hacked_twin(Tensor self, Tensor[] indices, Tensor values, bool accumulate=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "indices", "type": "Tensor[]" }, @@ -8337,7 +8529,7 @@ ] }, { - "name": "aten::index_put_", + "name": "aten::index_put_(Tensor(a!) self, Tensor?[] indices, Tensor values, bool accumulate=False) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "indices", "type": "Tensor?[]" }, @@ -8349,7 +8541,7 @@ ] }, { - "name": "aten::index_put_.hacked_twin", + "name": "aten::index_put_.hacked_twin(Tensor(a!) self, Tensor[] indices, Tensor values, bool accumulate=False) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "indices", "type": "Tensor[]" }, @@ -8361,7 +8553,7 @@ ] }, { - "name": "aten::index_reduce", + "name": "aten::index_reduce(Tensor self, int dim, Tensor index, Tensor source, str reduce, *, bool include_self=True) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -8375,7 +8567,7 @@ ] }, { - "name": "aten::index_reduce.out", + "name": "aten::index_reduce.out(Tensor self, int dim, Tensor index, Tensor source, str reduce, *, bool include_self=True, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -8389,7 +8581,7 @@ ] }, { - "name": "aten::index_reduce_", + "name": "aten::index_reduce_(Tensor(a!) self, int dim, Tensor index, Tensor source, str reduce, *, bool include_self=True) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -8403,7 +8595,7 @@ ] }, { - "name": "aten::index_select", + "name": "aten::index_select(Tensor self, int dim, Tensor index) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -8414,7 +8606,7 @@ ] }, { - "name": "aten::index_select.dimname", + "name": "aten::index_select.dimname(Tensor self, Dimname dim, Tensor index) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -8425,7 +8617,7 @@ ] }, { - "name": "aten::index_select.dimname_out", + "name": "aten::index_select.dimname_out(Tensor self, Dimname dim, Tensor index, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -8436,7 +8628,7 @@ ] }, { - "name": "aten::index_select.out", + "name": "aten::index_select.out(Tensor self, int dim, Tensor index, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -8447,7 +8639,7 @@ ] }, { - "name": "aten::index_select_backward", + "name": "aten::index_select_backward(Tensor grad, SymInt[] self_sizes, int dim, Tensor index) -> Tensor", "inputs": [ { "name": "grad", "type": "Tensor" }, { "name": "self_sizes", "type": "SymInt[]" }, @@ -8459,7 +8651,7 @@ ] }, { - "name": "aten::instance_norm", + "name": "aten::instance_norm(Tensor input, Tensor? weight, Tensor? bias, Tensor? running_mean, Tensor? running_var, bool use_input_stats, float momentum, float eps, bool cudnn_enabled) -> Tensor", "category": "Normalization", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -8477,7 +8669,7 @@ ] }, { - "name": "aten::int_repr", + "name": "aten::int_repr(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -8486,7 +8678,7 @@ ] }, { - "name": "aten::int_repr.out", + "name": "aten::int_repr.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -8495,7 +8687,7 @@ ] }, { - "name": "aten::inverse", + "name": "aten::inverse(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -8504,7 +8696,7 @@ ] }, { - "name": "aten::inverse.out", + "name": "aten::inverse.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -8513,7 +8705,7 @@ ] }, { - "name": "aten::is_contiguous", + "name": "aten::is_contiguous(Tensor self) -> bool", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -8522,7 +8714,7 @@ ] }, { - "name": "aten::is_contiguous.memory_format", + "name": "aten::is_contiguous.memory_format(Tensor self, MemoryFormat memory_format) -> bool", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "memory_format", "type": "MemoryFormat" } @@ -8532,7 +8724,7 @@ ] }, { - "name": "aten::is_floating_point", + "name": "aten::is_floating_point(Tensor self) -> bool", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -8541,7 +8733,7 @@ ] }, { - "name": "aten::isfinite", + "name": "aten::isfinite(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -8550,7 +8742,7 @@ ] }, { - "name": "aten::isinf", + "name": "aten::isinf(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -8559,7 +8751,7 @@ ] }, { - "name": "aten::isnan", + "name": "aten::isnan(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -8568,7 +8760,7 @@ ] }, { - "name": "aten::istft", + "name": "aten::istft(Tensor self, int n_fft, int? hop_length=None, int? win_length=None, Tensor? window=None, bool center=True, bool normalized=False, bool? onesided=None, int? length=None, bool return_complex=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "n_fft", "type": "int64" }, @@ -8586,7 +8778,7 @@ ] }, { - "name": "aten::item", + "name": "aten::item(Tensor self) -> Scalar", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -8595,7 +8787,7 @@ ] }, { - "name": "aten::items.Tensor", + "name": "aten::items.Tensor(Dict(Tensor, t) self) -> ((Tensor, t)[])", "inputs": [ { "name": "self", "type": "Dict(Tensor, t)" } ], @@ -8604,7 +8796,7 @@ ] }, { - "name": "aten::items.bool", + "name": "aten::items.bool(Dict(bool, t) self) -> ((bool, t)[])", "inputs": [ { "name": "self", "type": "Dict(boolean, t)" } ], @@ -8613,7 +8805,7 @@ ] }, { - "name": "aten::items.complex", + "name": "aten::items.complex(Dict(complex, t) self) -> ((complex, t)[])", "inputs": [ { "name": "self", "type": "Dict(complex, t)" } ], @@ -8622,7 +8814,7 @@ ] }, { - "name": "aten::items.float", + "name": "aten::items.float(Dict(float, t) self) -> ((float, t)[])", "inputs": [ { "name": "self", "type": "Dict(float32, t)" } ], @@ -8631,7 +8823,7 @@ ] }, { - "name": "aten::items.int", + "name": "aten::items.int(Dict(int, t) self) -> ((int, t)[])", "inputs": [ { "name": "self", "type": "Dict(int64, t)" } ], @@ -8640,7 +8832,7 @@ ] }, { - "name": "aten::items.str", + "name": "aten::items.str(Dict(str, t) self) -> ((str, t)[])", "inputs": [ { "name": "self", "type": "Dict(string, t)" } ], @@ -8649,7 +8841,7 @@ ] }, { - "name": "aten::join", + "name": "aten::join(str self, str[] values) -> str", "inputs": [ { "name": "self", "type": "string" }, { "name": "values", "type": "string[]" } @@ -8659,7 +8851,7 @@ ] }, { - "name": "aten::keys.Tensor", + "name": "aten::keys.Tensor(Dict(Tensor, t) self) -> Tensor[](*)", "inputs": [ { "name": "self", "type": "Dict(Tensor, t)" } ], @@ -8668,7 +8860,7 @@ ] }, { - "name": "aten::keys.bool", + "name": "aten::keys.bool(Dict(bool, t) self) -> bool[](*)", "inputs": [ { "name": "self", "type": "Dict(boolean, t)" } ], @@ -8677,7 +8869,7 @@ ] }, { - "name": "aten::keys.complex", + "name": "aten::keys.complex(Dict(complex, t) self) -> complex[](*)", "inputs": [ { "name": "self", "type": "Dict(complex, t)" } ], @@ -8686,7 +8878,7 @@ ] }, { - "name": "aten::keys.float", + "name": "aten::keys.float(Dict(float, t) self) -> float[](*)", "inputs": [ { "name": "self", "type": "Dict(float32, t)" } ], @@ -8695,7 +8887,7 @@ ] }, { - "name": "aten::keys.int", + "name": "aten::keys.int(Dict(int, t) self) -> int[](*)", "inputs": [ { "name": "self", "type": "Dict(int64, t)" } ], @@ -8704,7 +8896,7 @@ ] }, { - "name": "aten::keys.str", + "name": "aten::keys.str(Dict(str, t) self) -> str[](*)", "inputs": [ { "name": "self", "type": "Dict(string, t)" } ], @@ -8713,7 +8905,7 @@ ] }, { - "name": "aten::kl_div", + "name": "aten::kl_div(Tensor self, Tensor target, int reduction=Mean, *, bool log_target=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "target", "type": "Tensor" }, @@ -8725,7 +8917,7 @@ ] }, { - "name": "aten::kthvalue", + "name": "aten::kthvalue(Tensor self, int k, int dim=-1, bool keepdim=False) -> (Tensor values, Tensor indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "k", "type": "int64" }, @@ -8738,7 +8930,7 @@ ] }, { - "name": "aten::kthvalue.dimname", + "name": "aten::kthvalue.dimname(Tensor self, int k, Dimname dim, bool keepdim=False) -> (Tensor values, Tensor indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "k", "type": "int64" }, @@ -8751,7 +8943,7 @@ ] }, { - "name": "aten::kthvalue.dimname_out", + "name": "aten::kthvalue.dimname_out(Tensor self, int k, Dimname dim, bool keepdim=False, *, Tensor(a!) values, Tensor(b!) indices) -> (Tensor(a!) values, Tensor(b!) indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "k", "type": "int64" }, @@ -8764,7 +8956,7 @@ ] }, { - "name": "aten::kthvalue.values", + "name": "aten::kthvalue.values(Tensor self, int k, int dim=-1, bool keepdim=False, *, Tensor(a!) values, Tensor(b!) indices) -> (Tensor(a!) values, Tensor(b!) indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "k", "type": "int64" }, @@ -8777,7 +8969,7 @@ ] }, { - "name": "aten::l1_loss", + "name": "aten::l1_loss(Tensor self, Tensor target, int reduction=Mean) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "target", "type": "Tensor" }, @@ -8788,7 +8980,7 @@ ] }, { - "name": "aten::layer_norm", + "name": "aten::layer_norm(Tensor input, SymInt[] normalized_shape, Tensor? weight=None, Tensor? bias=None, float eps=1e-05, bool cudnn_enable=True) -> Tensor", "category": "Normalization", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -8803,7 +8995,7 @@ ] }, { - "name": "aten::le", + "name": "aten::le(Scalar a, Scalar b) -> bool", "inputs": [ { "name": "a", "type": "Scalar" }, { "name": "b", "type": "Scalar" } @@ -8813,7 +9005,7 @@ ] }, { - "name": "aten::le.Scalar", + "name": "aten::le.Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -8823,7 +9015,7 @@ ] }, { - "name": "aten::le.Scalar_out", + "name": "aten::le.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -8833,7 +9025,7 @@ ] }, { - "name": "aten::le.Tensor", + "name": "aten::le.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -8843,7 +9035,7 @@ ] }, { - "name": "aten::le.Tensor_out", + "name": "aten::le.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -8853,7 +9045,7 @@ ] }, { - "name": "aten::le.float", + "name": "aten::le.float(float a, float b) -> bool", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "float32" } @@ -8863,7 +9055,7 @@ ] }, { - "name": "aten::le.float_int", + "name": "aten::le.float_int(float a, int b) -> bool", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "int64" } @@ -8873,7 +9065,7 @@ ] }, { - "name": "aten::le.int", + "name": "aten::le.int(int a, int b) -> bool", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "int64" } @@ -8883,7 +9075,7 @@ ] }, { - "name": "aten::le.int_float", + "name": "aten::le.int_float(int a, float b) -> bool", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "float32" } @@ -8893,7 +9085,7 @@ ] }, { - "name": "aten::le.str", + "name": "aten::le.str(str a, str b) -> bool", "inputs": [ { "name": "a", "type": "string" }, { "name": "b", "type": "string" } @@ -8903,7 +9095,7 @@ ] }, { - "name": "aten::leaky_relu", + "name": "aten::leaky_relu(Tensor self, Scalar negative_slope=0.01) -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -8914,7 +9106,7 @@ ] }, { - "name": "aten::leaky_relu.out", + "name": "aten::leaky_relu.out(Tensor self, Scalar negative_slope=0.01, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "negative_slope", "type": "Scalar", "default": 0.01 } @@ -8924,7 +9116,7 @@ ] }, { - "name": "aten::leaky_relu_", + "name": "aten::leaky_relu_(Tensor(a!) self, Scalar negative_slope=0.01) -> Tensor(a!)", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -8935,7 +9127,7 @@ ] }, { - "name": "aten::len.Dict_Tensor", + "name": "aten::len.Dict_Tensor(Dict(Tensor, t) self) -> int", "inputs": [ { "name": "self", "type": "Dict(Tensor, t)" } ], @@ -8944,7 +9136,7 @@ ] }, { - "name": "aten::len.Dict_bool", + "name": "aten::len.Dict_bool(Dict(bool, t) self) -> int", "inputs": [ { "name": "self", "type": "Dict(boolean, t)" } ], @@ -8953,7 +9145,7 @@ ] }, { - "name": "aten::len.Dict_complex", + "name": "aten::len.Dict_complex(Dict(complex, t) self) -> int", "inputs": [ { "name": "self", "type": "Dict(complex, t)" } ], @@ -8962,7 +9154,7 @@ ] }, { - "name": "aten::len.Dict_float", + "name": "aten::len.Dict_float(Dict(float, t) self) -> int", "inputs": [ { "name": "self", "type": "Dict(float32, t)" } ], @@ -8971,7 +9163,7 @@ ] }, { - "name": "aten::len.Dict_int", + "name": "aten::len.Dict_int(Dict(int, t) self) -> int", "inputs": [ { "name": "self", "type": "Dict(int64, t)" } ], @@ -8980,7 +9172,7 @@ ] }, { - "name": "aten::len.Dict_str", + "name": "aten::len.Dict_str(Dict(str, t) self) -> int", "inputs": [ { "name": "self", "type": "Dict(string, t)" } ], @@ -8989,7 +9181,7 @@ ] }, { - "name": "aten::len.Tensor", + "name": "aten::len.Tensor(Tensor t) -> int", "inputs": [ { "name": "t", "type": "Tensor" } ], @@ -8998,7 +9190,7 @@ ] }, { - "name": "aten::len.any", + "name": "aten::len.any(Any[] a) -> int", "inputs": [ { "name": "a", "type": "Any[]" } ], @@ -9007,7 +9199,7 @@ ] }, { - "name": "aten::len.str", + "name": "aten::len.str(str s) -> int", "inputs": [ { "name": "s", "type": "string" } ], @@ -9016,7 +9208,7 @@ ] }, { - "name": "aten::len.t", + "name": "aten::len.t(t[] a) -> int", "inputs": [ { "name": "a", "type": "t[]" } ], @@ -9025,7 +9217,7 @@ ] }, { - "name": "aten::lerp.Scalar", + "name": "aten::lerp.Scalar(Tensor self, Tensor end, Scalar weight) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "end", "type": "Tensor" }, @@ -9036,7 +9228,7 @@ ] }, { - "name": "aten::lerp.Scalar_out", + "name": "aten::lerp.Scalar_out(Tensor self, Tensor end, Scalar weight, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "end", "type": "Tensor" }, @@ -9047,7 +9239,7 @@ ] }, { - "name": "aten::lerp.Tensor", + "name": "aten::lerp.Tensor(Tensor self, Tensor end, Tensor weight) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "end", "type": "Tensor" }, @@ -9058,7 +9250,7 @@ ] }, { - "name": "aten::lerp.Tensor_out", + "name": "aten::lerp.Tensor_out(Tensor self, Tensor end, Tensor weight, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "end", "type": "Tensor" }, @@ -9069,7 +9261,7 @@ ] }, { - "name": "aten::lift_fresh", + "name": "aten::lift_fresh(Tensor(a) self) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -9078,7 +9270,7 @@ ] }, { - "name": "aten::lift_fresh_copy", + "name": "aten::lift_fresh_copy(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -9087,7 +9279,7 @@ ] }, { - "name": "aten::linalg_cross", + "name": "aten::linalg_cross(Tensor self, Tensor other, *, int dim=-1) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -9098,7 +9290,7 @@ ] }, { - "name": "aten::linalg_cross.out", + "name": "aten::linalg_cross.out(Tensor self, Tensor other, *, int dim=-1, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -9109,7 +9301,7 @@ ] }, { - "name": "aten::linalg_inv", + "name": "aten::linalg_inv(Tensor A) -> Tensor", "inputs": [ { "name": "A", "type": "Tensor" } ], @@ -9118,7 +9310,7 @@ ] }, { - "name": "aten::linalg_inv.out", + "name": "aten::linalg_inv.out(Tensor A, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "A", "type": "Tensor" } ], @@ -9127,7 +9319,7 @@ ] }, { - "name": "aten::linalg_inv_ex", + "name": "aten::linalg_inv_ex(Tensor A, *, bool check_errors=False) -> (Tensor inverse, Tensor info)", "inputs": [ { "name": "A", "type": "Tensor" }, { "name": "check_errors", "type": "boolean", "default": false, "kwarg_only": true } @@ -9138,7 +9330,7 @@ ] }, { - "name": "aten::linalg_inv_ex.inverse", + "name": "aten::linalg_inv_ex.inverse(Tensor A, *, bool check_errors=False, Tensor(a!) inverse, Tensor(b!) info) -> (Tensor(a!) inverse, Tensor(b!) info)", "inputs": [ { "name": "A", "type": "Tensor" }, { "name": "check_errors", "type": "boolean", "default": false, "kwarg_only": true } @@ -9149,7 +9341,7 @@ ] }, { - "name": "aten::linalg_norm", + "name": "aten::linalg_norm(Tensor self, Scalar? ord=None, int[1]? dim=None, bool keepdim=False, *, ScalarType? dtype=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "ord", "type": "Scalar?", "default": null }, @@ -9162,7 +9354,7 @@ ] }, { - "name": "aten::linalg_norm.ord_str", + "name": "aten::linalg_norm.ord_str(Tensor self, str ord, int[1]? dim=None, bool keepdim=False, *, ScalarType? dtype=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "ord", "type": "string" }, @@ -9175,7 +9367,7 @@ ] }, { - "name": "aten::linalg_norm.ord_str_out", + "name": "aten::linalg_norm.ord_str_out(Tensor self, str ord, int[1]? dim=None, bool keepdim=False, *, ScalarType? dtype=None, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "ord", "type": "string" }, @@ -9188,7 +9380,7 @@ ] }, { - "name": "aten::linalg_norm.out", + "name": "aten::linalg_norm.out(Tensor self, Scalar? ord=None, int[1]? dim=None, bool keepdim=False, *, ScalarType? dtype=None, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "ord", "type": "Scalar?", "default": null }, @@ -9201,7 +9393,7 @@ ] }, { - "name": "aten::linalg_qr", + "name": "aten::linalg_qr(Tensor A, str mode='reduced') -> (Tensor Q, Tensor R)", "inputs": [ { "name": "A", "type": "Tensor" }, { "name": "mode", "type": "string", "default": "reduced" } @@ -9212,7 +9404,7 @@ ] }, { - "name": "aten::linalg_qr.out", + "name": "aten::linalg_qr.out(Tensor A, str mode='reduced', *, Tensor(a!) Q, Tensor(b!) R) -> (Tensor(a!) Q, Tensor(b!) R)", "inputs": [ { "name": "A", "type": "Tensor" }, { "name": "mode", "type": "string", "default": "reduced" } @@ -9223,7 +9415,7 @@ ] }, { - "name": "aten::linalg_solve", + "name": "aten::linalg_solve(Tensor A, Tensor B, *, bool left=True) -> Tensor", "inputs": [ { "name": "A", "type": "Tensor" }, { "name": "B", "type": "Tensor" }, @@ -9234,7 +9426,7 @@ ] }, { - "name": "aten::linalg_solve.out", + "name": "aten::linalg_solve.out(Tensor A, Tensor B, *, bool left=True, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "A", "type": "Tensor" }, { "name": "B", "type": "Tensor" }, @@ -9245,7 +9437,7 @@ ] }, { - "name": "aten::linalg_solve_ex", + "name": "aten::linalg_solve_ex(Tensor A, Tensor B, *, bool left=True, bool check_errors=False) -> (Tensor result, Tensor info)", "inputs": [ { "name": "A", "type": "Tensor" }, { "name": "B", "type": "Tensor" }, @@ -9258,7 +9450,7 @@ ] }, { - "name": "aten::linalg_solve_ex.out", + "name": "aten::linalg_solve_ex.out(Tensor A, Tensor B, *, bool left=True, bool check_errors=False, Tensor(a!) result, Tensor(b!) info) -> (Tensor(a!) result, Tensor(b!) info)", "inputs": [ { "name": "A", "type": "Tensor" }, { "name": "B", "type": "Tensor" }, @@ -9271,7 +9463,7 @@ ] }, { - "name": "aten::linalg_solve_triangular", + "name": "aten::linalg_solve_triangular(Tensor self, Tensor B, *, bool upper, bool left=True, bool unitriangular=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "B", "type": "Tensor" }, @@ -9284,7 +9476,7 @@ ] }, { - "name": "aten::linalg_solve_triangular.out", + "name": "aten::linalg_solve_triangular.out(Tensor self, Tensor B, *, bool upper, bool left=True, bool unitriangular=False, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "B", "type": "Tensor" }, @@ -9297,7 +9489,7 @@ ] }, { - "name": "aten::linalg_svd", + "name": "aten::linalg_svd(Tensor A, bool full_matrices=True, *, str? driver=None) -> (Tensor U, Tensor S, Tensor Vh)", "inputs": [ { "name": "A", "type": "Tensor" }, { "name": "full_matrices", "type": "boolean", "default": true }, @@ -9310,7 +9502,7 @@ ] }, { - "name": "aten::linalg_svd.U", + "name": "aten::linalg_svd.U(Tensor A, bool full_matrices=True, *, str? driver=None, Tensor(a!) U, Tensor(b!) S, Tensor(c!) Vh) -> (Tensor(a!) U, Tensor(b!) S, Tensor(c!) Vh)", "inputs": [ { "name": "A", "type": "Tensor" }, { "name": "full_matrices", "type": "boolean", "default": true }, @@ -9323,7 +9515,7 @@ ] }, { - "name": "aten::linalg_tensorinv", + "name": "aten::linalg_tensorinv(Tensor self, int ind=2) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "ind", "type": "int64", "default": 2 } @@ -9333,7 +9525,7 @@ ] }, { - "name": "aten::linalg_tensorinv.out", + "name": "aten::linalg_tensorinv.out(Tensor self, int ind=2, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "ind", "type": "int64", "default": 2 } @@ -9343,7 +9535,7 @@ ] }, { - "name": "aten::linalg_tensorsolve", + "name": "aten::linalg_tensorsolve(Tensor self, Tensor other, int[]? dims=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -9354,7 +9546,7 @@ ] }, { - "name": "aten::linalg_tensorsolve.out", + "name": "aten::linalg_tensorsolve.out(Tensor self, Tensor other, int[]? dims=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -9365,7 +9557,7 @@ ] }, { - "name": "aten::linalg_vector_norm", + "name": "aten::linalg_vector_norm(Tensor self, Scalar ord=2, int[1]? dim=None, bool keepdim=False, *, ScalarType? dtype=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "ord", "type": "Scalar", "default": 2 }, @@ -9378,7 +9570,7 @@ ] }, { - "name": "aten::linalg_vector_norm.out", + "name": "aten::linalg_vector_norm.out(Tensor self, Scalar ord=2, int[1]? dim=None, bool keepdim=False, *, ScalarType? dtype=None, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "ord", "type": "Scalar", "default": 2 }, @@ -9391,7 +9583,7 @@ ] }, { - "name": "aten::linear", + "name": "aten::linear(Tensor input, Tensor weight, Tensor? bias=None) -> Tensor", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -9403,7 +9595,7 @@ ] }, { - "name": "aten::linear.out", + "name": "aten::linear.out(Tensor input, Tensor weight, Tensor? bias=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "weight", "type": "Tensor" }, @@ -9414,7 +9606,7 @@ ] }, { - "name": "aten::linear_backward", + "name": "aten::linear_backward(Tensor self, Tensor grad_output, Tensor weight, bool[3] output_mask) -> (Tensor, Tensor, Tensor)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "grad_output", "type": "Tensor" }, @@ -9428,7 +9620,7 @@ ] }, { - "name": "aten::linspace", + "name": "aten::linspace(Scalar start, Scalar end, int steps, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "start", "type": "Scalar" }, { "name": "end", "type": "Scalar" }, @@ -9443,7 +9635,7 @@ ] }, { - "name": "aten::linspace.Scalar_Tensor", + "name": "aten::linspace.Scalar_Tensor(Scalar start, Tensor end, int steps, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "start", "type": "Scalar" }, { "name": "end", "type": "Tensor" }, @@ -9458,7 +9650,7 @@ ] }, { - "name": "aten::linspace.Scalar_Tensor_out", + "name": "aten::linspace.Scalar_Tensor_out(Scalar start, Tensor end, int steps, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "start", "type": "Scalar" }, { "name": "end", "type": "Tensor" }, @@ -9469,7 +9661,7 @@ ] }, { - "name": "aten::linspace.Tensor_Scalar", + "name": "aten::linspace.Tensor_Scalar(Tensor start, Scalar end, int steps, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "start", "type": "Tensor" }, { "name": "end", "type": "Scalar" }, @@ -9484,7 +9676,7 @@ ] }, { - "name": "aten::linspace.Tensor_Scalar_out", + "name": "aten::linspace.Tensor_Scalar_out(Tensor start, Scalar end, int steps, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "start", "type": "Tensor" }, { "name": "end", "type": "Scalar" }, @@ -9495,7 +9687,7 @@ ] }, { - "name": "aten::linspace.Tensor_Tensor", + "name": "aten::linspace.Tensor_Tensor(Tensor start, Tensor end, int steps, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "start", "type": "Tensor" }, { "name": "end", "type": "Tensor" }, @@ -9510,7 +9702,7 @@ ] }, { - "name": "aten::linspace.Tensor_Tensor_out", + "name": "aten::linspace.Tensor_Tensor_out(Tensor start, Tensor end, int steps, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "start", "type": "Tensor" }, { "name": "end", "type": "Tensor" }, @@ -9521,7 +9713,7 @@ ] }, { - "name": "aten::linspace.out", + "name": "aten::linspace.out(Scalar start, Scalar end, int steps, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "start", "type": "Scalar" }, { "name": "end", "type": "Scalar" }, @@ -9532,7 +9724,7 @@ ] }, { - "name": "aten::list", + "name": "aten::list(str t) -> str[]", "inputs": [ { "name": "t", "type": "string" } ], @@ -9541,7 +9733,7 @@ ] }, { - "name": "aten::list.t", + "name": "aten::list.t(t[] l) -> t[]", "inputs": [ { "name": "l", "type": "t[]" } ], @@ -9550,7 +9742,7 @@ ] }, { - "name": "aten::log", + "name": "aten::log(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -9559,7 +9751,7 @@ ] }, { - "name": "aten::log.out", + "name": "aten::log.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -9568,7 +9760,7 @@ ] }, { - "name": "aten::log10", + "name": "aten::log10(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -9577,7 +9769,7 @@ ] }, { - "name": "aten::log10.Scalar", + "name": "aten::log10.Scalar(Scalar a) -> Scalar", "inputs": [ { "name": "a", "type": "Scalar" } ], @@ -9586,7 +9778,7 @@ ] }, { - "name": "aten::log10.complex", + "name": "aten::log10.complex(complex a) -> complex", "inputs": [ { "name": "a", "type": "complex" } ], @@ -9595,7 +9787,7 @@ ] }, { - "name": "aten::log10.float", + "name": "aten::log10.float(float a) -> float", "inputs": [ { "name": "a", "type": "float32" } ], @@ -9604,7 +9796,7 @@ ] }, { - "name": "aten::log10.int", + "name": "aten::log10.int(int a) -> float", "inputs": [ { "name": "a", "type": "int64" } ], @@ -9613,7 +9805,7 @@ ] }, { - "name": "aten::log10.out", + "name": "aten::log10.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -9622,7 +9814,7 @@ ] }, { - "name": "aten::log10_", + "name": "aten::log10_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -9631,7 +9823,7 @@ ] }, { - "name": "aten::log1p", + "name": "aten::log1p(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -9640,7 +9832,7 @@ ] }, { - "name": "aten::log1p.out", + "name": "aten::log1p.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -9649,7 +9841,7 @@ ] }, { - "name": "aten::log1p_", + "name": "aten::log1p_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -9658,7 +9850,7 @@ ] }, { - "name": "aten::log2", + "name": "aten::log2(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -9667,7 +9859,7 @@ ] }, { - "name": "aten::log2.out", + "name": "aten::log2.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -9676,7 +9868,7 @@ ] }, { - "name": "aten::log2_", + "name": "aten::log2_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -9685,7 +9877,7 @@ ] }, { - "name": "aten::log_", + "name": "aten::log_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -9694,7 +9886,7 @@ ] }, { - "name": "aten::log_normal_", + "name": "aten::log_normal_(Tensor(a!) self, float mean=1, float std=2, *, Generator? generator=None) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "mean", "type": "float32", "default": 1 }, @@ -9706,7 +9898,7 @@ ] }, { - "name": "aten::log_sigmoid", + "name": "aten::log_sigmoid(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -9715,7 +9907,7 @@ ] }, { - "name": "aten::log_sigmoid.out", + "name": "aten::log_sigmoid.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -9724,7 +9916,7 @@ ] }, { - "name": "aten::log_sigmoid_backward", + "name": "aten::log_sigmoid_backward(Tensor grad_output, Tensor self, Tensor buffer) -> Tensor", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -9735,7 +9927,7 @@ ] }, { - "name": "aten::log_sigmoid_backward.grad_input", + "name": "aten::log_sigmoid_backward.grad_input(Tensor grad_output, Tensor self, Tensor buffer, *, Tensor(a!) grad_input) -> Tensor(a!)", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -9746,7 +9938,7 @@ ] }, { - "name": "aten::log_sigmoid_forward", + "name": "aten::log_sigmoid_forward(Tensor self) -> (Tensor output, Tensor buffer)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -9756,7 +9948,7 @@ ] }, { - "name": "aten::log_sigmoid_forward.output", + "name": "aten::log_sigmoid_forward.output(Tensor self, *, Tensor(a!) output, Tensor(b!) buffer) -> (Tensor(a!), Tensor(b!))", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -9766,7 +9958,7 @@ ] }, { - "name": "aten::log_softmax.Dimname", + "name": "aten::log_softmax.Dimname(Tensor self, Dimname dim, *, ScalarType? dtype=None) -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -9778,7 +9970,7 @@ ] }, { - "name": "aten::log_softmax.int", + "name": "aten::log_softmax.int(Tensor self, int dim, ScalarType? dtype=None) -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -9790,7 +9982,7 @@ ] }, { - "name": "aten::log_softmax.int_out", + "name": "aten::log_softmax.int_out(Tensor self, int dim, ScalarType? dtype=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -9801,7 +9993,7 @@ ] }, { - "name": "aten::logaddexp", + "name": "aten::logaddexp(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -9811,7 +10003,7 @@ ] }, { - "name": "aten::logaddexp.out", + "name": "aten::logaddexp.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -9821,7 +10013,7 @@ ] }, { - "name": "aten::logaddexp2", + "name": "aten::logaddexp2(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -9831,7 +10023,7 @@ ] }, { - "name": "aten::logaddexp2.out", + "name": "aten::logaddexp2.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -9841,7 +10033,7 @@ ] }, { - "name": "aten::logcumsumexp", + "name": "aten::logcumsumexp(Tensor self, int dim) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" } @@ -9851,7 +10043,7 @@ ] }, { - "name": "aten::logcumsumexp.dimname", + "name": "aten::logcumsumexp.dimname(Tensor self, Dimname dim) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" } @@ -9861,7 +10053,7 @@ ] }, { - "name": "aten::logcumsumexp.dimname_out", + "name": "aten::logcumsumexp.dimname_out(Tensor self, Dimname dim, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" } @@ -9871,7 +10063,7 @@ ] }, { - "name": "aten::logcumsumexp.out", + "name": "aten::logcumsumexp.out(Tensor self, int dim, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" } @@ -9881,7 +10073,7 @@ ] }, { - "name": "aten::logdet", + "name": "aten::logdet(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -9890,7 +10082,7 @@ ] }, { - "name": "aten::logical_and", + "name": "aten::logical_and(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -9900,7 +10092,7 @@ ] }, { - "name": "aten::logical_and.out", + "name": "aten::logical_and.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -9910,7 +10102,7 @@ ] }, { - "name": "aten::logical_and_", + "name": "aten::logical_and_(Tensor(a!) self, Tensor other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -9920,7 +10112,7 @@ ] }, { - "name": "aten::logical_not", + "name": "aten::logical_not(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -9929,7 +10121,7 @@ ] }, { - "name": "aten::logical_not.out", + "name": "aten::logical_not.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -9938,7 +10130,7 @@ ] }, { - "name": "aten::logical_not_", + "name": "aten::logical_not_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -9947,7 +10139,7 @@ ] }, { - "name": "aten::logical_or", + "name": "aten::logical_or(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -9957,7 +10149,7 @@ ] }, { - "name": "aten::logical_or.out", + "name": "aten::logical_or.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -9967,7 +10159,7 @@ ] }, { - "name": "aten::logical_or_", + "name": "aten::logical_or_(Tensor(a!) self, Tensor other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -9977,7 +10169,7 @@ ] }, { - "name": "aten::logical_xor", + "name": "aten::logical_xor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -9987,7 +10179,7 @@ ] }, { - "name": "aten::logical_xor.out", + "name": "aten::logical_xor.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -9997,7 +10189,7 @@ ] }, { - "name": "aten::logical_xor_", + "name": "aten::logical_xor_(Tensor(a!) self, Tensor other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -10007,7 +10199,7 @@ ] }, { - "name": "aten::logit", + "name": "aten::logit(Tensor self, float? eps=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "eps", "type": "float32?", "default": null } @@ -10017,7 +10209,7 @@ ] }, { - "name": "aten::logit.out", + "name": "aten::logit.out(Tensor self, float? eps=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "eps", "type": "float32?", "default": null } @@ -10027,7 +10219,7 @@ ] }, { - "name": "aten::logit_", + "name": "aten::logit_(Tensor(a!) self, float? eps=None) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "eps", "type": "float32?", "default": null } @@ -10037,7 +10229,7 @@ ] }, { - "name": "aten::logit_backward", + "name": "aten::logit_backward(Tensor grad_output, Tensor self, float? eps=None) -> Tensor", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -10048,7 +10240,7 @@ ] }, { - "name": "aten::logit_backward.grad_input", + "name": "aten::logit_backward.grad_input(Tensor grad_output, Tensor self, float? eps=None, *, Tensor(a!) grad_input) -> Tensor(a!)", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -10059,7 +10251,7 @@ ] }, { - "name": "aten::logspace", + "name": "aten::logspace(Scalar start, Scalar end, int steps, float base=10.0, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "start", "type": "Scalar" }, { "name": "end", "type": "Scalar" }, @@ -10075,7 +10267,7 @@ ] }, { - "name": "aten::logspace.Scalar_Tensor", + "name": "aten::logspace.Scalar_Tensor(Scalar start, Tensor end, int steps, float base=10.0, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "start", "type": "Scalar" }, { "name": "end", "type": "Tensor" }, @@ -10091,7 +10283,7 @@ ] }, { - "name": "aten::logspace.Scalar_Tensor_out", + "name": "aten::logspace.Scalar_Tensor_out(Scalar start, Tensor end, int steps, float base=10.0, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "start", "type": "Scalar" }, { "name": "end", "type": "Tensor" }, @@ -10103,7 +10295,7 @@ ] }, { - "name": "aten::logspace.Tensor_Scalar", + "name": "aten::logspace.Tensor_Scalar(Tensor start, Scalar end, int steps, float base=10.0, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "start", "type": "Tensor" }, { "name": "end", "type": "Scalar" }, @@ -10119,7 +10311,7 @@ ] }, { - "name": "aten::logspace.Tensor_Scalar_out", + "name": "aten::logspace.Tensor_Scalar_out(Tensor start, Scalar end, int steps, float base=10.0, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "start", "type": "Tensor" }, { "name": "end", "type": "Scalar" }, @@ -10131,7 +10323,7 @@ ] }, { - "name": "aten::logspace.Tensor_Tensor", + "name": "aten::logspace.Tensor_Tensor(Tensor start, Tensor end, int steps, float base=10.0, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "start", "type": "Tensor" }, { "name": "end", "type": "Tensor" }, @@ -10147,7 +10339,7 @@ ] }, { - "name": "aten::logspace.Tensor_Tensor_out", + "name": "aten::logspace.Tensor_Tensor_out(Tensor start, Tensor end, int steps, float base=10.0, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "start", "type": "Tensor" }, { "name": "end", "type": "Tensor" }, @@ -10159,7 +10351,7 @@ ] }, { - "name": "aten::logspace.out", + "name": "aten::logspace.out(Scalar start, Scalar end, int steps, float base=10.0, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "start", "type": "Scalar" }, { "name": "end", "type": "Scalar" }, @@ -10171,7 +10363,7 @@ ] }, { - "name": "aten::logsumexp", + "name": "aten::logsumexp(Tensor self, int[1] dim, bool keepdim=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]" }, @@ -10182,7 +10374,7 @@ ] }, { - "name": "aten::logsumexp.names", + "name": "aten::logsumexp.names(Tensor self, Dimname[1] dim, bool keepdim=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname[1]" }, @@ -10193,7 +10385,7 @@ ] }, { - "name": "aten::logsumexp.names_out", + "name": "aten::logsumexp.names_out(Tensor self, Dimname[1] dim, bool keepdim=False, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname[1]" }, @@ -10204,7 +10396,7 @@ ] }, { - "name": "aten::logsumexp.out", + "name": "aten::logsumexp.out(Tensor self, int[1] dim, bool keepdim=False, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]" }, @@ -10215,7 +10407,7 @@ ] }, { - "name": "aten::lstm.data", + "name": "aten::lstm.data(Tensor data, Tensor batch_sizes, Tensor[] hx, Tensor[] params, bool has_biases, int num_layers, float dropout, bool train, bool bidirectional) -> (Tensor, Tensor, Tensor)", "category": "Layer", "inputs": [ { "name": "data", "type": "Tensor" }, @@ -10235,7 +10427,7 @@ ] }, { - "name": "aten::lstm.input", + "name": "aten::lstm.input(Tensor input, Tensor[] hx, Tensor[] params, bool has_biases, int num_layers, float dropout, bool train, bool bidirectional, bool batch_first) -> (Tensor, Tensor, Tensor)", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -10255,7 +10447,7 @@ ] }, { - "name": "aten::lstm_cell", + "name": "aten::lstm_cell(Tensor input, Tensor[] hx, Tensor w_ih, Tensor w_hh, Tensor? b_ih=None, Tensor? b_hh=None) -> (Tensor, Tensor)", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -10271,7 +10463,7 @@ ] }, { - "name": "aten::lt", + "name": "aten::lt(Scalar a, Scalar b) -> bool", "inputs": [ { "name": "a", "type": "Scalar" }, { "name": "b", "type": "Scalar" } @@ -10281,7 +10473,7 @@ ] }, { - "name": "aten::lt.Scalar", + "name": "aten::lt.Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -10291,7 +10483,7 @@ ] }, { - "name": "aten::lt.Scalar_out", + "name": "aten::lt.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -10301,7 +10493,7 @@ ] }, { - "name": "aten::lt.Tensor", + "name": "aten::lt.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -10311,7 +10503,7 @@ ] }, { - "name": "aten::lt.Tensor_out", + "name": "aten::lt.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -10321,7 +10513,7 @@ ] }, { - "name": "aten::lt.float", + "name": "aten::lt.float(float a, float b) -> bool", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "float32" } @@ -10331,7 +10523,7 @@ ] }, { - "name": "aten::lt.float_int", + "name": "aten::lt.float_int(float a, int b) -> bool", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "int64" } @@ -10341,7 +10533,7 @@ ] }, { - "name": "aten::lt.int", + "name": "aten::lt.int(int a, int b) -> bool", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "int64" } @@ -10351,7 +10543,7 @@ ] }, { - "name": "aten::lt.int_float", + "name": "aten::lt.int_float(int a, float b) -> bool", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "float32" } @@ -10361,7 +10553,7 @@ ] }, { - "name": "aten::lt.str", + "name": "aten::lt.str(str a, str b) -> bool", "inputs": [ { "name": "a", "type": "string" }, { "name": "b", "type": "string" } @@ -10371,14 +10563,14 @@ ] }, { - "name": "aten::manual_seed", + "name": "aten::manual_seed(int seed) -> ()", "inputs": [ { "name": "seed", "type": "int64" } ], "outputs": [] }, { - "name": "aten::manual_seed.generator", + "name": "aten::manual_seed.generator(Generator(a!) self, int seed) -> Generator(a!)", "inputs": [ { "name": "self", "type": "Generator" }, { "name": "seed", "type": "int64" } @@ -10388,7 +10580,7 @@ ] }, { - "name": "aten::masked_fill.Scalar", + "name": "aten::masked_fill.Scalar(Tensor self, Tensor mask, Scalar value) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "mask", "type": "Tensor" }, @@ -10399,7 +10591,7 @@ ] }, { - "name": "aten::masked_fill.Tensor", + "name": "aten::masked_fill.Tensor(Tensor self, Tensor mask, Tensor value) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "mask", "type": "Tensor" }, @@ -10410,7 +10602,7 @@ ] }, { - "name": "aten::masked_fill_.Scalar", + "name": "aten::masked_fill_.Scalar(Tensor(a!) self, Tensor mask, Scalar value) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "mask", "type": "Tensor" }, @@ -10421,7 +10613,7 @@ ] }, { - "name": "aten::masked_fill_.Tensor", + "name": "aten::masked_fill_.Tensor(Tensor(a!) self, Tensor mask, Tensor value) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "mask", "type": "Tensor" }, @@ -10432,7 +10624,7 @@ ] }, { - "name": "aten::masked_scatter", + "name": "aten::masked_scatter(Tensor self, Tensor mask, Tensor source) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "mask", "type": "Tensor" }, @@ -10443,7 +10635,7 @@ ] }, { - "name": "aten::masked_scatter_", + "name": "aten::masked_scatter_(Tensor(a!) self, Tensor mask, Tensor source) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "mask", "type": "Tensor" }, @@ -10454,7 +10646,7 @@ ] }, { - "name": "aten::masked_select", + "name": "aten::masked_select(Tensor self, Tensor mask) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "mask", "type": "Tensor" } @@ -10464,7 +10656,7 @@ ] }, { - "name": "aten::masked_select.out", + "name": "aten::masked_select.out(Tensor self, Tensor mask, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "mask", "type": "Tensor" } @@ -10474,7 +10666,7 @@ ] }, { - "name": "aten::matmul", + "name": "aten::matmul(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -10484,7 +10676,7 @@ ] }, { - "name": "aten::matmul.out", + "name": "aten::matmul.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -10494,7 +10686,7 @@ ] }, { - "name": "aten::max", + "name": "aten::max(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -10503,7 +10695,7 @@ ] }, { - "name": "aten::max.dim", + "name": "aten::max.dim(Tensor self, int dim, bool keepdim=False) -> (Tensor values, Tensor indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -10515,7 +10707,7 @@ ] }, { - "name": "aten::max.dim_max", + "name": "aten::max.dim_max(Tensor self, int dim, bool keepdim=False, *, Tensor(a!) max, Tensor(b!) max_values) -> (Tensor(a!) values, Tensor(b!) indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -10527,7 +10719,7 @@ ] }, { - "name": "aten::max.names_dim", + "name": "aten::max.names_dim(Tensor self, Dimname dim, bool keepdim=False) -> (Tensor values, Tensor indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -10539,7 +10731,7 @@ ] }, { - "name": "aten::max.names_dim_max", + "name": "aten::max.names_dim_max(Tensor self, Dimname dim, bool keepdim=False, *, Tensor(a!) max, Tensor(b!) max_values) -> (Tensor(a!) values, Tensor(b!) indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -10551,7 +10743,7 @@ ] }, { - "name": "aten::max.other", + "name": "aten::max.other(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -10561,7 +10753,7 @@ ] }, { - "name": "aten::max.out", + "name": "aten::max.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -10571,7 +10763,7 @@ ] }, { - "name": "aten::max.unary_out", + "name": "aten::max.unary_out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -10580,7 +10772,7 @@ ] }, { - "name": "aten::max_pool1d", + "name": "aten::max_pool1d(Tensor self, int[1] kernel_size, int[1] stride=[], int[1] padding=0, int[1] dilation=1, bool ceil_mode=False) -> Tensor", "category": "Pool", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -10595,7 +10787,7 @@ ] }, { - "name": "aten::max_pool1d_with_indices", + "name": "aten::max_pool1d_with_indices(Tensor self, int[1] kernel_size, int[1] stride=[], int[1] padding=0, int[1] dilation=1, bool ceil_mode=False) -> (Tensor, Tensor)", "category": "Pool", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -10611,7 +10803,7 @@ ] }, { - "name": "aten::max_pool2d", + "name": "aten::max_pool2d(Tensor self, int[2] kernel_size, int[2] stride=[], int[2] padding=0, int[2] dilation=1, bool ceil_mode=False) -> Tensor", "category": "Pool", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -10626,7 +10818,7 @@ ] }, { - "name": "aten::max_pool2d_with_indices", + "name": "aten::max_pool2d_with_indices(Tensor self, int[2] kernel_size, int[2] stride=[], int[2] padding=0, int[2] dilation=1, bool ceil_mode=False) -> (Tensor, Tensor)", "category": "Pool", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -10642,7 +10834,7 @@ ] }, { - "name": "aten::max_pool2d_with_indices.out", + "name": "aten::max_pool2d_with_indices.out(Tensor self, int[2] kernel_size, int[2] stride=[], int[2] padding=0, int[2] dilation=1, bool ceil_mode=False, *, Tensor(a!) out, Tensor(b!) indices) -> (Tensor(a!), Tensor(b!))", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "kernel_size", "type": "int64[2]" }, @@ -10657,7 +10849,7 @@ ] }, { - "name": "aten::max_pool3d", + "name": "aten::max_pool3d(Tensor self, int[3] kernel_size, int[3] stride=[], int[3] padding=0, int[3] dilation=1, bool ceil_mode=False) -> Tensor", "category": "Pool", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -10672,7 +10864,7 @@ ] }, { - "name": "aten::max_pool3d_with_indices", + "name": "aten::max_pool3d_with_indices(Tensor self, int[3] kernel_size, int[3] stride=[], int[3] padding=0, int[3] dilation=1, bool ceil_mode=False) -> (Tensor, Tensor)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "kernel_size", "type": "int64[3]" }, @@ -10687,7 +10879,7 @@ ] }, { - "name": "aten::max_pool3d_with_indices.out", + "name": "aten::max_pool3d_with_indices.out(Tensor self, int[3] kernel_size, int[3] stride=[], int[3] padding=0, int[3] dilation=1, bool ceil_mode=False, *, Tensor(a!) out, Tensor(b!) indices) -> (Tensor(a!), Tensor(b!))", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "kernel_size", "type": "int64[3]" }, @@ -10702,7 +10894,7 @@ ] }, { - "name": "aten::max_pool3d_with_indices_backward", + "name": "aten::max_pool3d_with_indices_backward(Tensor grad_output, Tensor self, int[3] kernel_size, int[3] stride, int[3] padding, int[3] dilation, bool ceil_mode, Tensor indices) -> Tensor", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -10718,7 +10910,7 @@ ] }, { - "name": "aten::max_pool3d_with_indices_backward.grad_input", + "name": "aten::max_pool3d_with_indices_backward.grad_input(Tensor grad_output, Tensor self, int[3] kernel_size, int[3] stride, int[3] padding, int[3] dilation, bool ceil_mode, Tensor indices, *, Tensor(a!) grad_input) -> Tensor(a!)", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -10734,7 +10926,7 @@ ] }, { - "name": "aten::max_unpool2d", + "name": "aten::max_unpool2d(Tensor self, Tensor indices, SymInt[2] output_size) -> Tensor", "category": "Pool", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -10746,7 +10938,7 @@ ] }, { - "name": "aten::max_unpool2d.out", + "name": "aten::max_unpool2d.out(Tensor self, Tensor indices, SymInt[2] output_size, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "indices", "type": "Tensor" }, @@ -10757,7 +10949,7 @@ ] }, { - "name": "aten::max_unpool3d", + "name": "aten::max_unpool3d(Tensor self, Tensor indices, SymInt[3] output_size, int[3] stride, int[3] padding) -> Tensor", "category": "Pool", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -10771,7 +10963,7 @@ ] }, { - "name": "aten::max_unpool3d.out", + "name": "aten::max_unpool3d.out(Tensor self, Tensor indices, SymInt[3] output_size, int[3] stride, int[3] padding, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "indices", "type": "Tensor" }, @@ -10784,7 +10976,7 @@ ] }, { - "name": "aten::maximum", + "name": "aten::maximum(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -10794,7 +10986,7 @@ ] }, { - "name": "aten::maximum.out", + "name": "aten::maximum.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -10804,7 +10996,7 @@ ] }, { - "name": "aten::mean", + "name": "aten::mean(Tensor self, *, ScalarType? dtype=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true } @@ -10814,7 +11006,7 @@ ] }, { - "name": "aten::mean.dim", + "name": "aten::mean.dim(Tensor self, int[1]? dim, bool keepdim=False, *, ScalarType? dtype=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]?" }, @@ -10826,7 +11018,7 @@ ] }, { - "name": "aten::mean.dtype_out", + "name": "aten::mean.dtype_out(Tensor self, *, ScalarType? dtype=None, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true } @@ -10836,7 +11028,7 @@ ] }, { - "name": "aten::mean.names_dim", + "name": "aten::mean.names_dim(Tensor self, Dimname[1] dim, bool keepdim=False, *, ScalarType? dtype=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname[1]" }, @@ -10848,7 +11040,7 @@ ] }, { - "name": "aten::mean.names_out", + "name": "aten::mean.names_out(Tensor self, Dimname[1] dim, bool keepdim=False, *, ScalarType? dtype=None, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname[1]" }, @@ -10860,7 +11052,7 @@ ] }, { - "name": "aten::mean.out", + "name": "aten::mean.out(Tensor self, int[1]? dim, bool keepdim=False, *, ScalarType? dtype=None, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]?" }, @@ -10872,7 +11064,7 @@ ] }, { - "name": "aten::median", + "name": "aten::median(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -10881,7 +11073,7 @@ ] }, { - "name": "aten::median.dim", + "name": "aten::median.dim(Tensor self, int dim, bool keepdim=False) -> (Tensor values, Tensor indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -10893,7 +11085,7 @@ ] }, { - "name": "aten::median.dim_values", + "name": "aten::median.dim_values(Tensor self, int dim, bool keepdim=False, *, Tensor(a!) values, Tensor(b!) indices) -> (Tensor(a!) values, Tensor(b!) indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -10905,7 +11097,7 @@ ] }, { - "name": "aten::median.names_dim", + "name": "aten::median.names_dim(Tensor self, Dimname dim, bool keepdim=False) -> (Tensor values, Tensor indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -10917,7 +11109,7 @@ ] }, { - "name": "aten::median.names_dim_values", + "name": "aten::median.names_dim_values(Tensor self, Dimname dim, bool keepdim=False, *, Tensor(a!) values, Tensor(b!) indices) -> (Tensor(a!) values, Tensor(b!) indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -10929,7 +11121,7 @@ ] }, { - "name": "aten::meshgrid", + "name": "aten::meshgrid(Tensor[] tensors) -> Tensor[]", "category": "Tensor", "inputs": [ { "name": "tensors", "type": "Tensor[]" } @@ -10939,7 +11131,7 @@ ] }, { - "name": "aten::meshgrid.indexing", + "name": "aten::meshgrid.indexing(Tensor[] tensors, *, str indexing) -> Tensor[]", "category": "Tensor", "inputs": [ { "name": "tensors", "type": "Tensor[]" }, @@ -10950,7 +11142,7 @@ ] }, { - "name": "aten::min", + "name": "aten::min(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -10959,7 +11151,7 @@ ] }, { - "name": "aten::min.dim", + "name": "aten::min.dim(Tensor self, int dim, bool keepdim=False) -> (Tensor values, Tensor indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64", "default": 0 }, @@ -10971,7 +11163,7 @@ ] }, { - "name": "aten::min.dim_min", + "name": "aten::min.dim_min(Tensor self, int dim, bool keepdim=False, *, Tensor(a!) min, Tensor(b!) min_indices) -> (Tensor(a!) values, Tensor(b!) indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -10983,7 +11175,7 @@ ] }, { - "name": "aten::min.names_dim", + "name": "aten::min.names_dim(Tensor self, Dimname dim, bool keepdim=False) -> (Tensor values, Tensor indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -10995,7 +11187,7 @@ ] }, { - "name": "aten::min.names_dim_min", + "name": "aten::min.names_dim_min(Tensor self, Dimname dim, bool keepdim=False, *, Tensor(a!) min, Tensor(b!) min_indices) -> (Tensor(a!) values, Tensor(b!) indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -11007,7 +11199,7 @@ ] }, { - "name": "aten::min.other", + "name": "aten::min.other(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -11017,7 +11209,7 @@ ] }, { - "name": "aten::min.out", + "name": "aten::min.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -11027,7 +11219,7 @@ ] }, { - "name": "aten::min.unary_out", + "name": "aten::min.unary_out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -11036,7 +11228,7 @@ ] }, { - "name": "aten::minimum", + "name": "aten::minimum(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -11046,7 +11238,7 @@ ] }, { - "name": "aten::minimum.out", + "name": "aten::minimum.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -11056,7 +11248,7 @@ ] }, { - "name": "aten::mish", + "name": "aten::mish(Tensor self) -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" } @@ -11066,7 +11258,7 @@ ] }, { - "name": "aten::mish.out", + "name": "aten::mish.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -11075,7 +11267,7 @@ ] }, { - "name": "aten::mish_", + "name": "aten::mish_(Tensor(a!) self) -> Tensor(a!)", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" } @@ -11085,7 +11277,7 @@ ] }, { - "name": "aten::mkldnn_reorder_conv2d_weight", + "name": "aten::mkldnn_reorder_conv2d_weight(Tensor self, SymInt[2] padding=0, SymInt[2] stride=1, SymInt[2] dilation=1, SymInt groups=1, SymInt[]? input_size=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "padding", "type": "SymInt[2]", "default": 0 }, @@ -11099,7 +11291,7 @@ ] }, { - "name": "aten::mm", + "name": "aten::mm(Tensor self, Tensor mat2) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "mat2", "type": "Tensor" } @@ -11109,7 +11301,7 @@ ] }, { - "name": "aten::mm.out", + "name": "aten::mm.out(Tensor self, Tensor mat2, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "mat2", "type": "Tensor" } @@ -11119,7 +11311,7 @@ ] }, { - "name": "aten::moveaxis.int", + "name": "aten::moveaxis.int(Tensor(a) self, int source, int destination) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "source", "type": "int64" }, @@ -11130,7 +11322,7 @@ ] }, { - "name": "aten::moveaxis.intlist", + "name": "aten::moveaxis.intlist(Tensor(a) self, int[] source, int[] destination) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "source", "type": "int64[]" }, @@ -11141,7 +11333,7 @@ ] }, { - "name": "aten::movedim.int", + "name": "aten::movedim.int(Tensor(a) self, int source, int destination) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "source", "type": "int64" }, @@ -11152,7 +11344,7 @@ ] }, { - "name": "aten::movedim.intlist", + "name": "aten::movedim.intlist(Tensor(a) self, int[] source, int[] destination) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "source", "type": "int64[]" }, @@ -11163,7 +11355,7 @@ ] }, { - "name": "aten::mse_loss", + "name": "aten::mse_loss(Tensor self, Tensor target, int reduction=Mean) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "target", "type": "Tensor" }, @@ -11174,7 +11366,7 @@ ] }, { - "name": "aten::mse_loss.out", + "name": "aten::mse_loss.out(Tensor self, Tensor target, int reduction=Mean, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "target", "type": "Tensor" }, @@ -11185,7 +11377,7 @@ ] }, { - "name": "aten::mse_loss_backward", + "name": "aten::mse_loss_backward(Tensor grad_output, Tensor self, Tensor target, int reduction) -> Tensor", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -11197,7 +11389,7 @@ ] }, { - "name": "aten::mse_loss_backward.grad_input", + "name": "aten::mse_loss_backward.grad_input(Tensor grad_output, Tensor self, Tensor target, int reduction, *, Tensor(a!) grad_input) -> Tensor(a!)", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -11209,7 +11401,7 @@ ] }, { - "name": "aten::mul", + "name": "aten::mul(Scalar a, Scalar b) -> Scalar", "inputs": [ { "name": "a", "type": "Scalar" }, { "name": "b", "type": "Scalar" } @@ -11219,7 +11411,7 @@ ] }, { - "name": "aten::mul.Scalar", + "name": "aten::mul.Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -11229,17 +11421,17 @@ ] }, { - "name": "aten::mul.ScalarT", + "name": "aten::mul.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ - { "name": "input", "type": "Tensor[]" }, + { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } ], "outputs": [ - { "type": "Tensor[]" } + { "type": "Tensor" } ] }, { - "name": "aten::mul.Tensor", + "name": "aten::mul.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -11249,7 +11441,57 @@ ] }, { - "name": "aten::mul.float_int", + "name": "aten::mul.complex(complex a, complex b) -> complex", + "inputs": [ + { "name": "a", "type": "complex" }, + { "name": "b", "type": "complex" } + ], + "outputs": [ + { "type": "complex" } + ] + }, + { + "name": "aten::mul.complex_float(complex a, float b) -> complex", + "inputs": [ + { "name": "a", "type": "complex" }, + { "name": "b", "type": "float32" } + ], + "outputs": [ + { "type": "complex" } + ] + }, + { + "name": "aten::mul.complex_int(complex a, int b) -> complex", + "inputs": [ + { "name": "a", "type": "complex" }, + { "name": "b", "type": "int64" } + ], + "outputs": [ + { "type": "complex" } + ] + }, + { + "name": "aten::mul.float(float a, float b) -> float", + "inputs": [ + { "name": "a", "type": "float32" }, + { "name": "b", "type": "float32" } + ], + "outputs": [ + { "type": "float32" } + ] + }, + { + "name": "aten::mul.float_complex(float a, complex b) -> complex", + "inputs": [ + { "name": "a", "type": "float32" }, + { "name": "b", "type": "complex" } + ], + "outputs": [ + { "type": "complex" } + ] + }, + { + "name": "aten::mul.float_int(float a, int b) -> float", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "int64" } @@ -11259,7 +11501,7 @@ ] }, { - "name": "aten::mul.int", + "name": "aten::mul.int(int a, int b) -> int", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "int64" } @@ -11269,7 +11511,17 @@ ] }, { - "name": "aten::mul.int_float", + "name": "aten::mul.int_complex(int a, complex b) -> complex", + "inputs": [ + { "name": "a", "type": "int64" }, + { "name": "b", "type": "complex" } + ], + "outputs": [ + { "type": "complex" } + ] + }, + { + "name": "aten::mul.int_float(int a, float b) -> float", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "float32" } @@ -11279,7 +11531,7 @@ ] }, { - "name": "aten::mul.left_t", + "name": "aten::mul.left_t(t[] l, int n) -> t[]", "inputs": [ { "name": "l", "type": "t[]" }, { "name": "n", "type": "int64" } @@ -11289,7 +11541,7 @@ ] }, { - "name": "aten::mul.out", + "name": "aten::mul.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -11299,7 +11551,7 @@ ] }, { - "name": "aten::mul.right_", + "name": "aten::mul.right_(int n, t[] l) -> t[]", "inputs": [ { "name": "n", "type": "int64" }, { "name": "l", "type": "t[]" } @@ -11309,7 +11561,7 @@ ] }, { - "name": "aten::mul_.Scalar", + "name": "aten::mul_.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -11319,7 +11571,7 @@ ] }, { - "name": "aten::mul_.Tensor", + "name": "aten::mul_.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -11329,7 +11581,7 @@ ] }, { - "name": "aten::mul_.t", + "name": "aten::mul_.t(t[](a!) l, int n) -> t[](a!)", "inputs": [ { "name": "l", "type": "t[]" }, { "name": "n", "type": "int64" } @@ -11339,7 +11591,7 @@ ] }, { - "name": "aten::multinomial", + "name": "aten::multinomial(Tensor self, int num_samples, bool replacement=False, *, Generator? generator=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "num_samples", "type": "int64" }, @@ -11351,7 +11603,7 @@ ] }, { - "name": "aten::multinomial.out", + "name": "aten::multinomial.out(Tensor self, int num_samples, bool replacement=False, *, Generator? generator=None, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "num_samples", "type": "int64" }, @@ -11363,7 +11615,7 @@ ] }, { - "name": "aten::multiply.Scalar", + "name": "aten::multiply.Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -11373,7 +11625,7 @@ ] }, { - "name": "aten::multiply.Tensor", + "name": "aten::multiply.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -11383,7 +11635,7 @@ ] }, { - "name": "aten::multiply.out", + "name": "aten::multiply.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -11393,7 +11645,7 @@ ] }, { - "name": "aten::multiply_.Scalar", + "name": "aten::multiply_.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -11403,7 +11655,7 @@ ] }, { - "name": "aten::multiply_.Tensor", + "name": "aten::multiply_.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -11413,7 +11665,7 @@ ] }, { - "name": "aten::mv", + "name": "aten::mv(Tensor self, Tensor vec) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "vec", "type": "Tensor" } @@ -11423,7 +11675,7 @@ ] }, { - "name": "aten::mv.out", + "name": "aten::mv.out(Tensor self, Tensor vec, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "vec", "type": "Tensor" } @@ -11433,7 +11685,7 @@ ] }, { - "name": "aten::mvlgamma", + "name": "aten::mvlgamma(Tensor self, int p) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "p", "type": "int64" } @@ -11443,7 +11695,7 @@ ] }, { - "name": "aten::mvlgamma.out", + "name": "aten::mvlgamma.out(Tensor self, int p, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "p", "type": "int64" } @@ -11453,7 +11705,7 @@ ] }, { - "name": "aten::mvlgamma_", + "name": "aten::mvlgamma_(Tensor(a!) self, int p) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "p", "type": "int64" } @@ -11463,7 +11715,7 @@ ] }, { - "name": "aten::nan_to_num", + "name": "aten::nan_to_num(Tensor self, float? nan=None, float? posinf=None, float? neginf=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "nan", "type": "float32?", "default": null }, @@ -11475,7 +11727,7 @@ ] }, { - "name": "aten::nan_to_num.out", + "name": "aten::nan_to_num.out(Tensor self, float? nan=None, float? posinf=None, float? neginf=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "nan", "type": "float32?", "default": null }, @@ -11487,7 +11739,7 @@ ] }, { - "name": "aten::nan_to_num_", + "name": "aten::nan_to_num_(Tensor(a!) self, float? nan=None, float? posinf=None, float? neginf=None) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "nan", "type": "float32?", "default": null }, @@ -11499,7 +11751,7 @@ ] }, { - "name": "aten::narrow", + "name": "aten::narrow(Tensor(a) self, int dim, SymInt start, SymInt length) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -11511,7 +11763,7 @@ ] }, { - "name": "aten::narrow.Tensor", + "name": "aten::narrow.Tensor(Tensor(a) self, int dim, Tensor start, SymInt length) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -11523,7 +11775,7 @@ ] }, { - "name": "aten::narrow_copy", + "name": "aten::narrow_copy(Tensor self, int dim, SymInt start, SymInt length) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -11535,7 +11787,7 @@ ] }, { - "name": "aten::narrow_copy.out", + "name": "aten::narrow_copy.out(Tensor self, int dim, SymInt start, SymInt length, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -11547,7 +11799,7 @@ ] }, { - "name": "aten::native_layer_norm", + "name": "aten::native_layer_norm(Tensor input, SymInt[] normalized_shape, Tensor? weight, Tensor? bias, float eps) -> (Tensor, Tensor, Tensor)", "category": "Normalization", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -11563,7 +11815,17 @@ ] }, { - "name": "aten::ne.Scalar", + "name": "aten::ne(Scalar a, Scalar b) -> bool", + "inputs": [ + { "name": "a", "type": "Scalar" }, + { "name": "b", "type": "Scalar" } + ], + "outputs": [ + { "type": "boolean" } + ] + }, + { + "name": "aten::ne.Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -11573,7 +11835,7 @@ ] }, { - "name": "aten::ne.Scalar_out", + "name": "aten::ne.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -11583,7 +11845,7 @@ ] }, { - "name": "aten::ne.Tensor", + "name": "aten::ne.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -11593,7 +11855,7 @@ ] }, { - "name": "aten::ne.Tensor_list", + "name": "aten::ne.Tensor_list(Tensor[] a, Tensor[] b) -> bool", "inputs": [ { "name": "a", "type": "Tensor[]" }, { "name": "b", "type": "Tensor[]" } @@ -11603,7 +11865,7 @@ ] }, { - "name": "aten::ne.Tensor_out", + "name": "aten::ne.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -11613,7 +11875,7 @@ ] }, { - "name": "aten::ne.bool", + "name": "aten::ne.bool(bool a, bool b) -> bool", "inputs": [ { "name": "a", "type": "boolean" }, { "name": "b", "type": "boolean" } @@ -11623,7 +11885,7 @@ ] }, { - "name": "aten::ne.bool_list", + "name": "aten::ne.bool_list(bool[] a, bool[] b) -> bool", "inputs": [ { "name": "a", "type": "boolean[]" }, { "name": "b", "type": "boolean[]" } @@ -11633,7 +11895,27 @@ ] }, { - "name": "aten::ne.device", + "name": "aten::ne.complex(complex a, complex b) -> bool", + "inputs": [ + { "name": "a", "type": "complex" }, + { "name": "b", "type": "complex" } + ], + "outputs": [ + { "type": "boolean" } + ] + }, + { + "name": "aten::ne.complex_float(complex a, float b) -> bool", + "inputs": [ + { "name": "a", "type": "complex" }, + { "name": "b", "type": "float32" } + ], + "outputs": [ + { "type": "boolean" } + ] + }, + { + "name": "aten::ne.device(Device a, Device b) -> bool", "inputs": [ { "name": "a", "type": "Device" }, { "name": "b", "type": "Device" } @@ -11643,7 +11925,7 @@ ] }, { - "name": "aten::ne.enum", + "name": "aten::ne.enum(AnyEnumType a, AnyEnumType b) -> bool", "inputs": [ { "name": "a", "type": "AnyEnumType" }, { "name": "b", "type": "AnyEnumType" } @@ -11653,7 +11935,7 @@ ] }, { - "name": "aten::ne.float", + "name": "aten::ne.float(float a, float b) -> bool", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "float32" } @@ -11663,7 +11945,27 @@ ] }, { - "name": "aten::ne.float_list", + "name": "aten::ne.float_complex(float a, complex b) -> bool", + "inputs": [ + { "name": "a", "type": "float32" }, + { "name": "b", "type": "complex" } + ], + "outputs": [ + { "type": "boolean" } + ] + }, + { + "name": "aten::ne.float_int(float a, int b) -> bool", + "inputs": [ + { "name": "a", "type": "float32" }, + { "name": "b", "type": "int64" } + ], + "outputs": [ + { "type": "boolean" } + ] + }, + { + "name": "aten::ne.float_list(float[] a, float[] b) -> bool", "inputs": [ { "name": "a", "type": "float32[]" }, { "name": "b", "type": "float32[]" } @@ -11673,7 +11975,7 @@ ] }, { - "name": "aten::ne.int", + "name": "aten::ne.int(int a, int b) -> bool", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "int64" } @@ -11683,7 +11985,17 @@ ] }, { - "name": "aten::ne.int_list", + "name": "aten::ne.int_float(int a, float b) -> bool", + "inputs": [ + { "name": "a", "type": "int64" }, + { "name": "b", "type": "float32" } + ], + "outputs": [ + { "type": "boolean" } + ] + }, + { + "name": "aten::ne.int_list(int[] a, int[] b) -> bool", "inputs": [ { "name": "a", "type": "int64[]" }, { "name": "b", "type": "int64[]" } @@ -11693,7 +12005,7 @@ ] }, { - "name": "aten::ne.str", + "name": "aten::ne.str(str a, str b) -> bool", "inputs": [ { "name": "a", "type": "string" }, { "name": "b", "type": "string" } @@ -11703,7 +12015,7 @@ ] }, { - "name": "aten::ne.str_list", + "name": "aten::ne.str_list(str[] a, str[] b) -> bool", "inputs": [ { "name": "a", "type": "string[]" }, { "name": "b", "type": "string[]" } @@ -11713,7 +12025,7 @@ ] }, { - "name": "aten::neg", + "name": "aten::neg(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -11722,7 +12034,7 @@ ] }, { - "name": "aten::neg.Scalar", + "name": "aten::neg.Scalar(Scalar a) -> Scalar", "inputs": [ { "name": "a", "type": "Scalar" } ], @@ -11731,7 +12043,7 @@ ] }, { - "name": "aten::neg.complex", + "name": "aten::neg.complex(complex a) -> complex", "inputs": [ { "name": "a", "type": "complex" } ], @@ -11740,7 +12052,7 @@ ] }, { - "name": "aten::neg.float", + "name": "aten::neg.float(float a) -> float", "inputs": [ { "name": "a", "type": "float32" } ], @@ -11749,7 +12061,7 @@ ] }, { - "name": "aten::neg.int", + "name": "aten::neg.int(int a) -> int", "inputs": [ { "name": "a", "type": "int64" } ], @@ -11758,7 +12070,7 @@ ] }, { - "name": "aten::neg.out", + "name": "aten::neg.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -11767,7 +12079,7 @@ ] }, { - "name": "aten::nested_to_padded_tensor", + "name": "aten::nested_to_padded_tensor(Tensor self, float padding, int[]? output_size=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "padding", "type": "float32" }, @@ -11778,7 +12090,7 @@ ] }, { - "name": "aten::new_empty", + "name": "aten::new_empty(Tensor self, SymInt[] size, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "size", "type": "SymInt[]" }, @@ -11792,7 +12104,7 @@ ] }, { - "name": "aten::new_empty_strided", + "name": "aten::new_empty_strided(Tensor self, SymInt[] size, SymInt[] stride, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "size", "type": "SymInt[]" }, @@ -11807,7 +12119,7 @@ ] }, { - "name": "aten::new_full", + "name": "aten::new_full(Tensor self, SymInt[] size, Scalar fill_value, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "size", "type": "SymInt[]" }, @@ -11822,7 +12134,7 @@ ] }, { - "name": "aten::new_ones", + "name": "aten::new_ones(Tensor self, SymInt[] size, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "size", "type": "SymInt[]" }, @@ -11836,7 +12148,7 @@ ] }, { - "name": "aten::new_zeros", + "name": "aten::new_zeros(Tensor self, SymInt[] size, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "size", "type": "SymInt[]" }, @@ -11850,7 +12162,7 @@ ] }, { - "name": "aten::nll_loss_nd", + "name": "aten::nll_loss_nd(Tensor self, Tensor target, Tensor? weight=None, int reduction=Mean, SymInt ignore_index=-100) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "target", "type": "Tensor" }, @@ -11863,7 +12175,7 @@ ] }, { - "name": "aten::nonzero", + "name": "aten::nonzero(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -11872,7 +12184,7 @@ ] }, { - "name": "aten::nonzero.out", + "name": "aten::nonzero.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -11881,7 +12193,7 @@ ] }, { - "name": "aten::nonzero_numpy", + "name": "aten::nonzero_numpy(Tensor self) -> Tensor[]", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -11890,7 +12202,7 @@ ] }, { - "name": "aten::norm.Scalar", + "name": "aten::norm.Scalar(Tensor self, Scalar p=2) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "p", "type": "Scalar", "default": 2 } @@ -11900,7 +12212,7 @@ ] }, { - "name": "aten::norm.ScalarOpt_dim", + "name": "aten::norm.ScalarOpt_dim(Tensor self, Scalar? p, int[1] dim, bool keepdim=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "p", "type": "Scalar?" }, @@ -11912,7 +12224,7 @@ ] }, { - "name": "aten::norm.ScalarOpt_dim_dtype", + "name": "aten::norm.ScalarOpt_dim_dtype(Tensor self, Scalar? p, int[1] dim, bool keepdim, *, ScalarType dtype) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "p", "type": "Scalar?" }, @@ -11925,7 +12237,7 @@ ] }, { - "name": "aten::norm.ScalarOpt_dtype", + "name": "aten::norm.ScalarOpt_dtype(Tensor self, Scalar? p, *, ScalarType dtype) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "p", "type": "Scalar?" }, @@ -11936,7 +12248,7 @@ ] }, { - "name": "aten::norm.dtype_out", + "name": "aten::norm.dtype_out(Tensor self, Scalar? p, int[1] dim, bool keepdim, *, ScalarType dtype, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "p", "type": "Scalar?" }, @@ -11949,7 +12261,7 @@ ] }, { - "name": "aten::norm.names_ScalarOpt_dim", + "name": "aten::norm.names_ScalarOpt_dim(Tensor self, Scalar? p, Dimname[1] dim, bool keepdim=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "p", "type": "Scalar?" }, @@ -11961,7 +12273,7 @@ ] }, { - "name": "aten::norm.names_ScalarOpt_dim_dtype", + "name": "aten::norm.names_ScalarOpt_dim_dtype(Tensor self, Scalar? p, Dimname[1] dim, bool keepdim, *, ScalarType dtype) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "p", "type": "Scalar?" }, @@ -11974,7 +12286,7 @@ ] }, { - "name": "aten::norm.names_dtype_out", + "name": "aten::norm.names_dtype_out(Tensor self, Scalar? p, Dimname[1] dim, bool keepdim, *, ScalarType dtype, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "p", "type": "Scalar?" }, @@ -11987,7 +12299,7 @@ ] }, { - "name": "aten::norm.names_out", + "name": "aten::norm.names_out(Tensor self, Scalar? p, Dimname[1] dim, bool keepdim=False, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "p", "type": "Scalar?" }, @@ -11999,7 +12311,7 @@ ] }, { - "name": "aten::norm.out", + "name": "aten::norm.out(Tensor self, Scalar? p, int[1] dim, bool keepdim=False, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "p", "type": "Scalar?" }, @@ -12011,7 +12323,7 @@ ] }, { - "name": "aten::normal.Tensor_Tensor", + "name": "aten::normal.Tensor_Tensor(Tensor mean, Tensor std, *, Generator? generator=None) -> Tensor", "inputs": [ { "name": "mean", "type": "Tensor" }, { "name": "std", "type": "Tensor" }, @@ -12022,7 +12334,7 @@ ] }, { - "name": "aten::normal.Tensor_Tensor_out", + "name": "aten::normal.Tensor_Tensor_out(Tensor mean, Tensor std, *, Generator? generator=None, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "mean", "type": "Tensor" }, { "name": "std", "type": "Tensor" }, @@ -12033,7 +12345,7 @@ ] }, { - "name": "aten::normal.Tensor_float", + "name": "aten::normal.Tensor_float(Tensor mean, float std=1, *, Generator? generator=None) -> Tensor", "inputs": [ { "name": "mean", "type": "Tensor" }, { "name": "std", "type": "float32", "default": 1 }, @@ -12044,7 +12356,7 @@ ] }, { - "name": "aten::normal.Tensor_float_out", + "name": "aten::normal.Tensor_float_out(Tensor mean, float std=1, *, Generator? generator=None, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "mean", "type": "Tensor" }, { "name": "std", "type": "float32", "default": 1 }, @@ -12055,7 +12367,7 @@ ] }, { - "name": "aten::normal.float_Tensor", + "name": "aten::normal.float_Tensor(float mean, Tensor std, *, Generator? generator=None) -> Tensor", "inputs": [ { "name": "mean", "type": "float32" }, { "name": "std", "type": "Tensor" }, @@ -12066,7 +12378,7 @@ ] }, { - "name": "aten::normal.float_Tensor_out", + "name": "aten::normal.float_Tensor_out(float mean, Tensor std, *, Generator? generator=None, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "mean", "type": "float32" }, { "name": "std", "type": "Tensor" }, @@ -12077,7 +12389,7 @@ ] }, { - "name": "aten::normal.float_float", + "name": "aten::normal.float_float(float mean, float std, SymInt[] size, *, Generator? generator=None, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "mean", "type": "float32", "default": 0 }, { "name": "std", "type": "float32" }, @@ -12093,7 +12405,7 @@ ] }, { - "name": "aten::normal.float_float_out", + "name": "aten::normal.float_float_out(float mean, float std, SymInt[] size, *, Generator? generator=None, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "mean", "type": "float32" }, { "name": "std", "type": "float32" }, @@ -12105,7 +12417,7 @@ ] }, { - "name": "aten::normal_", + "name": "aten::normal_(Tensor(a!) self, float mean=0, float std=1, *, Generator? generator=None) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "mean", "type": "float32", "default": 0 }, @@ -12117,7 +12429,7 @@ ] }, { - "name": "aten::numel", + "name": "aten::numel(Tensor self) -> int", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -12126,7 +12438,7 @@ ] }, { - "name": "aten::numpy_T", + "name": "aten::numpy_T(Tensor(a) self) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -12135,7 +12447,7 @@ ] }, { - "name": "aten::numpy_T.a", + "name": "aten::numpy_T.a(Tensor(a) self) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -12144,7 +12456,7 @@ ] }, { - "name": "aten::one_hot", + "name": "aten::one_hot(Tensor self, int num_classes=-1) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "num_classes", "type": "int64", "default": -1 } @@ -12154,7 +12466,7 @@ ] }, { - "name": "aten::ones", + "name": "aten::ones(SymInt[] size, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "size", "type": "SymInt[]" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -12167,7 +12479,7 @@ ] }, { - "name": "aten::ones.names", + "name": "aten::ones.names(int[] size, *, Dimname[]? names, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "size", "type": "int64[]" }, { "name": "names", "type": "Dimname[]?", "kwarg_only": true }, @@ -12181,7 +12493,7 @@ ] }, { - "name": "aten::ones.out", + "name": "aten::ones.out(SymInt[] size, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "size", "type": "SymInt[]" } ], @@ -12190,7 +12502,7 @@ ] }, { - "name": "aten::ones_like", + "name": "aten::ones_like(Tensor self, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None, MemoryFormat? memory_format=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -12204,7 +12516,7 @@ ] }, { - "name": "aten::outer", + "name": "aten::outer(Tensor self, Tensor vec2) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "vec2", "type": "Tensor" } @@ -12214,7 +12526,7 @@ ] }, { - "name": "aten::outer.out", + "name": "aten::outer.out(Tensor self, Tensor vec2, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "vec2", "type": "Tensor" } @@ -12224,7 +12536,7 @@ ] }, { - "name": "aten::pad", + "name": "aten::pad(Tensor self, SymInt[] pad, str mode=\"constant\", float? value=None) -> Tensor", "category": "Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -12237,7 +12549,7 @@ ] }, { - "name": "aten::pad_sequence", + "name": "aten::pad_sequence(Tensor[] sequences, bool batch_first=False, float padding_value=0.0, str padding_side=\"right\") -> Tensor", "inputs": [ { "name": "sequences", "type": "Tensor[]" }, { "name": "batch_first", "type": "boolean", "default": false }, @@ -12249,7 +12561,7 @@ ] }, { - "name": "aten::pairwise_distance", + "name": "aten::pairwise_distance(Tensor x1, Tensor x2, float p=2, float eps=1e-06, bool keepdim=False) -> Tensor", "inputs": [ { "name": "x1", "type": "Tensor" }, { "name": "x2", "type": "Tensor" }, @@ -12262,7 +12574,7 @@ ] }, { - "name": "aten::pdist", + "name": "aten::pdist(Tensor self, float p=2) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "p", "type": "float32", "default": 2 } @@ -12272,7 +12584,7 @@ ] }, { - "name": "aten::permute", + "name": "aten::permute(Tensor(a) self, int[] dims) -> Tensor(a)", "category": "Shape", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -12283,7 +12595,7 @@ ] }, { - "name": "aten::pin_memory", + "name": "aten::pin_memory(Tensor(a) self, Device? device=None) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "device", "type": "Device?", "default": null } @@ -12293,7 +12605,7 @@ ] }, { - "name": "aten::pinverse", + "name": "aten::pinverse(Tensor self, float rcond=1e-15) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "rcond", "type": "float32", "default": 1e-15 } @@ -12303,7 +12615,7 @@ ] }, { - "name": "aten::pixel_shuffle", + "name": "aten::pixel_shuffle(Tensor self, int upscale_factor) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "upscale_factor", "type": "int64" } @@ -12313,7 +12625,7 @@ ] }, { - "name": "aten::pixel_unshuffle", + "name": "aten::pixel_unshuffle(Tensor self, int downscale_factor) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "downscale_factor", "type": "int64" } @@ -12323,7 +12635,7 @@ ] }, { - "name": "aten::polar", + "name": "aten::polar(Tensor abs, Tensor angle) -> Tensor", "inputs": [ { "name": "abs", "type": "Tensor" }, { "name": "angle", "type": "Tensor" } @@ -12333,7 +12645,7 @@ ] }, { - "name": "aten::polar.out", + "name": "aten::polar.out(Tensor abs, Tensor angle, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "abs", "type": "Tensor" }, { "name": "angle", "type": "Tensor" } @@ -12343,7 +12655,7 @@ ] }, { - "name": "aten::pop.t", + "name": "aten::pop.t(t[](a!) self, int idx=-1) -> t(*)", "inputs": [ { "name": "self", "type": "t[]" }, { "name": "idx", "type": "int64", "default": -1 } @@ -12353,7 +12665,7 @@ ] }, { - "name": "aten::pow.Scalar", + "name": "aten::pow.Scalar(Scalar self, Tensor exponent) -> Tensor", "inputs": [ { "name": "self", "type": "Scalar" }, { "name": "exponent", "type": "Tensor" } @@ -12363,7 +12675,7 @@ ] }, { - "name": "aten::pow.Scalar_Scalar", + "name": "aten::pow.Scalar_Scalar(Scalar a, Scalar b) -> float", "inputs": [ { "name": "a", "type": "Scalar" }, { "name": "b", "type": "Scalar" } @@ -12373,7 +12685,7 @@ ] }, { - "name": "aten::pow.Scalar_out", + "name": "aten::pow.Scalar_out(Scalar self, Tensor exponent, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Scalar" }, { "name": "exponent", "type": "Tensor" } @@ -12383,7 +12695,7 @@ ] }, { - "name": "aten::pow.Tensor_Scalar", + "name": "aten::pow.Tensor_Scalar(Tensor self, Scalar exponent) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "exponent", "type": "Scalar" } @@ -12393,7 +12705,7 @@ ] }, { - "name": "aten::pow.Tensor_Scalar_out", + "name": "aten::pow.Tensor_Scalar_out(Tensor self, Scalar exponent, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "exponent", "type": "Scalar" } @@ -12403,7 +12715,7 @@ ] }, { - "name": "aten::pow.Tensor_Tensor", + "name": "aten::pow.Tensor_Tensor(Tensor self, Tensor exponent) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "exponent", "type": "Tensor" } @@ -12413,7 +12725,7 @@ ] }, { - "name": "aten::pow.Tensor_Tensor_out", + "name": "aten::pow.Tensor_Tensor_out(Tensor self, Tensor exponent, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "exponent", "type": "Tensor" } @@ -12423,7 +12735,7 @@ ] }, { - "name": "aten::pow.complex", + "name": "aten::pow.complex(complex a, complex b) -> complex", "inputs": [ { "name": "a", "type": "complex" }, { "name": "b", "type": "complex" } @@ -12433,7 +12745,7 @@ ] }, { - "name": "aten::pow.complex_float", + "name": "aten::pow.complex_float(complex a, float b) -> complex", "inputs": [ { "name": "a", "type": "complex" }, { "name": "b", "type": "float32" } @@ -12443,7 +12755,7 @@ ] }, { - "name": "aten::pow.float", + "name": "aten::pow.float(float a, float b) -> float", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "float32" } @@ -12453,7 +12765,7 @@ ] }, { - "name": "aten::pow.float_complex", + "name": "aten::pow.float_complex(float a, complex b) -> complex", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "complex" } @@ -12463,7 +12775,7 @@ ] }, { - "name": "aten::pow.float_int", + "name": "aten::pow.float_int(float a, int b) -> float", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "int64" } @@ -12473,7 +12785,7 @@ ] }, { - "name": "aten::pow.int", + "name": "aten::pow.int(int a, int b) -> float", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "int64" } @@ -12483,7 +12795,7 @@ ] }, { - "name": "aten::pow.int_float", + "name": "aten::pow.int_float(int a, float b) -> float", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "float32" } @@ -12493,7 +12805,7 @@ ] }, { - "name": "aten::pow.int_to_int", + "name": "aten::pow.int_to_int(int a, int b) -> int", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "int64" } @@ -12503,7 +12815,7 @@ ] }, { - "name": "aten::pow_.Scalar", + "name": "aten::pow_.Scalar(Tensor(a!) self, Scalar exponent) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "exponent", "type": "Scalar" } @@ -12513,7 +12825,7 @@ ] }, { - "name": "aten::pow_.Tensor", + "name": "aten::pow_.Tensor(Tensor(a!) self, Tensor exponent) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "exponent", "type": "Tensor" } @@ -12523,7 +12835,7 @@ ] }, { - "name": "aten::prelu", + "name": "aten::prelu(Tensor self, Tensor weight) -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -12534,7 +12846,7 @@ ] }, { - "name": "aten::prod", + "name": "aten::prod(Tensor self, *, ScalarType? dtype=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true } @@ -12544,7 +12856,7 @@ ] }, { - "name": "aten::prod.Dimname_out", + "name": "aten::prod.Dimname_out(Tensor self, Dimname dim, bool keepdim=False, *, ScalarType? dtype=None, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -12556,7 +12868,7 @@ ] }, { - "name": "aten::prod.dim_Dimname", + "name": "aten::prod.dim_Dimname(Tensor self, Dimname dim, bool keepdim=False, *, ScalarType? dtype=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -12568,7 +12880,7 @@ ] }, { - "name": "aten::prod.dim_int", + "name": "aten::prod.dim_int(Tensor self, int dim, bool keepdim=False, *, ScalarType? dtype=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -12580,7 +12892,7 @@ ] }, { - "name": "aten::prod.int_out", + "name": "aten::prod.int_out(Tensor self, int dim, bool keepdim=False, *, ScalarType? dtype=None, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -12592,7 +12904,7 @@ ] }, { - "name": "aten::put_", + "name": "aten::put_(Tensor(a!) self, Tensor index, Tensor source, bool accumulate=False) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "index", "type": "Tensor" }, @@ -12604,7 +12916,7 @@ ] }, { - "name": "aten::quantile", + "name": "aten::quantile(Tensor self, Tensor q, int? dim=None, bool keepdim=False, *, str interpolation='linear') -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "q", "type": "Tensor" }, @@ -12617,7 +12929,7 @@ ] }, { - "name": "aten::quantile.out", + "name": "aten::quantile.out(Tensor self, Tensor q, int? dim=None, bool keepdim=False, *, str interpolation='linear', Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "q", "type": "Tensor" }, @@ -12630,7 +12942,7 @@ ] }, { - "name": "aten::quantile.scalar", + "name": "aten::quantile.scalar(Tensor self, float q, int? dim=None, bool keepdim=False, *, str interpolation='linear') -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "q", "type": "float32" }, @@ -12643,7 +12955,7 @@ ] }, { - "name": "aten::quantile.scalar_out", + "name": "aten::quantile.scalar_out(Tensor self, float q, int? dim=None, bool keepdim=False, *, str interpolation='linear', Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "q", "type": "float32" }, @@ -12656,7 +12968,7 @@ ] }, { - "name": "aten::quantize_per_channel", + "name": "aten::quantize_per_channel(Tensor self, Tensor scales, Tensor zero_points, int axis, ScalarType dtype) -> Tensor", "category": "Quantization", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -12670,7 +12982,7 @@ ] }, { - "name": "aten::quantize_per_tensor", + "name": "aten::quantize_per_tensor(Tensor self, float scale, int zero_point, ScalarType dtype) -> Tensor", "category": "Quantization", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -12683,7 +12995,7 @@ ] }, { - "name": "aten::quantize_per_tensor.tensor_qparams", + "name": "aten::quantize_per_tensor.tensor_qparams(Tensor self, Tensor scale, Tensor zero_point, ScalarType dtype) -> Tensor", "category": "Quantization", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -12696,7 +13008,7 @@ ] }, { - "name": "aten::quantize_per_tensor.tensors", + "name": "aten::quantize_per_tensor.tensors(Tensor[] tensors, Tensor scales, Tensor zero_points, ScalarType dtype) -> Tensor[]", "category": "Quantization", "inputs": [ { "name": "tensors", "type": "Tensor[]" }, @@ -12709,7 +13021,7 @@ ] }, { - "name": "aten::quantize_per_tensor_dynamic", + "name": "aten::quantize_per_tensor_dynamic(Tensor self, ScalarType dtype, bool reduce_range) -> Tensor", "category": "Quantization", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -12721,7 +13033,7 @@ ] }, { - "name": "aten::quantized_gru.data", + "name": "aten::quantized_gru.data(Tensor data, Tensor batch_sizes, Tensor hx, __torch__.torch.classes.rnn.CellParamsBase[] params, bool has_biases, int num_layers, float dropout, bool train, bool bidirectional) -> (Tensor, Tensor)", "category": "Layer", "inputs": [ { "name": "data", "type": "Tensor" }, @@ -12740,7 +13052,7 @@ ] }, { - "name": "aten::quantized_gru.data_legacy", + "name": "aten::quantized_gru.data_legacy(Tensor data, Tensor batch_sizes, Tensor hx, Tensor[] params, bool has_biases, int num_layers, float dropout, bool train, bool bidirectional) -> (Tensor, Tensor)", "category": "Layer", "inputs": [ { "name": "data", "type": "Tensor" }, @@ -12759,7 +13071,7 @@ ] }, { - "name": "aten::quantized_gru.input", + "name": "aten::quantized_gru.input(Tensor input, Tensor hx, __torch__.torch.classes.rnn.CellParamsBase[] params, bool has_biases, int num_layers, float dropout, bool train, bool bidirectional, bool batch_first) -> (Tensor, Tensor)", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -12778,7 +13090,7 @@ ] }, { - "name": "aten::quantized_gru.input_legacy", + "name": "aten::quantized_gru.input_legacy(Tensor input, Tensor hx, Tensor[] params, bool has_biases, int num_layers, float dropout, bool train, bool bidirectional, bool batch_first) -> (Tensor, Tensor)", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -12797,7 +13109,7 @@ ] }, { - "name": "aten::quantized_lstm.data", + "name": "aten::quantized_lstm.data(Tensor data, Tensor batch_sizes, Tensor[] hx, __torch__.torch.classes.rnn.CellParamsBase[] params, bool has_biases, int num_layers, float dropout, bool train, bool bidirectional, *, ScalarType? dtype=None, bool use_dynamic=False) -> (Tensor, Tensor, Tensor)", "category": "Layer", "inputs": [ { "name": "data", "type": "Tensor" }, @@ -12819,7 +13131,7 @@ ] }, { - "name": "aten::quantized_lstm.data_legacy", + "name": "aten::quantized_lstm.data_legacy(Tensor data, Tensor batch_sizes, Tensor[] hx, Tensor[] params, bool has_biases, int num_layers, float dropout, bool train, bool bidirectional, *, ScalarType? dtype=None, bool use_dynamic=False) -> (Tensor, Tensor, Tensor)", "category": "Layer", "inputs": [ { "name": "data", "type": "Tensor" }, @@ -12841,7 +13153,7 @@ ] }, { - "name": "aten::quantized_lstm.input", + "name": "aten::quantized_lstm.input(Tensor input, Tensor[] hx, __torch__.torch.classes.rnn.CellParamsBase[] params, bool has_biases, int num_layers, float dropout, bool train, bool bidirectional, bool batch_first, *, ScalarType? dtype=None, bool use_dynamic=False) -> (Tensor, Tensor, Tensor)", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -12863,7 +13175,7 @@ ] }, { - "name": "aten::quantized_lstm.input_legacy", + "name": "aten::quantized_lstm.input_legacy(Tensor input, Tensor[] hx, Tensor[] params, bool has_biases, int num_layers, float dropout, bool train, bool bidirectional, bool batch_first, *, ScalarType? dtype=None, bool use_dynamic=False) -> (Tensor, Tensor, Tensor)", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -12885,7 +13197,7 @@ ] }, { - "name": "aten::quantized_lstm_cell", + "name": "aten::quantized_lstm_cell(Tensor input, Tensor[] hx, Tensor w_ih, Tensor w_hh, Tensor b_ih, Tensor b_hh, Tensor packed_ih, Tensor packed_hh, Tensor col_offsets_ih, Tensor col_offsets_hh, Scalar scale_ih, Scalar scale_hh, Scalar zero_point_ih, Scalar zero_point_hh) -> (Tensor, Tensor)", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "hx", "type": "Tensor[]" }, @@ -12908,7 +13220,7 @@ ] }, { - "name": "aten::rand", + "name": "aten::rand(SymInt[] size, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "size", "type": "SymInt[]" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -12921,7 +13233,7 @@ ] }, { - "name": "aten::rand.generator", + "name": "aten::rand.generator(SymInt[] size, *, Generator? generator, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "size", "type": "SymInt[]", "default": null }, { "name": "generator", "type": "Generator?", "kwarg_only": true }, @@ -12935,7 +13247,7 @@ ] }, { - "name": "aten::rand.generator_out", + "name": "aten::rand.generator_out(SymInt[] size, *, Generator? generator, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "size", "type": "SymInt[]" }, { "name": "generator", "type": "Generator?", "kwarg_only": true } @@ -12945,7 +13257,7 @@ ] }, { - "name": "aten::rand.generator_with_names", + "name": "aten::rand.generator_with_names(SymInt[] size, *, Generator? generator, Dimname[]? names, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "size", "type": "SymInt[]" }, { "name": "generator", "type": "Generator?", "kwarg_only": true }, @@ -12960,7 +13272,7 @@ ] }, { - "name": "aten::rand.names", + "name": "aten::rand.names(SymInt[] size, *, Dimname[]? names, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "size", "type": "SymInt[]" }, { "name": "names", "type": "Dimname[]?", "kwarg_only": true }, @@ -12974,7 +13286,7 @@ ] }, { - "name": "aten::rand.out", + "name": "aten::rand.out(SymInt[] size, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "size", "type": "SymInt[]" } ], @@ -12983,7 +13295,7 @@ ] }, { - "name": "aten::rand_like", + "name": "aten::rand_like(Tensor self, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None, MemoryFormat? memory_format=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -12997,7 +13309,7 @@ ] }, { - "name": "aten::randint", + "name": "aten::randint(SymInt high, SymInt[] size, *, ScalarType? dtype=long, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "high", "type": "SymInt" }, { "name": "size", "type": "SymInt[]" }, @@ -13011,7 +13323,7 @@ ] }, { - "name": "aten::randint.generator", + "name": "aten::randint.generator(SymInt high, SymInt[] size, *, Generator? generator, ScalarType? dtype=long, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "high", "type": "SymInt" }, { "name": "size", "type": "SymInt[]" }, @@ -13026,7 +13338,7 @@ ] }, { - "name": "aten::randint.generator_out", + "name": "aten::randint.generator_out(SymInt high, SymInt[] size, *, Generator? generator, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "high", "type": "SymInt" }, { "name": "size", "type": "SymInt[]" }, @@ -13037,7 +13349,7 @@ ] }, { - "name": "aten::randint.low", + "name": "aten::randint.low(SymInt low, SymInt high, SymInt[] size, *, ScalarType? dtype=long, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "low", "type": "SymInt", "default": null }, { "name": "high", "type": "SymInt", "default": null }, @@ -13052,7 +13364,7 @@ ] }, { - "name": "aten::randint.low_generator", + "name": "aten::randint.low_generator(SymInt low, SymInt high, SymInt[] size, *, Generator? generator, ScalarType? dtype=long, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "low", "type": "SymInt", "default": null }, { "name": "high", "type": "SymInt", "default": null }, @@ -13068,7 +13380,7 @@ ] }, { - "name": "aten::randint.low_generator_out", + "name": "aten::randint.low_generator_out(SymInt low, SymInt high, SymInt[] size, *, Generator? generator, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "low", "type": "SymInt" }, { "name": "high", "type": "SymInt" }, @@ -13080,7 +13392,7 @@ ] }, { - "name": "aten::randint.low_out", + "name": "aten::randint.low_out(SymInt low, SymInt high, SymInt[] size, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "low", "type": "SymInt" }, { "name": "high", "type": "SymInt" }, @@ -13091,7 +13403,7 @@ ] }, { - "name": "aten::randint.out", + "name": "aten::randint.out(SymInt high, SymInt[] size, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "high", "type": "SymInt" }, { "name": "size", "type": "SymInt[]" } @@ -13101,7 +13413,7 @@ ] }, { - "name": "aten::randint_like", + "name": "aten::randint_like(Tensor self, SymInt high, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None, MemoryFormat? memory_format=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "high", "type": "SymInt" }, @@ -13116,7 +13428,7 @@ ] }, { - "name": "aten::randint_like.low_dtype", + "name": "aten::randint_like.low_dtype(Tensor self, SymInt low, SymInt high, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None, MemoryFormat? memory_format=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "low", "type": "SymInt" }, @@ -13132,7 +13444,7 @@ ] }, { - "name": "aten::randn", + "name": "aten::randn(SymInt[] size, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "size", "type": "SymInt[]" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -13145,7 +13457,7 @@ ] }, { - "name": "aten::randn.generator", + "name": "aten::randn.generator(SymInt[] size, *, Generator? generator, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "size", "type": "SymInt[]" }, { "name": "generator", "type": "Generator?", "kwarg_only": true }, @@ -13159,7 +13471,7 @@ ] }, { - "name": "aten::randn.generator_out", + "name": "aten::randn.generator_out(SymInt[] size, *, Generator? generator, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "size", "type": "SymInt[]" }, { "name": "generator", "type": "Generator?", "kwarg_only": true } @@ -13169,7 +13481,7 @@ ] }, { - "name": "aten::randn.generator_with_names", + "name": "aten::randn.generator_with_names(SymInt[] size, *, Generator? generator, Dimname[]? names, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "size", "type": "SymInt[]" }, { "name": "generator", "type": "Generator?", "kwarg_only": true }, @@ -13184,7 +13496,7 @@ ] }, { - "name": "aten::randn.names", + "name": "aten::randn.names(SymInt[] size, *, Dimname[]? names, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "size", "type": "SymInt[]" }, { "name": "names", "type": "Dimname[]?", "kwarg_only": true }, @@ -13198,7 +13510,7 @@ ] }, { - "name": "aten::randn.out", + "name": "aten::randn.out(SymInt[] size, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "size", "type": "SymInt[]" } ], @@ -13207,7 +13519,7 @@ ] }, { - "name": "aten::randn_like", + "name": "aten::randn_like(Tensor self, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None, MemoryFormat? memory_format=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -13221,7 +13533,7 @@ ] }, { - "name": "aten::random_", + "name": "aten::random_(Tensor(a!) self, *, Generator? generator=None) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "generator", "type": "Generator?", "default": null, "kwarg_only": true } @@ -13231,7 +13543,7 @@ ] }, { - "name": "aten::random_.from", + "name": "aten::random_.from(Tensor(a!) self, int from, int? to, *, Generator? generator=None) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "from", "type": "int64" }, @@ -13243,7 +13555,7 @@ ] }, { - "name": "aten::random_.to", + "name": "aten::random_.to(Tensor(a!) self, int to, *, Generator? generator=None) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "to", "type": "int64" }, @@ -13254,7 +13566,7 @@ ] }, { - "name": "aten::randperm", + "name": "aten::randperm(SymInt n, *, ScalarType? dtype=long, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "n", "type": "SymInt" }, { "name": "dtype", "type": "ScalarType?", "default": "long", "kwarg_only": true }, @@ -13267,7 +13579,7 @@ ] }, { - "name": "aten::randperm.generator", + "name": "aten::randperm.generator(SymInt n, *, Generator? generator, ScalarType? dtype=long, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "n", "type": "SymInt" }, { "name": "generator", "type": "Generator?", "kwarg_only": true }, @@ -13281,7 +13593,7 @@ ] }, { - "name": "aten::randperm.generator_out", + "name": "aten::randperm.generator_out(SymInt n, *, Generator? generator, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "n", "type": "SymInt" }, { "name": "generator", "type": "Generator?", "kwarg_only": true } @@ -13291,7 +13603,7 @@ ] }, { - "name": "aten::randperm.out", + "name": "aten::randperm.out(SymInt n, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "n", "type": "SymInt" } ], @@ -13300,7 +13612,7 @@ ] }, { - "name": "aten::real", + "name": "aten::real(Tensor(a) self) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -13309,7 +13621,7 @@ ] }, { - "name": "aten::reciprocal", + "name": "aten::reciprocal(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -13318,7 +13630,7 @@ ] }, { - "name": "aten::reciprocal.out", + "name": "aten::reciprocal.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -13327,7 +13639,7 @@ ] }, { - "name": "aten::reflection_pad1d", + "name": "aten::reflection_pad1d(Tensor self, SymInt[2] padding) -> Tensor", "category": "Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -13338,7 +13650,7 @@ ] }, { - "name": "aten::reflection_pad1d.out", + "name": "aten::reflection_pad1d.out(Tensor self, SymInt[2] padding, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "padding", "type": "SymInt[2]" } @@ -13348,7 +13660,7 @@ ] }, { - "name": "aten::reflection_pad2d", + "name": "aten::reflection_pad2d(Tensor self, SymInt[4] padding) -> Tensor", "category": "Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -13359,7 +13671,7 @@ ] }, { - "name": "aten::reflection_pad2d.out", + "name": "aten::reflection_pad2d.out(Tensor self, SymInt[4] padding, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "padding", "type": "SymInt[4]" } @@ -13369,7 +13681,7 @@ ] }, { - "name": "aten::reflection_pad3d", + "name": "aten::reflection_pad3d(Tensor self, SymInt[6] padding) -> Tensor", "category": "Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -13380,7 +13692,7 @@ ] }, { - "name": "aten::reflection_pad3d.out", + "name": "aten::reflection_pad3d.out(Tensor self, SymInt[6] padding, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "padding", "type": "SymInt[6]" } @@ -13390,7 +13702,7 @@ ] }, { - "name": "aten::relu", + "name": "aten::relu(Tensor self) -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" } @@ -13400,7 +13712,7 @@ ] }, { - "name": "aten::relu6", + "name": "aten::relu6(Tensor self) -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" } @@ -13410,7 +13722,7 @@ ] }, { - "name": "aten::relu6_", + "name": "aten::relu6_(Tensor(a!) self) -> Tensor(a!)", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" } @@ -13420,7 +13732,7 @@ ] }, { - "name": "aten::relu_", + "name": "aten::relu_(Tensor(a!) self) -> Tensor(a!)", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" } @@ -13430,7 +13742,7 @@ ] }, { - "name": "aten::remainder", + "name": "aten::remainder(Scalar a, Scalar b) -> Scalar", "inputs": [ { "name": "a", "type": "Scalar" }, { "name": "b", "type": "Scalar" } @@ -13440,7 +13752,7 @@ ] }, { - "name": "aten::remainder.Scalar", + "name": "aten::remainder.Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -13450,7 +13762,7 @@ ] }, { - "name": "aten::remainder.Scalar_Tensor", + "name": "aten::remainder.Scalar_Tensor(Scalar self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Scalar" }, { "name": "other", "type": "Tensor" } @@ -13460,7 +13772,7 @@ ] }, { - "name": "aten::remainder.Scalar_Tensor_out", + "name": "aten::remainder.Scalar_Tensor_out(Scalar self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Scalar" }, { "name": "other", "type": "Tensor" } @@ -13470,7 +13782,7 @@ ] }, { - "name": "aten::remainder.Scalar_out", + "name": "aten::remainder.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -13480,7 +13792,7 @@ ] }, { - "name": "aten::remainder.Tensor", + "name": "aten::remainder.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -13490,7 +13802,7 @@ ] }, { - "name": "aten::remainder.Tensor_out", + "name": "aten::remainder.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -13500,7 +13812,7 @@ ] }, { - "name": "aten::remainder.float", + "name": "aten::remainder.float(float a, float b) -> float", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "float32" } @@ -13510,7 +13822,7 @@ ] }, { - "name": "aten::remainder.float_int", + "name": "aten::remainder.float_int(float a, int b) -> float", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "int64" } @@ -13520,7 +13832,7 @@ ] }, { - "name": "aten::remainder.int", + "name": "aten::remainder.int(int a, int b) -> int", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "int64" } @@ -13530,7 +13842,7 @@ ] }, { - "name": "aten::remainder.int_float", + "name": "aten::remainder.int_float(int a, float b) -> float", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "float32" } @@ -13540,7 +13852,7 @@ ] }, { - "name": "aten::remainder_.Scalar", + "name": "aten::remainder_.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -13550,7 +13862,7 @@ ] }, { - "name": "aten::remainder_.Tensor", + "name": "aten::remainder_.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -13560,7 +13872,7 @@ ] }, { - "name": "aten::renorm", + "name": "aten::renorm(Tensor self, Scalar p, int dim, Scalar maxnorm) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "p", "type": "Scalar" }, @@ -13572,7 +13884,7 @@ ] }, { - "name": "aten::renorm.out", + "name": "aten::renorm.out(Tensor self, Scalar p, int dim, Scalar maxnorm, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "p", "type": "Scalar" }, @@ -13584,7 +13896,7 @@ ] }, { - "name": "aten::repeat", + "name": "aten::repeat(Tensor self, SymInt[] repeats) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "repeats", "type": "SymInt[]" } @@ -13594,7 +13906,7 @@ ] }, { - "name": "aten::repeat_interleave.Tensor", + "name": "aten::repeat_interleave.Tensor(Tensor repeats, *, SymInt? output_size=None) -> Tensor", "inputs": [ { "name": "repeats", "type": "Tensor" }, { "name": "output_size", "type": "SymInt?", "default": null, "kwarg_only": true } @@ -13604,7 +13916,7 @@ ] }, { - "name": "aten::repeat_interleave.self_Tensor", + "name": "aten::repeat_interleave.self_Tensor(Tensor self, Tensor repeats, int? dim=None, *, SymInt? output_size=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "repeats", "type": "Tensor" }, @@ -13616,7 +13928,7 @@ ] }, { - "name": "aten::repeat_interleave.self_int", + "name": "aten::repeat_interleave.self_int(Tensor self, SymInt repeats, int? dim=None, *, SymInt? output_size=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "repeats", "type": "SymInt" }, @@ -13628,7 +13940,7 @@ ] }, { - "name": "aten::replace", + "name": "aten::replace(str self, str old, str new, int max=-1) -> str", "inputs": [ { "name": "self", "type": "string" }, { "name": "old", "type": "string" }, @@ -13640,7 +13952,7 @@ ] }, { - "name": "aten::replication_pad1d", + "name": "aten::replication_pad1d(Tensor self, SymInt[2] padding) -> Tensor", "category": "Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -13651,7 +13963,7 @@ ] }, { - "name": "aten::replication_pad1d.out", + "name": "aten::replication_pad1d.out(Tensor self, SymInt[2] padding, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "padding", "type": "SymInt[2]" } @@ -13661,7 +13973,7 @@ ] }, { - "name": "aten::replication_pad2d", + "name": "aten::replication_pad2d(Tensor self, SymInt[4] padding) -> Tensor", "category": "Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -13672,7 +13984,7 @@ ] }, { - "name": "aten::replication_pad2d.out", + "name": "aten::replication_pad2d.out(Tensor self, SymInt[4] padding, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "padding", "type": "SymInt[4]" } @@ -13682,7 +13994,7 @@ ] }, { - "name": "aten::replication_pad3d", + "name": "aten::replication_pad3d(Tensor self, SymInt[6] padding) -> Tensor", "category": "Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -13693,7 +14005,7 @@ ] }, { - "name": "aten::replication_pad3d.out", + "name": "aten::replication_pad3d.out(Tensor self, SymInt[6] padding, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "padding", "type": "SymInt[6]" } @@ -13703,7 +14015,7 @@ ] }, { - "name": "aten::requires_grad_", + "name": "aten::requires_grad_(Tensor(a!) self, bool requires_grad=True) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "requires_grad", "type": "boolean", "default": true } @@ -13713,7 +14025,7 @@ ] }, { - "name": "aten::reshape", + "name": "aten::reshape(Tensor(a) self, SymInt[] shape) -> Tensor(a)", "category": "Shape", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -13724,7 +14036,7 @@ ] }, { - "name": "aten::reshape_as", + "name": "aten::reshape_as(Tensor(a) self, Tensor other) -> Tensor(a)", "category": "Shape", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -13735,7 +14047,7 @@ ] }, { - "name": "aten::resize_", + "name": "aten::resize_(Tensor(a!) self, SymInt[] size, *, MemoryFormat? memory_format=None) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "size", "type": "SymInt[]" }, @@ -13746,7 +14058,7 @@ ] }, { - "name": "aten::resolve_conj", + "name": "aten::resolve_conj(Tensor(a) self) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -13755,7 +14067,7 @@ ] }, { - "name": "aten::resolve_neg", + "name": "aten::resolve_neg(Tensor(a) self) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -13764,14 +14076,14 @@ ] }, { - "name": "aten::reverse.t", + "name": "aten::reverse.t(t[](a!) self) -> ()", "inputs": [ { "name": "self", "type": "t[]" } ], "outputs": [] }, { - "name": "aten::rnn_relu.data", + "name": "aten::rnn_relu.data(Tensor data, Tensor batch_sizes, Tensor hx, Tensor[] params, bool has_biases, int num_layers, float dropout, bool train, bool bidirectional) -> (Tensor, Tensor)", "inputs": [ { "name": "data", "type": "Tensor" }, { "name": "batch_sizes", "type": "Tensor" }, @@ -13789,7 +14101,7 @@ ] }, { - "name": "aten::rnn_relu.input", + "name": "aten::rnn_relu.input(Tensor input, Tensor hx, Tensor[] params, bool has_biases, int num_layers, float dropout, bool train, bool bidirectional, bool batch_first) -> (Tensor, Tensor)", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -13808,7 +14120,7 @@ ] }, { - "name": "aten::rnn_tanh.data", + "name": "aten::rnn_tanh.data(Tensor data, Tensor batch_sizes, Tensor hx, Tensor[] params, bool has_biases, int num_layers, float dropout, bool train, bool bidirectional) -> (Tensor, Tensor)", "inputs": [ { "name": "data", "type": "Tensor" }, { "name": "batch_sizes", "type": "Tensor" }, @@ -13826,7 +14138,7 @@ ] }, { - "name": "aten::rnn_tanh.input", + "name": "aten::rnn_tanh.input(Tensor input, Tensor hx, Tensor[] params, bool has_biases, int num_layers, float dropout, bool train, bool bidirectional, bool batch_first) -> (Tensor, Tensor)", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -13845,7 +14157,7 @@ ] }, { - "name": "aten::rnn_tanh_cell", + "name": "aten::rnn_tanh_cell(Tensor input, Tensor hx, Tensor w_ih, Tensor w_hh, Tensor? b_ih=None, Tensor? b_hh=None) -> Tensor", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "hx", "type": "Tensor" }, @@ -13859,7 +14171,7 @@ ] }, { - "name": "aten::roll", + "name": "aten::roll(Tensor self, SymInt[1] shifts, int[1] dims=[]) -> Tensor", "category": "Layer", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -13871,7 +14183,7 @@ ] }, { - "name": "aten::rot90", + "name": "aten::rot90(Tensor self, int k=1, int[] dims=[0,1]) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "k", "type": "int64", "default": 1 }, @@ -13882,7 +14194,7 @@ ] }, { - "name": "aten::round", + "name": "aten::round(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -13891,7 +14203,7 @@ ] }, { - "name": "aten::round.decimals", + "name": "aten::round.decimals(Tensor self, *, int decimals) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "decimals", "type": "int64", "kwarg_only": true } @@ -13901,7 +14213,7 @@ ] }, { - "name": "aten::round.decimals_out", + "name": "aten::round.decimals_out(Tensor self, *, int decimals, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "decimals", "type": "int64", "kwarg_only": true } @@ -13911,7 +14223,7 @@ ] }, { - "name": "aten::round.out", + "name": "aten::round.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -13920,7 +14232,7 @@ ] }, { - "name": "aten::round_", + "name": "aten::round_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -13929,7 +14241,7 @@ ] }, { - "name": "aten::round_.decimals", + "name": "aten::round_.decimals(Tensor(a!) self, *, int decimals) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "decimals", "type": "int64", "kwarg_only": true } @@ -13939,7 +14251,7 @@ ] }, { - "name": "aten::rrelu", + "name": "aten::rrelu(Tensor self, Scalar lower=0.125, Scalar upper=0.3333333333333333, bool training=False, Generator? generator=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "lower", "type": "Scalar", "default": 0.125 }, @@ -13952,7 +14264,7 @@ ] }, { - "name": "aten::rsqrt", + "name": "aten::rsqrt(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -13961,7 +14273,7 @@ ] }, { - "name": "aten::rsqrt.out", + "name": "aten::rsqrt.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -13970,7 +14282,7 @@ ] }, { - "name": "aten::rsqrt_", + "name": "aten::rsqrt_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -13979,7 +14291,7 @@ ] }, { - "name": "aten::rsub.Scalar", + "name": "aten::rsub.Scalar(Tensor self, Scalar other, Scalar alpha=1) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" }, @@ -13990,7 +14302,7 @@ ] }, { - "name": "aten::rsub.Tensor", + "name": "aten::rsub.Tensor(Tensor self, Tensor other, *, Scalar alpha=1) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -14001,7 +14313,7 @@ ] }, { - "name": "aten::scalar_tensor", + "name": "aten::scalar_tensor(Scalar s, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "s", "type": "Scalar" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -14014,7 +14326,7 @@ ] }, { - "name": "aten::scaled_dot_product_attention", + "name": "aten::scaled_dot_product_attention(Tensor query, Tensor key, Tensor value, Tensor? attn_mask=None, float dropout_p=0.0, bool is_causal=False, *, float? scale=None, bool enable_gqa=False) -> Tensor", "category": "Attention", "inputs": [ { "name": "query", "type": "Tensor" }, @@ -14031,7 +14343,7 @@ ] }, { - "name": "aten::scatter.dimname_src", + "name": "aten::scatter.dimname_src(Tensor self, Dimname dim, Tensor index, Tensor src) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -14043,7 +14355,7 @@ ] }, { - "name": "aten::scatter.dimname_value", + "name": "aten::scatter.dimname_value(Tensor self, Dimname dim, Tensor index, Scalar value) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -14055,7 +14367,7 @@ ] }, { - "name": "aten::scatter.reduce", + "name": "aten::scatter.reduce(Tensor self, int dim, Tensor index, Tensor src, *, str reduce) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -14068,7 +14380,7 @@ ] }, { - "name": "aten::scatter.reduce_out", + "name": "aten::scatter.reduce_out(Tensor self, int dim, Tensor index, Tensor src, *, str reduce, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -14081,7 +14393,7 @@ ] }, { - "name": "aten::scatter.src", + "name": "aten::scatter.src(Tensor self, int dim, Tensor index, Tensor src) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -14093,7 +14405,7 @@ ] }, { - "name": "aten::scatter.src_out", + "name": "aten::scatter.src_out(Tensor self, int dim, Tensor index, Tensor src, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -14105,7 +14417,7 @@ ] }, { - "name": "aten::scatter.value", + "name": "aten::scatter.value(Tensor self, int dim, Tensor index, Scalar value) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -14117,7 +14429,7 @@ ] }, { - "name": "aten::scatter.value_out", + "name": "aten::scatter.value_out(Tensor self, int dim, Tensor index, Scalar value, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -14129,7 +14441,7 @@ ] }, { - "name": "aten::scatter.value_reduce", + "name": "aten::scatter.value_reduce(Tensor self, int dim, Tensor index, Scalar value, *, str reduce) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -14142,7 +14454,7 @@ ] }, { - "name": "aten::scatter.value_reduce_out", + "name": "aten::scatter.value_reduce_out(Tensor self, int dim, Tensor index, Scalar value, *, str reduce, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -14155,7 +14467,7 @@ ] }, { - "name": "aten::scatter_.reduce", + "name": "aten::scatter_.reduce(Tensor(a!) self, int dim, Tensor index, Tensor src, *, str reduce) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -14168,7 +14480,7 @@ ] }, { - "name": "aten::scatter_.src", + "name": "aten::scatter_.src(Tensor(a!) self, int dim, Tensor index, Tensor src) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -14180,7 +14492,7 @@ ] }, { - "name": "aten::scatter_.value", + "name": "aten::scatter_.value(Tensor(a!) self, int dim, Tensor index, Scalar value) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -14192,7 +14504,7 @@ ] }, { - "name": "aten::scatter_.value_reduce", + "name": "aten::scatter_.value_reduce(Tensor(a!) self, int dim, Tensor index, Scalar value, *, str reduce) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -14205,7 +14517,7 @@ ] }, { - "name": "aten::scatter_add", + "name": "aten::scatter_add(Tensor self, int dim, Tensor index, Tensor src) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -14217,7 +14529,7 @@ ] }, { - "name": "aten::scatter_add.dimname", + "name": "aten::scatter_add.dimname(Tensor self, Dimname dim, Tensor index, Tensor src) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -14229,7 +14541,7 @@ ] }, { - "name": "aten::scatter_add.out", + "name": "aten::scatter_add.out(Tensor self, int dim, Tensor index, Tensor src, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -14241,7 +14553,7 @@ ] }, { - "name": "aten::scatter_add_", + "name": "aten::scatter_add_(Tensor(a!) self, int dim, Tensor index, Tensor src) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -14253,7 +14565,7 @@ ] }, { - "name": "aten::scatter_reduce.two", + "name": "aten::scatter_reduce.two(Tensor self, int dim, Tensor index, Tensor src, str reduce, *, bool include_self=True) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -14267,7 +14579,7 @@ ] }, { - "name": "aten::scatter_reduce.two_out", + "name": "aten::scatter_reduce.two_out(Tensor self, int dim, Tensor index, Tensor src, str reduce, *, bool include_self=True, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -14281,7 +14593,7 @@ ] }, { - "name": "aten::scatter_reduce_.two", + "name": "aten::scatter_reduce_.two(Tensor(a!) self, int dim, Tensor index, Tensor src, str reduce, *, bool include_self=True) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -14295,7 +14607,7 @@ ] }, { - "name": "aten::searchsorted.Scalar", + "name": "aten::searchsorted.Scalar(Tensor sorted_sequence, Scalar self, *, bool out_int32=False, bool right=False, str? side=None, Tensor? sorter=None) -> Tensor", "inputs": [ { "name": "sorted_sequence", "type": "Tensor" }, { "name": "self", "type": "Scalar" }, @@ -14309,7 +14621,7 @@ ] }, { - "name": "aten::searchsorted.Scalar_out", + "name": "aten::searchsorted.Scalar_out(Tensor sorted_sequence, Scalar self, *, bool out_int32=False, bool right=False, str? side=None, Tensor? sorter=None, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "sorted_sequence", "type": "Tensor" }, { "name": "self", "type": "Scalar" }, @@ -14323,7 +14635,7 @@ ] }, { - "name": "aten::searchsorted.Tensor", + "name": "aten::searchsorted.Tensor(Tensor sorted_sequence, Tensor self, *, bool out_int32=False, bool right=False, str? side=None, Tensor? sorter=None) -> Tensor", "inputs": [ { "name": "sorted_sequence", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -14337,7 +14649,7 @@ ] }, { - "name": "aten::searchsorted.Tensor_out", + "name": "aten::searchsorted.Tensor_out(Tensor sorted_sequence, Tensor self, *, bool out_int32=False, bool right=False, str? side=None, Tensor? sorter=None, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "sorted_sequence", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -14351,7 +14663,7 @@ ] }, { - "name": "aten::select.Dimname", + "name": "aten::select.Dimname(Tensor(a) self, Dimname dim, int index) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -14362,7 +14674,7 @@ ] }, { - "name": "aten::select.int", + "name": "aten::select.int(Tensor(a) self, int dim, SymInt index) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -14373,7 +14685,7 @@ ] }, { - "name": "aten::select.t", + "name": "aten::select.t(t[](a) list, int idx) -> t(*)", "inputs": [ { "name": "list", "type": "t[]" }, { "name": "idx", "type": "int64" } @@ -14383,7 +14695,7 @@ ] }, { - "name": "aten::select_backward", + "name": "aten::select_backward(Tensor grad_output, SymInt[] input_sizes, int dim, SymInt index) -> Tensor", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "input_sizes", "type": "SymInt[]" }, @@ -14395,7 +14707,7 @@ ] }, { - "name": "aten::select_copy.int", + "name": "aten::select_copy.int(Tensor self, int dim, SymInt index) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -14406,7 +14718,7 @@ ] }, { - "name": "aten::select_scatter", + "name": "aten::select_scatter(Tensor self, Tensor src, int dim, SymInt index) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "src", "type": "Tensor" }, @@ -14418,7 +14730,7 @@ ] }, { - "name": "aten::selu", + "name": "aten::selu(Tensor self) -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" } @@ -14428,7 +14740,7 @@ ] }, { - "name": "aten::selu_", + "name": "aten::selu_(Tensor(a!) self) -> Tensor(a!)", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" } @@ -14438,7 +14750,7 @@ ] }, { - "name": "aten::sigmoid", + "name": "aten::sigmoid(Tensor self) -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" } @@ -14448,7 +14760,7 @@ ] }, { - "name": "aten::sigmoid.out", + "name": "aten::sigmoid.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -14457,7 +14769,7 @@ ] }, { - "name": "aten::sigmoid_", + "name": "aten::sigmoid_(Tensor(a!) self) -> Tensor(a!)", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" } @@ -14467,7 +14779,7 @@ ] }, { - "name": "aten::sign", + "name": "aten::sign(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -14476,7 +14788,7 @@ ] }, { - "name": "aten::sign.out", + "name": "aten::sign.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -14485,7 +14797,7 @@ ] }, { - "name": "aten::sign_", + "name": "aten::sign_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -14494,7 +14806,7 @@ ] }, { - "name": "aten::signbit", + "name": "aten::signbit(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -14503,7 +14815,7 @@ ] }, { - "name": "aten::signbit.out", + "name": "aten::signbit.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -14512,7 +14824,7 @@ ] }, { - "name": "aten::silu", + "name": "aten::silu(Tensor self) -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" } @@ -14522,7 +14834,7 @@ ] }, { - "name": "aten::silu.out", + "name": "aten::silu.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -14531,7 +14843,7 @@ ] }, { - "name": "aten::silu_", + "name": "aten::silu_(Tensor(a!) self) -> Tensor(a!)", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" } @@ -14541,7 +14853,7 @@ ] }, { - "name": "aten::silu_backward", + "name": "aten::silu_backward(Tensor grad_output, Tensor self) -> Tensor", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "self", "type": "Tensor" } @@ -14551,7 +14863,7 @@ ] }, { - "name": "aten::silu_backward.grad_input", + "name": "aten::silu_backward.grad_input(Tensor grad_output, Tensor self, *, Tensor(a!) grad_input) -> Tensor(a!)", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "self", "type": "Tensor" } @@ -14561,7 +14873,7 @@ ] }, { - "name": "aten::sin", + "name": "aten::sin(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -14570,7 +14882,7 @@ ] }, { - "name": "aten::sin.out", + "name": "aten::sin.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -14579,7 +14891,7 @@ ] }, { - "name": "aten::sinh", + "name": "aten::sinh(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -14588,7 +14900,7 @@ ] }, { - "name": "aten::sinh.out", + "name": "aten::sinh.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -14597,7 +14909,7 @@ ] }, { - "name": "aten::size", + "name": "aten::size(Tensor self) -> int[]", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -14606,7 +14918,7 @@ ] }, { - "name": "aten::size.Dimname", + "name": "aten::size.Dimname(Tensor self, Dimname dim) -> int", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" } @@ -14616,7 +14928,7 @@ ] }, { - "name": "aten::size.int", + "name": "aten::size.int(Tensor self, int dim) -> int", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" } @@ -14626,7 +14938,7 @@ ] }, { - "name": "aten::slice.Tensor", + "name": "aten::slice.Tensor(Tensor(a) self, int dim=0, SymInt? start=None, SymInt? end=None, SymInt step=1) -> Tensor(a)", "category": "Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -14640,7 +14952,7 @@ ] }, { - "name": "aten::slice.str", + "name": "aten::slice.str(str string, int? start=None, int? end=None, int step=1) -> str", "inputs": [ { "name": "string", "type": "string" }, { "name": "start", "type": "int64?", "default": null }, @@ -14652,7 +14964,7 @@ ] }, { - "name": "aten::slice.t", + "name": "aten::slice.t(t[] l, int? start=None, int? end=None, int step=1) -> t[]", "inputs": [ { "name": "l", "type": "t[]" }, { "name": "start", "type": "int64?", "default": null }, @@ -14664,7 +14976,7 @@ ] }, { - "name": "aten::slice_copy.Tensor", + "name": "aten::slice_copy.Tensor(Tensor self, int dim=0, SymInt? start=None, SymInt? end=None, SymInt step=1) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64", "default": 0 }, @@ -14677,7 +14989,7 @@ ] }, { - "name": "aten::slice_scatter", + "name": "aten::slice_scatter(Tensor self, Tensor src, int dim=0, SymInt? start=None, SymInt? end=None, SymInt step=1) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "src", "type": "Tensor" }, @@ -14691,7 +15003,7 @@ ] }, { - "name": "aten::smooth_l1_loss", + "name": "aten::smooth_l1_loss(Tensor self, Tensor target, int reduction=Mean, float beta=1.0) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "target", "type": "Tensor" }, @@ -14703,7 +15015,7 @@ ] }, { - "name": "aten::smooth_l1_loss.out", + "name": "aten::smooth_l1_loss.out(Tensor self, Tensor target, int reduction=Mean, float beta=1.0, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "target", "type": "Tensor" }, @@ -14715,7 +15027,7 @@ ] }, { - "name": "aten::smooth_l1_loss_backward", + "name": "aten::smooth_l1_loss_backward(Tensor grad_output, Tensor self, Tensor target, int reduction, float beta) -> Tensor", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -14728,7 +15040,7 @@ ] }, { - "name": "aten::smooth_l1_loss_backward.grad_input", + "name": "aten::smooth_l1_loss_backward.grad_input(Tensor grad_output, Tensor self, Tensor target, int reduction, float beta, *, Tensor(a!) grad_input) -> Tensor(a!)", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -14741,7 +15053,7 @@ ] }, { - "name": "aten::softmax.Dimname", + "name": "aten::softmax.Dimname(Tensor self, Dimname dim, *, ScalarType? dtype=None) -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -14753,7 +15065,7 @@ ] }, { - "name": "aten::softmax.int", + "name": "aten::softmax.int(Tensor self, int dim, ScalarType? dtype=None) -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -14765,7 +15077,7 @@ ] }, { - "name": "aten::softmax.int_out", + "name": "aten::softmax.int_out(Tensor self, int dim, ScalarType? dtype=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -14776,7 +15088,7 @@ ] }, { - "name": "aten::softplus", + "name": "aten::softplus(Tensor self, Scalar beta=1, Scalar threshold=20) -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -14788,7 +15100,7 @@ ] }, { - "name": "aten::softplus.out", + "name": "aten::softplus.out(Tensor self, Scalar beta=1, Scalar threshold=20, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "beta", "type": "Scalar", "default": 1 }, @@ -14799,7 +15111,7 @@ ] }, { - "name": "aten::softshrink", + "name": "aten::softshrink(Tensor self, Scalar lambd=0.5) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "lambd", "type": "Scalar", "default": 0.5 } @@ -14809,7 +15121,7 @@ ] }, { - "name": "aten::softshrink.out", + "name": "aten::softshrink.out(Tensor self, Scalar lambd=0.5, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "lambd", "type": "Scalar", "default": 0.5 } @@ -14819,7 +15131,7 @@ ] }, { - "name": "aten::sort", + "name": "aten::sort(Tensor self, int dim=-1, bool descending=False) -> (Tensor values, Tensor indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64", "default": -1 }, @@ -14831,7 +15143,7 @@ ] }, { - "name": "aten::sort.Tensor", + "name": "aten::sort.Tensor(Tensor[](a!) self, bool reverse=False) -> ()", "inputs": [ { "name": "self", "type": "Tensor[]" }, { "name": "reverse", "type": "boolean", "default": false } @@ -14839,7 +15151,7 @@ "outputs": [] }, { - "name": "aten::sort.any", + "name": "aten::sort.any(t[](a!) self, bool reverse=False) -> ()", "inputs": [ { "name": "self", "type": "t[]" }, { "name": "reverse", "type": "boolean", "default": false } @@ -14847,7 +15159,7 @@ "outputs": [] }, { - "name": "aten::sort.bool", + "name": "aten::sort.bool(bool[](a!) self, bool reverse=False) -> ()", "inputs": [ { "name": "self", "type": "boolean[]" }, { "name": "reverse", "type": "boolean", "default": false } @@ -14855,7 +15167,7 @@ "outputs": [] }, { - "name": "aten::sort.dimname", + "name": "aten::sort.dimname(Tensor self, Dimname dim, bool descending=False) -> (Tensor values, Tensor indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -14867,7 +15179,7 @@ ] }, { - "name": "aten::sort.dimname_stable", + "name": "aten::sort.dimname_stable(Tensor self, *, bool? stable, Dimname dim, bool descending=False) -> (Tensor values, Tensor indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "stable", "type": "boolean?", "kwarg_only": true }, @@ -14880,7 +15192,7 @@ ] }, { - "name": "aten::sort.dimname_values", + "name": "aten::sort.dimname_values(Tensor self, Dimname dim, bool descending=False, *, Tensor(a!) values, Tensor(b!) indices) -> (Tensor(a!) values, Tensor(b!) indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" }, @@ -14892,7 +15204,7 @@ ] }, { - "name": "aten::sort.dimname_values_stable", + "name": "aten::sort.dimname_values_stable(Tensor self, *, bool? stable, Dimname dim, bool descending=False, Tensor(a!) values, Tensor(b!) indices) -> (Tensor(a!) values, Tensor(b!) indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "stable", "type": "boolean?", "kwarg_only": true }, @@ -14905,7 +15217,7 @@ ] }, { - "name": "aten::sort.float", + "name": "aten::sort.float(float[](a!) self, bool reverse=False) -> ()", "inputs": [ { "name": "self", "type": "float32[]" }, { "name": "reverse", "type": "boolean", "default": false } @@ -14913,7 +15225,7 @@ "outputs": [] }, { - "name": "aten::sort.int", + "name": "aten::sort.int(int[](a!) self, bool reverse=False) -> ()", "inputs": [ { "name": "self", "type": "int64[]" }, { "name": "reverse", "type": "boolean", "default": false } @@ -14921,7 +15233,7 @@ "outputs": [] }, { - "name": "aten::sort.stable", + "name": "aten::sort.stable(Tensor self, *, bool? stable, int dim=-1, bool descending=False) -> (Tensor values, Tensor indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "stable", "type": "boolean?", "kwarg_only": true }, @@ -14934,7 +15246,7 @@ ] }, { - "name": "aten::sort.str", + "name": "aten::sort.str(str[](a!) self, bool reverse=False) -> ()", "inputs": [ { "name": "self", "type": "string[]" }, { "name": "reverse", "type": "boolean", "default": false } @@ -14942,7 +15254,7 @@ "outputs": [] }, { - "name": "aten::sort.values", + "name": "aten::sort.values(Tensor self, int dim=-1, bool descending=False, *, Tensor(a!) values, Tensor(b!) indices) -> (Tensor(a!) values, Tensor(b!) indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64", "default": -1 }, @@ -14954,7 +15266,7 @@ ] }, { - "name": "aten::sort.values_stable", + "name": "aten::sort.values_stable(Tensor self, *, bool? stable, int dim=-1, bool descending=False, Tensor(a!) values, Tensor(b!) indices) -> (Tensor(a!) values, Tensor(b!) indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "stable", "type": "boolean?", "kwarg_only": true }, @@ -14967,7 +15279,7 @@ ] }, { - "name": "aten::special_expit", + "name": "aten::special_expit(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -14976,7 +15288,7 @@ ] }, { - "name": "aten::special_expit.out", + "name": "aten::special_expit.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -14985,7 +15297,7 @@ ] }, { - "name": "aten::split", + "name": "aten::split(Tensor(a -> *) self, int[] split_sizes, int dim=0) -> Tensor(a)[]", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "split_sizes", "type": "int64[]" }, @@ -14996,7 +15308,7 @@ ] }, { - "name": "aten::split.Tensor", + "name": "aten::split.Tensor(Tensor(a -> *) self, SymInt split_size, int dim=0) -> Tensor(a)[]", "category": "Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -15008,7 +15320,7 @@ ] }, { - "name": "aten::split.sizes", + "name": "aten::split.sizes(Tensor(a -> *) self, SymInt[] split_size, int dim=0) -> Tensor(a)[]", "category": "Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -15020,7 +15332,7 @@ ] }, { - "name": "aten::split.str", + "name": "aten::split.str(str self, str? separator=None, int max=-1) -> str[]", "inputs": [ { "name": "self", "type": "string" }, { "name": "separator", "type": "string?", "default": null }, @@ -15031,7 +15343,7 @@ ] }, { - "name": "aten::split_copy.Tensor", + "name": "aten::split_copy.Tensor(Tensor self, SymInt split_size, int dim=0) -> Tensor[]", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "split_size", "type": "SymInt" }, @@ -15042,7 +15354,7 @@ ] }, { - "name": "aten::split_copy.Tensor_out", + "name": "aten::split_copy.Tensor_out(Tensor self, SymInt split_size, int dim=0, *, Tensor(a!)[] out) -> ()", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "split_size", "type": "SymInt" }, @@ -15051,7 +15363,7 @@ "outputs": [] }, { - "name": "aten::split_with_sizes", + "name": "aten::split_with_sizes(Tensor(a -> *) self, SymInt[] split_sizes, int dim=0) -> Tensor(a)[]", "category": "Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -15063,7 +15375,7 @@ ] }, { - "name": "aten::split_with_sizes_copy", + "name": "aten::split_with_sizes_copy(Tensor self, SymInt[] split_sizes, int dim=0) -> Tensor[]", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "split_sizes", "type": "SymInt[]" }, @@ -15074,7 +15386,7 @@ ] }, { - "name": "aten::split_with_sizes_copy.out", + "name": "aten::split_with_sizes_copy.out(Tensor self, SymInt[] split_sizes, int dim=0, *, Tensor(a!)[] out) -> ()", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "split_sizes", "type": "SymInt[]" }, @@ -15083,7 +15395,7 @@ "outputs": [] }, { - "name": "aten::splitlines", + "name": "aten::splitlines(str self, bool keepends=False) -> str[]", "inputs": [ { "name": "self", "type": "string" }, { "name": "keepends", "type": "boolean", "default": false } @@ -15093,7 +15405,7 @@ ] }, { - "name": "aten::sqrt", + "name": "aten::sqrt(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -15102,7 +15414,7 @@ ] }, { - "name": "aten::sqrt.Scalar", + "name": "aten::sqrt.Scalar(Scalar a) -> Scalar", "inputs": [ { "name": "a", "type": "Scalar" } ], @@ -15111,7 +15423,7 @@ ] }, { - "name": "aten::sqrt.complex", + "name": "aten::sqrt.complex(complex a) -> complex", "inputs": [ { "name": "a", "type": "complex" } ], @@ -15120,7 +15432,7 @@ ] }, { - "name": "aten::sqrt.float", + "name": "aten::sqrt.float(float a) -> float", "inputs": [ { "name": "a", "type": "float32" } ], @@ -15129,7 +15441,7 @@ ] }, { - "name": "aten::sqrt.int", + "name": "aten::sqrt.int(int a) -> float", "inputs": [ { "name": "a", "type": "int64" } ], @@ -15138,7 +15450,7 @@ ] }, { - "name": "aten::sqrt.out", + "name": "aten::sqrt.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -15147,7 +15459,7 @@ ] }, { - "name": "aten::sqrt_", + "name": "aten::sqrt_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -15156,7 +15468,7 @@ ] }, { - "name": "aten::square", + "name": "aten::square(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -15165,7 +15477,7 @@ ] }, { - "name": "aten::square.out", + "name": "aten::square.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -15174,7 +15486,7 @@ ] }, { - "name": "aten::square_", + "name": "aten::square_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -15183,7 +15495,7 @@ ] }, { - "name": "aten::squeeze", + "name": "aten::squeeze(Tensor(a) self) -> Tensor(a)", "category": "Transform", "inputs": [ { "name": "self", "type": "Tensor" } @@ -15193,7 +15505,7 @@ ] }, { - "name": "aten::squeeze.dim", + "name": "aten::squeeze.dim(Tensor(a) self, int dim) -> Tensor(a)", "category": "Transform", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -15204,7 +15516,7 @@ ] }, { - "name": "aten::squeeze.dimname", + "name": "aten::squeeze.dimname(Tensor(a) self, Dimname dim) -> Tensor(a)", "category": "Transform", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -15215,7 +15527,7 @@ ] }, { - "name": "aten::squeeze.dims", + "name": "aten::squeeze.dims(Tensor(a) self, int[] dim) -> Tensor(a)", "category": "Transform", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -15226,7 +15538,7 @@ ] }, { - "name": "aten::squeeze_", + "name": "aten::squeeze_(Tensor(a!) self) -> Tensor(a!)", "category": "Transform", "inputs": [ { "name": "self", "type": "Tensor" } @@ -15236,7 +15548,7 @@ ] }, { - "name": "aten::squeeze_.dim", + "name": "aten::squeeze_.dim(Tensor(a!) self, int dim) -> Tensor(a!)", "category": "Transform", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -15247,7 +15559,7 @@ ] }, { - "name": "aten::squeeze_.dimname", + "name": "aten::squeeze_.dimname(Tensor(a!) self, Dimname dim) -> Tensor(a!)", "category": "Transform", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -15258,7 +15570,7 @@ ] }, { - "name": "aten::squeeze_.dims", + "name": "aten::squeeze_.dims(Tensor(a!) self, int[] dim) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[]" } @@ -15268,7 +15580,7 @@ ] }, { - "name": "aten::stack", + "name": "aten::stack(Tensor[] tensors, int dim=0) -> Tensor", "category": "Tensor", "inputs": [ { "name": "tensors", "type": "Tensor[]" }, @@ -15279,7 +15591,7 @@ ] }, { - "name": "aten::stack.out", + "name": "aten::stack.out(Tensor[] tensors, int dim=0, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "tensors", "type": "Tensor[]" }, { "name": "dim", "type": "int64", "default": 0 } @@ -15289,7 +15601,7 @@ ] }, { - "name": "aten::std", + "name": "aten::std(Tensor self, bool unbiased=True) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "unbiased", "type": "boolean", "default": true } @@ -15299,7 +15611,7 @@ ] }, { - "name": "aten::std.correction", + "name": "aten::std.correction(Tensor self, int[1]? dim=None, *, Scalar? correction=None, bool keepdim=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]?", "default": null }, @@ -15311,7 +15623,7 @@ ] }, { - "name": "aten::std.correction_names", + "name": "aten::std.correction_names(Tensor self, Dimname[1] dim, *, Scalar? correction=None, bool keepdim=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname[1]" }, @@ -15323,7 +15635,7 @@ ] }, { - "name": "aten::std.correction_names_out", + "name": "aten::std.correction_names_out(Tensor self, Dimname[1] dim, *, Scalar? correction=None, bool keepdim=False, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname[1]" }, @@ -15335,7 +15647,7 @@ ] }, { - "name": "aten::std.correction_out", + "name": "aten::std.correction_out(Tensor self, int[1]? dim=None, *, Scalar? correction=None, bool keepdim=False, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]?", "default": null }, @@ -15347,7 +15659,7 @@ ] }, { - "name": "aten::std.dim", + "name": "aten::std.dim(Tensor self, int[1]? dim, bool unbiased=True, bool keepdim=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]?" }, @@ -15359,7 +15671,7 @@ ] }, { - "name": "aten::std.names_dim", + "name": "aten::std.names_dim(Tensor self, Dimname[1] dim, bool unbiased=True, bool keepdim=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname[1]" }, @@ -15371,7 +15683,7 @@ ] }, { - "name": "aten::std.names_out", + "name": "aten::std.names_out(Tensor self, Dimname[1] dim, bool unbiased=True, bool keepdim=False, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname[1]" }, @@ -15383,7 +15695,7 @@ ] }, { - "name": "aten::std.out", + "name": "aten::std.out(Tensor self, int[1]? dim, bool unbiased=True, bool keepdim=False, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]?" }, @@ -15395,7 +15707,7 @@ ] }, { - "name": "aten::std_mean", + "name": "aten::std_mean(Tensor self, bool unbiased=True) -> (Tensor, Tensor)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "unbiased", "type": "boolean", "default": true } @@ -15406,7 +15718,7 @@ ] }, { - "name": "aten::std_mean.correction", + "name": "aten::std_mean.correction(Tensor self, int[1]? dim=None, *, Scalar? correction=None, bool keepdim=False) -> (Tensor, Tensor)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]?", "default": null }, @@ -15419,7 +15731,7 @@ ] }, { - "name": "aten::std_mean.correction_names", + "name": "aten::std_mean.correction_names(Tensor self, Dimname[1] dim, *, Scalar? correction=None, bool keepdim=False) -> (Tensor, Tensor)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname[1]" }, @@ -15432,7 +15744,7 @@ ] }, { - "name": "aten::std_mean.dim", + "name": "aten::std_mean.dim(Tensor self, int[1]? dim, bool unbiased=True, bool keepdim=False) -> (Tensor, Tensor)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]?" }, @@ -15445,7 +15757,7 @@ ] }, { - "name": "aten::std_mean.names_dim", + "name": "aten::std_mean.names_dim(Tensor self, Dimname[1] dim, bool unbiased=True, bool keepdim=False) -> (Tensor, Tensor)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname[1]" }, @@ -15458,7 +15770,7 @@ ] }, { - "name": "aten::stft", + "name": "aten::stft(Tensor self, int n_fft, int? hop_length=None, int? win_length=None, Tensor? window=None, bool normalized=False, bool? onesided=None, bool? return_complex=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "n_fft", "type": "int64" }, @@ -15474,7 +15786,7 @@ ] }, { - "name": "aten::stft.center", + "name": "aten::stft.center(Tensor self, int n_fft, int? hop_length=None, int? win_length=None, Tensor? window=None, bool center=True, str pad_mode=\"reflect\", bool normalized=False, bool? onesided=None, bool? return_complex=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "n_fft", "type": "int64" }, @@ -15492,7 +15804,7 @@ ] }, { - "name": "aten::str", + "name": "aten::str(t elem) -> str", "inputs": [ { "name": "elem", "type": "t" } ], @@ -15501,7 +15813,7 @@ ] }, { - "name": "aten::stride", + "name": "aten::stride(Tensor self) -> int[]", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -15510,7 +15822,7 @@ ] }, { - "name": "aten::stride.Dimname", + "name": "aten::stride.Dimname(Tensor self, Dimname dim) -> int", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" } @@ -15520,7 +15832,7 @@ ] }, { - "name": "aten::stride.int", + "name": "aten::stride.int(Tensor self, int dim) -> int", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" } @@ -15530,7 +15842,7 @@ ] }, { - "name": "aten::strip", + "name": "aten::strip(str self, str chars=' \\\\n\\\\t\\\\f\\\\v') -> str", "inputs": [ { "name": "self", "type": "string" }, { "name": "chars", "type": "string", "default": " \\\\n\\\\t\\\\f\\\\v" } @@ -15540,7 +15852,17 @@ ] }, { - "name": "aten::sub.Scalar", + "name": "aten::sub(Scalar a, Scalar b) -> Scalar", + "inputs": [ + { "name": "a", "type": "Scalar" }, + { "name": "b", "type": "Scalar" } + ], + "outputs": [ + { "type": "Scalar" } + ] + }, + { + "name": "aten::sub.Scalar(Tensor self, Scalar other, Scalar alpha=1) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" }, @@ -15551,7 +15873,18 @@ ] }, { - "name": "aten::sub.Tensor", + "name": "aten::sub.Scalar_out(Tensor self, Scalar other, Scalar alpha=1, *, Tensor(a!) out) -> Tensor(a!)", + "inputs": [ + { "name": "self", "type": "Tensor" }, + { "name": "other", "type": "Scalar" }, + { "name": "alpha", "type": "Scalar", "default": 1 } + ], + "outputs": [ + { "type": "Tensor" } + ] + }, + { + "name": "aten::sub.Tensor(Tensor self, Tensor other, *, Scalar alpha=1) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -15562,7 +15895,37 @@ ] }, { - "name": "aten::sub.float", + "name": "aten::sub.complex(complex a, complex b) -> complex", + "inputs": [ + { "name": "a", "type": "complex" }, + { "name": "b", "type": "complex" } + ], + "outputs": [ + { "type": "complex" } + ] + }, + { + "name": "aten::sub.complex_float(complex a, float b) -> complex", + "inputs": [ + { "name": "a", "type": "complex" }, + { "name": "b", "type": "float32" } + ], + "outputs": [ + { "type": "complex" } + ] + }, + { + "name": "aten::sub.complex_int(complex a, int b) -> complex", + "inputs": [ + { "name": "a", "type": "complex" }, + { "name": "b", "type": "int64" } + ], + "outputs": [ + { "type": "complex" } + ] + }, + { + "name": "aten::sub.float(float a, float b) -> float", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "float32" } @@ -15572,7 +15935,27 @@ ] }, { - "name": "aten::sub.int", + "name": "aten::sub.float_complex(float a, complex b) -> complex", + "inputs": [ + { "name": "a", "type": "float32" }, + { "name": "b", "type": "complex" } + ], + "outputs": [ + { "type": "complex" } + ] + }, + { + "name": "aten::sub.float_int(float a, int b) -> float", + "inputs": [ + { "name": "a", "type": "float32" }, + { "name": "b", "type": "int64" } + ], + "outputs": [ + { "type": "float32" } + ] + }, + { + "name": "aten::sub.int(int a, int b) -> int", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "int64" } @@ -15582,28 +15965,38 @@ ] }, { - "name": "aten::sub.out", + "name": "aten::sub.int_complex(int a, complex b) -> complex", "inputs": [ - { "name": "self", "type": "Tensor" }, - { "name": "other", "type": "Tensor" }, - { "name": "alpha", "type": "Scalar", "default": 1, "kwarg_only": true } + { "name": "a", "type": "int64" }, + { "name": "b", "type": "complex" } ], "outputs": [ - { "type": "Tensor" } + { "type": "complex" } ] }, { - "name": "aten::sub.str", + "name": "aten::sub.int_float(int a, float b) -> float", "inputs": [ - { "name": "a", "type": "string" }, - { "name": "b", "type": "string" } + { "name": "a", "type": "int64" }, + { "name": "b", "type": "float32" } + ], + "outputs": [ + { "type": "float32" } + ] + }, + { + "name": "aten::sub.out(Tensor self, Tensor other, *, Scalar alpha=1, Tensor(a!) out) -> Tensor(a!)", + "inputs": [ + { "name": "self", "type": "Tensor" }, + { "name": "other", "type": "Tensor" }, + { "name": "alpha", "type": "Scalar", "default": 1, "kwarg_only": true } ], "outputs": [ { "type": "Tensor" } ] }, { - "name": "aten::sub_.Scalar", + "name": "aten::sub_.Scalar(Tensor(a!) self, Scalar other, Scalar alpha=1) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" }, @@ -15614,7 +16007,7 @@ ] }, { - "name": "aten::sub_.Tensor", + "name": "aten::sub_.Tensor(Tensor(a!) self, Tensor other, *, Scalar alpha=1) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -15625,7 +16018,7 @@ ] }, { - "name": "aten::sum", + "name": "aten::sum(Tensor self, *, ScalarType? dtype=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true } @@ -15635,10 +16028,10 @@ ] }, { - "name": "aten::sum.DimnameList_out", + "name": "aten::sum.DimnameList_out(Tensor self, str[1] dim, bool keepdim=False, *, ScalarType? dtype=None, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, - { "name": "dim", "type": "Dimname[1]" }, + { "name": "dim", "type": "string[1]" }, { "name": "keepdim", "type": "boolean", "default": false }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true } ], @@ -15647,7 +16040,7 @@ ] }, { - "name": "aten::sum.IntList_out", + "name": "aten::sum.IntList_out(Tensor self, int[1]? dim, bool keepdim=False, *, ScalarType? dtype=None, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]?" }, @@ -15659,7 +16052,7 @@ ] }, { - "name": "aten::sum.bool", + "name": "aten::sum.bool(bool[] self) -> int", "inputs": [ { "name": "self", "type": "boolean[]" } ], @@ -15668,7 +16061,7 @@ ] }, { - "name": "aten::sum.complex", + "name": "aten::sum.complex(complex[] self) -> complex", "inputs": [ { "name": "self", "type": "complex[]" } ], @@ -15677,10 +16070,10 @@ ] }, { - "name": "aten::sum.dim_DimnameList", + "name": "aten::sum.dim_DimnameList(Tensor self, str[1] dim, bool keepdim=False, *, ScalarType? dtype=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, - { "name": "dim", "type": "Dimname[1]" }, + { "name": "dim", "type": "string[1]" }, { "name": "keepdim", "type": "boolean", "default": false }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true } ], @@ -15689,7 +16082,7 @@ ] }, { - "name": "aten::sum.dim_IntList", + "name": "aten::sum.dim_IntList(Tensor self, int[1]? dim, bool keepdim=False, *, ScalarType? dtype=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]?" }, @@ -15701,7 +16094,7 @@ ] }, { - "name": "aten::sum.float", + "name": "aten::sum.float(float[] self) -> float", "inputs": [ { "name": "self", "type": "float32[]" } ], @@ -15710,7 +16103,7 @@ ] }, { - "name": "aten::sum.int", + "name": "aten::sum.int(int[] self) -> int", "inputs": [ { "name": "self", "type": "int64[]" } ], @@ -15719,7 +16112,17 @@ ] }, { - "name": "aten::swapaxes", + "name": "aten::sum.out(Tensor self, *, ScalarType? dtype=None, Tensor(a!) out) -> Tensor(a!)", + "inputs": [ + { "name": "self", "type": "Tensor" }, + { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true } + ], + "outputs": [ + { "type": "Tensor" } + ] + }, + { + "name": "aten::swapaxes(Tensor(a) self, int axis0, int axis1) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "axis0", "type": "int64" }, @@ -15730,7 +16133,7 @@ ] }, { - "name": "aten::swapaxes_", + "name": "aten::swapaxes_(Tensor(a!) self, int axis0, int axis1) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "axis0", "type": "int64" }, @@ -15741,7 +16144,7 @@ ] }, { - "name": "aten::sym_size", + "name": "aten::sym_size(Tensor self) -> SymInt[]", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -15750,7 +16153,7 @@ ] }, { - "name": "aten::sym_size.int", + "name": "aten::sym_size.int(Tensor self, int dim) -> SymInt", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" } @@ -15760,7 +16163,7 @@ ] }, { - "name": "aten::t", + "name": "aten::t(Tensor(a) self) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -15769,7 +16172,7 @@ ] }, { - "name": "aten::take", + "name": "aten::take(Tensor self, Tensor index) -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -15780,7 +16183,7 @@ ] }, { - "name": "aten::take.out", + "name": "aten::take.out(Tensor self, Tensor index, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "index", "type": "Tensor" } @@ -15790,7 +16193,7 @@ ] }, { - "name": "aten::take_along_dim", + "name": "aten::take_along_dim(Tensor self, Tensor indices, int? dim=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "indices", "type": "Tensor" }, @@ -15801,7 +16204,7 @@ ] }, { - "name": "aten::take_along_dim.out", + "name": "aten::take_along_dim.out(Tensor self, Tensor indices, int? dim=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "indices", "type": "Tensor" }, @@ -15812,7 +16215,7 @@ ] }, { - "name": "aten::tan", + "name": "aten::tan(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -15821,7 +16224,7 @@ ] }, { - "name": "aten::tan.out", + "name": "aten::tan.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -15830,7 +16233,7 @@ ] }, { - "name": "aten::tan_", + "name": "aten::tan_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -15839,7 +16242,7 @@ ] }, { - "name": "aten::tanh", + "name": "aten::tanh(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -15848,7 +16251,7 @@ ] }, { - "name": "aten::tanh.out", + "name": "aten::tanh.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -15857,7 +16260,7 @@ ] }, { - "name": "aten::tanh_", + "name": "aten::tanh_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -15866,7 +16269,7 @@ ] }, { - "name": "aten::tensor", + "name": "aten::tensor(t[] data, *, ScalarType? dtype=None, Device? device=None, bool requires_grad=False) -> Tensor", "inputs": [ { "name": "data", "type": "t[]" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -15878,11 +16281,11 @@ ] }, { - "name": "aten::tensor.bool", + "name": "aten::tensor.bool(bool t, *, ScalarType? dtype=None, Device? device=None, bool requires_grad=False) -> Tensor", "inputs": [ { "name": "t", "type": "boolean" }, - { "name": "dtype", "type": "ScalarType", "optional": true, "default": null, "kwarg_only": true }, - { "name": "device", "type": "Device", "optional": true, "default": null, "kwarg_only": true }, + { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, + { "name": "device", "type": "Device?", "default": null, "kwarg_only": true }, { "name": "requires_grad", "type": "boolean", "default": false, "kwarg_only": true } ], "outputs": [ @@ -15890,11 +16293,23 @@ ] }, { - "name": "aten::tensor.float", + "name": "aten::tensor.complex(complex t, *, ScalarType? dtype=None, Device? device=None, bool requires_grad=False) -> Tensor", + "inputs": [ + { "name": "t", "type": "complex" }, + { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, + { "name": "device", "type": "Device?", "default": null, "kwarg_only": true }, + { "name": "requires_grad", "type": "boolean", "default": false, "kwarg_only": true } + ], + "outputs": [ + { "type": "Tensor" } + ] + }, + { + "name": "aten::tensor.float(float t, *, ScalarType? dtype=None, Device? device=None, bool requires_grad=False) -> Tensor", "inputs": [ { "name": "t", "type": "float32" }, - { "name": "dtype", "type": "ScalarType", "optional": true, "default": null, "kwarg_only": true }, - { "name": "device", "type": "Device", "optional": true, "default": null, "kwarg_only": true }, + { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, + { "name": "device", "type": "Device?", "default": null, "kwarg_only": true }, { "name": "requires_grad", "type": "boolean", "default": false, "kwarg_only": true } ], "outputs": [ @@ -15902,11 +16317,11 @@ ] }, { - "name": "aten::tensor.int", + "name": "aten::tensor.int(int t, *, ScalarType? dtype=None, Device? device=None, bool requires_grad=False) -> Tensor", "inputs": [ { "name": "t", "type": "int64" }, - { "name": "dtype", "type": "ScalarType", "optional": true, "default": null, "kwarg_only": true }, - { "name": "device", "type": "Device", "optional": true, "default": null, "kwarg_only": true }, + { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, + { "name": "device", "type": "Device?", "default": null, "kwarg_only": true }, { "name": "requires_grad", "type": "boolean", "default": false, "kwarg_only": true } ], "outputs": [ @@ -15914,7 +16329,7 @@ ] }, { - "name": "aten::tensor_split.indices", + "name": "aten::tensor_split.indices(Tensor(a -> *) self, SymInt[] indices, int dim=0) -> Tensor(a)[]", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "indices", "type": "SymInt[]" }, @@ -15925,7 +16340,7 @@ ] }, { - "name": "aten::tensor_split.sections", + "name": "aten::tensor_split.sections(Tensor(a -> *) self, SymInt sections, int dim=0) -> Tensor(a)[]", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "sections", "type": "SymInt" }, @@ -15936,7 +16351,7 @@ ] }, { - "name": "aten::tensor_split.tensor_indices_or_sections", + "name": "aten::tensor_split.tensor_indices_or_sections(Tensor(a -> *) self, Tensor tensor_indices_or_sections, int dim=0) -> Tensor(a)[]", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "tensor_indices_or_sections", "type": "Tensor" }, @@ -15947,7 +16362,7 @@ ] }, { - "name": "aten::tensordot", + "name": "aten::tensordot(Tensor self, Tensor other, int[] dims_self, int[] dims_other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -15959,7 +16374,7 @@ ] }, { - "name": "aten::tensordot.out", + "name": "aten::tensordot.out(Tensor self, Tensor other, int[] dims_self, int[] dims_other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -15971,7 +16386,7 @@ ] }, { - "name": "aten::threshold", + "name": "aten::threshold(Tensor self, Scalar threshold, Scalar value) -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -15983,7 +16398,7 @@ ] }, { - "name": "aten::threshold.out", + "name": "aten::threshold.out(Tensor self, Scalar threshold, Scalar value, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "threshold", "type": "Scalar" }, @@ -15994,7 +16409,7 @@ ] }, { - "name": "aten::threshold_", + "name": "aten::threshold_(Tensor(a!) self, Scalar threshold, Scalar value) -> Tensor(a!)", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -16006,7 +16421,7 @@ ] }, { - "name": "aten::tile", + "name": "aten::tile(Tensor self, SymInt[] dims) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dims", "type": "SymInt[]" } @@ -16016,7 +16431,7 @@ ] }, { - "name": "aten::to.device", + "name": "aten::to.device(Tensor(a) self, Device device, ScalarType dtype, bool non_blocking=False, bool copy=False, MemoryFormat? memory_format=None) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "device", "type": "Device" }, @@ -16030,7 +16445,7 @@ ] }, { - "name": "aten::to.dtype", + "name": "aten::to.dtype(Tensor(a) self, ScalarType dtype, bool non_blocking=False, bool copy=False, MemoryFormat? memory_format=None) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dtype", "type": "ScalarType" }, @@ -16043,7 +16458,7 @@ ] }, { - "name": "aten::to.dtype_layout", + "name": "aten::to.dtype_layout(Tensor(a) self, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None, bool non_blocking=False, bool copy=False, MemoryFormat? memory_format=None) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -16059,7 +16474,7 @@ ] }, { - "name": "aten::to.other", + "name": "aten::to.other(Tensor(a) self, Tensor other, bool non_blocking=False, bool copy=False, MemoryFormat? memory_format=None) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" }, @@ -16072,7 +16487,7 @@ ] }, { - "name": "aten::to.prim_Device", + "name": "aten::to.prim_Device(Tensor(a) self, Device? device, int? dtype=None, bool non_blocking=False, bool copy=False) -> Tensor(a|b)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "device", "type": "Device?" }, @@ -16085,7 +16500,7 @@ ] }, { - "name": "aten::to.prim_dtype", + "name": "aten::to.prim_dtype(Tensor(a) self, int? dtype=None, bool non_blocking=False, bool copy=False) -> Tensor(a|b)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dtype", "type": "int64?", "default": null }, @@ -16097,7 +16512,7 @@ ] }, { - "name": "aten::to.prim_other", + "name": "aten::to.prim_other(Tensor(a) self, bool non_blocking=False, bool copy=False) -> Tensor(a|b)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "non_blocking", "type": "boolean", "default": false }, @@ -16108,7 +16523,7 @@ ] }, { - "name": "aten::to_dense", + "name": "aten::to_dense(Tensor self, ScalarType? dtype=None, *, bool? masked_grad=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dtype", "type": "ScalarType?", "default": null }, @@ -16119,7 +16534,7 @@ ] }, { - "name": "aten::to_dense_backward", + "name": "aten::to_dense_backward(Tensor grad, Tensor input, bool? masked_grad=None) -> Tensor", "inputs": [ { "name": "grad", "type": "Tensor" }, { "name": "input", "type": "Tensor" }, @@ -16130,7 +16545,7 @@ ] }, { - "name": "aten::to_mkldnn", + "name": "aten::to_mkldnn(Tensor self, ScalarType? dtype=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dtype", "type": "ScalarType?", "default": null } @@ -16140,7 +16555,7 @@ ] }, { - "name": "aten::to_mkldnn_backward", + "name": "aten::to_mkldnn_backward(Tensor grad, Tensor input) -> Tensor", "inputs": [ { "name": "grad", "type": "Tensor" }, { "name": "input", "type": "Tensor" } @@ -16150,7 +16565,7 @@ ] }, { - "name": "aten::to_padded_tensor", + "name": "aten::to_padded_tensor(Tensor self, float padding, SymInt[]? output_size=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "padding", "type": "float32" }, @@ -16161,7 +16576,7 @@ ] }, { - "name": "aten::to_sparse", + "name": "aten::to_sparse(Tensor self, *, Layout? layout=None, int[2]? blocksize=None, int? dense_dim=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "layout", "type": "Layout?", "default": null, "kwarg_only": true }, @@ -16173,7 +16588,7 @@ ] }, { - "name": "aten::to_sparse.sparse_dim", + "name": "aten::to_sparse.sparse_dim(Tensor self, int sparse_dim) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "sparse_dim", "type": "int64" } @@ -16183,7 +16598,7 @@ ] }, { - "name": "aten::to_sparse_bsc", + "name": "aten::to_sparse_bsc(Tensor self, int[2] blocksize, int? dense_dim=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "blocksize", "type": "int64[2]" }, @@ -16194,7 +16609,7 @@ ] }, { - "name": "aten::to_sparse_bsr", + "name": "aten::to_sparse_bsr(Tensor self, int[2] blocksize, int? dense_dim=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "blocksize", "type": "int64[2]" }, @@ -16205,7 +16620,7 @@ ] }, { - "name": "aten::to_sparse_csc", + "name": "aten::to_sparse_csc(Tensor self, int? dense_dim=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dense_dim", "type": "int64?", "default": null } @@ -16215,7 +16630,7 @@ ] }, { - "name": "aten::to_sparse_csr", + "name": "aten::to_sparse_csr(Tensor self, int? dense_dim=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dense_dim", "type": "int64?", "default": null } @@ -16225,7 +16640,7 @@ ] }, { - "name": "aten::topk", + "name": "aten::topk(Tensor self, SymInt k, int dim=-1, bool largest=True, bool sorted=True) -> (Tensor values, Tensor indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "k", "type": "SymInt" }, @@ -16239,7 +16654,7 @@ ] }, { - "name": "aten::topk.values", + "name": "aten::topk.values(Tensor self, SymInt k, int dim=-1, bool largest=True, bool sorted=True, *, Tensor(a!) values, Tensor(b!) indices) -> (Tensor(a!) values, Tensor(b!) indices)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "k", "type": "SymInt" }, @@ -16253,7 +16668,7 @@ ] }, { - "name": "aten::transpose.Dimname", + "name": "aten::transpose.Dimname(Tensor(a) self, Dimname dim0, Dimname dim1) -> Tensor(a)", "category": "Transform", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -16265,7 +16680,7 @@ ] }, { - "name": "aten::transpose.int", + "name": "aten::transpose.int(Tensor(a) self, int dim0, int dim1) -> Tensor(a)", "category": "Transform", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -16277,7 +16692,7 @@ ] }, { - "name": "aten::transpose_", + "name": "aten::transpose_(Tensor(a!) self, int dim0, int dim1) -> Tensor(a!)", "category": "Transform", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -16289,7 +16704,7 @@ ] }, { - "name": "aten::transpose_copy.int", + "name": "aten::transpose_copy.int(Tensor self, int dim0, int dim1) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim0", "type": "int64" }, @@ -16300,7 +16715,7 @@ ] }, { - "name": "aten::tril", + "name": "aten::tril(Tensor self, int diagonal=0) -> Tensor", "category": "Layer", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -16311,7 +16726,7 @@ ] }, { - "name": "aten::tril.out", + "name": "aten::tril.out(Tensor self, int diagonal=0, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "diagonal", "type": "int64", "default": 0 } @@ -16321,7 +16736,7 @@ ] }, { - "name": "aten::tril_", + "name": "aten::tril_(Tensor(a!) self, int diagonal=0) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "diagonal", "type": "int64", "default": 0 } @@ -16331,7 +16746,7 @@ ] }, { - "name": "aten::tril_indices", + "name": "aten::tril_indices(int row, int col, int offset=0, *, ScalarType? dtype=long, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "category": "Layer", "inputs": [ { "name": "row", "type": "int64" }, @@ -16347,7 +16762,7 @@ ] }, { - "name": "aten::triu", + "name": "aten::triu(Tensor self, int diagonal=0) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "diagonal", "type": "int64", "default": 0 } @@ -16357,7 +16772,7 @@ ] }, { - "name": "aten::triu.out", + "name": "aten::triu.out(Tensor self, int diagonal=0, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "diagonal", "type": "int64", "default": 0 } @@ -16367,7 +16782,7 @@ ] }, { - "name": "aten::triu_", + "name": "aten::triu_(Tensor(a!) self, int diagonal=0) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "diagonal", "type": "int64", "default": 0 } @@ -16377,7 +16792,7 @@ ] }, { - "name": "aten::triu_indices", + "name": "aten::triu_indices(int row, int col, int offset=0, *, ScalarType? dtype=long, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "row", "type": "int64" }, { "name": "col", "type": "int64" }, @@ -16392,7 +16807,7 @@ ] }, { - "name": "aten::true_divide.Scalar", + "name": "aten::true_divide.Scalar(Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -16402,7 +16817,7 @@ ] }, { - "name": "aten::true_divide.Tensor", + "name": "aten::true_divide.Tensor(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -16412,7 +16827,7 @@ ] }, { - "name": "aten::true_divide.out", + "name": "aten::true_divide.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -16422,7 +16837,7 @@ ] }, { - "name": "aten::true_divide_.Scalar", + "name": "aten::true_divide_.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Scalar" } @@ -16432,7 +16847,7 @@ ] }, { - "name": "aten::true_divide_.Tensor", + "name": "aten::true_divide_.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -16442,7 +16857,7 @@ ] }, { - "name": "aten::type_as", + "name": "aten::type_as(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -16452,7 +16867,7 @@ ] }, { - "name": "aten::unbind.Dimname", + "name": "aten::unbind.Dimname(Tensor(a -> *) self, Dimname dim) -> Tensor(a)[]", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname" } @@ -16462,7 +16877,7 @@ ] }, { - "name": "aten::unbind.int", + "name": "aten::unbind.int(Tensor(a -> *) self, int dim=0) -> Tensor(a)[]", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64", "default": 0 } @@ -16472,7 +16887,7 @@ ] }, { - "name": "aten::unflatten.Dimname", + "name": "aten::unflatten.Dimname(Tensor(a) self, Dimname dim, SymInt[] sizes, Dimname[] names) -> Tensor(a)", "category": "Shape", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -16485,7 +16900,7 @@ ] }, { - "name": "aten::unflatten.int", + "name": "aten::unflatten.int(Tensor(a) self, int dim, SymInt[] sizes) -> Tensor(a)", "category": "Shape", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -16497,7 +16912,7 @@ ] }, { - "name": "aten::unfold", + "name": "aten::unfold(Tensor(a) self, int dimension, int size, int step) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dimension", "type": "int64" }, @@ -16509,7 +16924,7 @@ ] }, { - "name": "aten::uniform_", + "name": "aten::uniform_(Tensor(a!) self, float from=0, float to=1, *, Generator? generator=None) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "from", "type": "float32", "default": 0 }, @@ -16521,7 +16936,7 @@ ] }, { - "name": "aten::unique_consecutive", + "name": "aten::unique_consecutive(Tensor self, bool return_inverse=False, bool return_counts=False, int? dim=None) -> (Tensor, Tensor, Tensor)", "category": "Layer", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -16536,7 +16951,7 @@ ] }, { - "name": "aten::unique_dim", + "name": "aten::unique_dim(Tensor self, int dim, bool sorted=True, bool return_inverse=False, bool return_counts=False) -> (Tensor, Tensor, Tensor)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -16551,7 +16966,7 @@ ] }, { - "name": "aten::unique_dim_consecutive", + "name": "aten::unique_dim_consecutive(Tensor self, int dim, bool return_inverse=False, bool return_counts=False) -> (Tensor, Tensor, Tensor)", "category": "Layer", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -16566,7 +16981,7 @@ ] }, { - "name": "aten::unsafe_chunk", + "name": "aten::unsafe_chunk(Tensor self, int chunks, int dim=0) -> Tensor[]", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "chunks", "type": "int64" }, @@ -16577,7 +16992,7 @@ ] }, { - "name": "aten::unsafe_split.Tensor", + "name": "aten::unsafe_split.Tensor(Tensor self, SymInt split_size, int dim=0) -> Tensor[]", "category": "Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -16589,7 +17004,7 @@ ] }, { - "name": "aten::unsqueeze", + "name": "aten::unsqueeze(Tensor(a) self, int dim) -> Tensor(a)", "category": "Transform", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -16600,7 +17015,7 @@ ] }, { - "name": "aten::unsqueeze_", + "name": "aten::unsqueeze_(Tensor(a!) self, int dim) -> Tensor(a!)", "category": "Transform", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -16611,7 +17026,7 @@ ] }, { - "name": "aten::unsqueeze_copy", + "name": "aten::unsqueeze_copy(Tensor self, int dim) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64" } @@ -16621,7 +17036,7 @@ ] }, { - "name": "aten::upsample_bicubic2d", + "name": "aten::upsample_bicubic2d(Tensor self, SymInt[2] output_size, bool align_corners, float? scales_h=None, float? scales_w=None) -> Tensor", "category": "Layer", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -16635,7 +17050,7 @@ ] }, { - "name": "aten::upsample_bicubic2d.out", + "name": "aten::upsample_bicubic2d.out(Tensor self, SymInt[2] output_size, bool align_corners, float? scales_h=None, float? scales_w=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[2]" }, @@ -16648,7 +17063,7 @@ ] }, { - "name": "aten::upsample_bicubic2d.vec", + "name": "aten::upsample_bicubic2d.vec(Tensor input, SymInt[]? output_size, bool align_corners, float[]? scale_factors) -> Tensor", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -16661,7 +17076,7 @@ ] }, { - "name": "aten::upsample_bilinear2d", + "name": "aten::upsample_bilinear2d(Tensor self, SymInt[2] output_size, bool align_corners, float? scales_h=None, float? scales_w=None) -> Tensor", "category": "Layer", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -16675,7 +17090,7 @@ ] }, { - "name": "aten::upsample_bilinear2d.out", + "name": "aten::upsample_bilinear2d.out(Tensor self, SymInt[2] output_size, bool align_corners, float? scales_h=None, float? scales_w=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[2]" }, @@ -16688,7 +17103,7 @@ ] }, { - "name": "aten::upsample_bilinear2d.vec", + "name": "aten::upsample_bilinear2d.vec(Tensor input, SymInt[]? output_size, bool align_corners, float[]? scale_factors) -> Tensor", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -16701,7 +17116,7 @@ ] }, { - "name": "aten::upsample_bilinear2d_backward", + "name": "aten::upsample_bilinear2d_backward(Tensor grad_output, SymInt[2] output_size, SymInt[4] input_size, bool align_corners, float? scales_h=None, float? scales_w=None) -> Tensor", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[2]" }, @@ -16715,7 +17130,7 @@ ] }, { - "name": "aten::upsample_bilinear2d_backward.grad_input", + "name": "aten::upsample_bilinear2d_backward.grad_input(Tensor grad_output, SymInt[2] output_size, SymInt[4] input_size, bool align_corners, float? scales_h=None, float? scales_w=None, *, Tensor(a!) grad_input) -> Tensor(a!)", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[2]" }, @@ -16729,7 +17144,7 @@ ] }, { - "name": "aten::upsample_linear1d", + "name": "aten::upsample_linear1d(Tensor self, SymInt[1] output_size, bool align_corners, float? scales=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[1]" }, @@ -16741,7 +17156,7 @@ ] }, { - "name": "aten::upsample_linear1d.out", + "name": "aten::upsample_linear1d.out(Tensor self, SymInt[1] output_size, bool align_corners, float? scales=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[1]" }, @@ -16753,7 +17168,7 @@ ] }, { - "name": "aten::upsample_linear1d.vec", + "name": "aten::upsample_linear1d.vec(Tensor input, SymInt[]? output_size, bool align_corners, float[]? scale_factors) -> Tensor", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[]?" }, @@ -16765,7 +17180,7 @@ ] }, { - "name": "aten::upsample_nearest1d", + "name": "aten::upsample_nearest1d(Tensor self, SymInt[1] output_size, float? scales=None) -> Tensor", "category": "Layer", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -16777,7 +17192,7 @@ ] }, { - "name": "aten::upsample_nearest1d.out", + "name": "aten::upsample_nearest1d.out(Tensor self, SymInt[1] output_size, float? scales=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[1]" }, @@ -16788,7 +17203,7 @@ ] }, { - "name": "aten::upsample_nearest1d.vec", + "name": "aten::upsample_nearest1d.vec(Tensor input, SymInt[]? output_size, float[]? scale_factors) -> Tensor", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -16800,7 +17215,7 @@ ] }, { - "name": "aten::upsample_nearest2d", + "name": "aten::upsample_nearest2d(Tensor self, SymInt[2] output_size, float? scales_h=None, float? scales_w=None) -> Tensor", "category": "Layer", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -16813,7 +17228,7 @@ ] }, { - "name": "aten::upsample_nearest2d.out", + "name": "aten::upsample_nearest2d.out(Tensor self, SymInt[2] output_size, float? scales_h=None, float? scales_w=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[2]" }, @@ -16825,7 +17240,7 @@ ] }, { - "name": "aten::upsample_nearest2d.vec", + "name": "aten::upsample_nearest2d.vec(Tensor input, SymInt[]? output_size, float[]? scale_factors) -> Tensor", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -16837,7 +17252,7 @@ ] }, { - "name": "aten::upsample_nearest2d_backward", + "name": "aten::upsample_nearest2d_backward(Tensor grad_output, SymInt[2] output_size, SymInt[4] input_size, float? scales_h=None, float? scales_w=None) -> Tensor", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[2]" }, @@ -16850,7 +17265,7 @@ ] }, { - "name": "aten::upsample_nearest2d_backward.grad_input", + "name": "aten::upsample_nearest2d_backward.grad_input(Tensor grad_output, SymInt[2] output_size, SymInt[4] input_size, float? scales_h=None, float? scales_w=None, *, Tensor(a!) grad_input) -> Tensor(a!)", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[2]" }, @@ -16863,7 +17278,7 @@ ] }, { - "name": "aten::upsample_nearest3d", + "name": "aten::upsample_nearest3d(Tensor self, SymInt[3] output_size, float? scales_d=None, float? scales_h=None, float? scales_w=None) -> Tensor", "category": "Layer", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -16877,7 +17292,7 @@ ] }, { - "name": "aten::upsample_nearest3d.out", + "name": "aten::upsample_nearest3d.out(Tensor self, SymInt[3] output_size, float? scales_d=None, float? scales_h=None, float? scales_w=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[3]" }, @@ -16890,7 +17305,7 @@ ] }, { - "name": "aten::upsample_nearest3d.vec", + "name": "aten::upsample_nearest3d.vec(Tensor input, SymInt[]? output_size, float[]? scale_factors) -> Tensor", "category": "Layer", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -16902,7 +17317,7 @@ ] }, { - "name": "aten::upsample_trilinear3d", + "name": "aten::upsample_trilinear3d(Tensor self, SymInt[3] output_size, bool align_corners, float? scales_d=None, float? scales_h=None, float? scales_w=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[3]" }, @@ -16916,7 +17331,7 @@ ] }, { - "name": "aten::upsample_trilinear3d.out", + "name": "aten::upsample_trilinear3d.out(Tensor self, SymInt[3] output_size, bool align_corners, float? scales_d=None, float? scales_h=None, float? scales_w=None, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[3]" }, @@ -16930,7 +17345,7 @@ ] }, { - "name": "aten::upsample_trilinear3d.vec", + "name": "aten::upsample_trilinear3d.vec(Tensor input, SymInt[]? output_size, bool align_corners, float[]? scale_factors) -> Tensor", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[]?" }, @@ -16942,7 +17357,7 @@ ] }, { - "name": "aten::upsample_trilinear3d_backward", + "name": "aten::upsample_trilinear3d_backward(Tensor grad_output, SymInt[3] output_size, SymInt[5] input_size, bool align_corners, float? scales_d=None, float? scales_h=None, float? scales_w=None) -> Tensor", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[3]" }, @@ -16957,7 +17372,7 @@ ] }, { - "name": "aten::upsample_trilinear3d_backward.grad_input", + "name": "aten::upsample_trilinear3d_backward.grad_input(Tensor grad_output, SymInt[3] output_size, SymInt[5] input_size, bool align_corners, float? scales_d=None, float? scales_h=None, float? scales_w=None, *, Tensor(a!) grad_input) -> Tensor(a!)", "inputs": [ { "name": "grad_output", "type": "Tensor" }, { "name": "output_size", "type": "SymInt[3]" }, @@ -16972,7 +17387,7 @@ ] }, { - "name": "aten::values", + "name": "aten::values(Tensor(a) self) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -16981,7 +17396,7 @@ ] }, { - "name": "aten::values.Tensor", + "name": "aten::values.Tensor(Dict(Tensor, t) self) -> t[](*)", "inputs": [ { "name": "self", "type": "Dict(Tensor, t)" } ], @@ -16990,7 +17405,7 @@ ] }, { - "name": "aten::values.bool", + "name": "aten::values.bool(Dict(bool, t) self) -> t[](*)", "inputs": [ { "name": "self", "type": "Dict(boolean, t)" } ], @@ -16999,7 +17414,7 @@ ] }, { - "name": "aten::values.complex", + "name": "aten::values.complex(Dict(complex, t) self) -> t[](*)", "inputs": [ { "name": "self", "type": "Dict(complex, t)" } ], @@ -17008,7 +17423,7 @@ ] }, { - "name": "aten::values.float", + "name": "aten::values.float(Dict(float, t) self) -> t[](*)", "inputs": [ { "name": "self", "type": "Dict(float32, t)" } ], @@ -17017,7 +17432,7 @@ ] }, { - "name": "aten::values.int", + "name": "aten::values.int(Dict(int, t) self) -> t[](*)", "inputs": [ { "name": "self", "type": "Dict(int64, t)" } ], @@ -17026,7 +17441,7 @@ ] }, { - "name": "aten::values.str", + "name": "aten::values.str(Dict(str, t) self) -> t[](*)", "inputs": [ { "name": "self", "type": "Dict(string, t)" } ], @@ -17035,7 +17450,7 @@ ] }, { - "name": "aten::var", + "name": "aten::var(Tensor self, bool unbiased=True) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "unbiased", "type": "boolean", "default": true } @@ -17045,7 +17460,7 @@ ] }, { - "name": "aten::var.correction", + "name": "aten::var.correction(Tensor self, int[1]? dim=None, *, Scalar? correction=None, bool keepdim=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]?", "default": null }, @@ -17057,7 +17472,7 @@ ] }, { - "name": "aten::var.correction_names", + "name": "aten::var.correction_names(Tensor self, Dimname[1] dim, *, Scalar? correction=None, bool keepdim=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname[1]" }, @@ -17069,7 +17484,7 @@ ] }, { - "name": "aten::var.correction_names_out", + "name": "aten::var.correction_names_out(Tensor self, Dimname[1] dim, *, Scalar? correction=None, bool keepdim=False, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname[1]" }, @@ -17081,7 +17496,7 @@ ] }, { - "name": "aten::var.correction_out", + "name": "aten::var.correction_out(Tensor self, int[1]? dim=None, *, Scalar? correction=None, bool keepdim=False, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]?", "default": null }, @@ -17093,7 +17508,7 @@ ] }, { - "name": "aten::var.dim", + "name": "aten::var.dim(Tensor self, int[1]? dim, bool unbiased=True, bool keepdim=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]?" }, @@ -17105,7 +17520,7 @@ ] }, { - "name": "aten::var.names_dim", + "name": "aten::var.names_dim(Tensor self, Dimname[1] dim, bool unbiased=True, bool keepdim=False) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname[1]" }, @@ -17117,7 +17532,7 @@ ] }, { - "name": "aten::var.names_out", + "name": "aten::var.names_out(Tensor self, Dimname[1] dim, bool unbiased=True, bool keepdim=False, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname[1]" }, @@ -17129,7 +17544,7 @@ ] }, { - "name": "aten::var.out", + "name": "aten::var.out(Tensor self, int[1]? dim, bool unbiased=True, bool keepdim=False, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]?" }, @@ -17141,7 +17556,7 @@ ] }, { - "name": "aten::var_mean", + "name": "aten::var_mean(Tensor self, bool unbiased=True) -> (Tensor, Tensor)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "unbiased", "type": "boolean", "default": true } @@ -17152,7 +17567,7 @@ ] }, { - "name": "aten::var_mean.correction", + "name": "aten::var_mean.correction(Tensor self, int[1]? dim=None, *, Scalar? correction=None, bool keepdim=False) -> (Tensor, Tensor)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]?", "default": null }, @@ -17165,7 +17580,7 @@ ] }, { - "name": "aten::var_mean.correction_names", + "name": "aten::var_mean.correction_names(Tensor self, Dimname[1] dim, *, Scalar? correction=None, bool keepdim=False) -> (Tensor, Tensor)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname[1]" }, @@ -17178,7 +17593,7 @@ ] }, { - "name": "aten::var_mean.dim", + "name": "aten::var_mean.dim(Tensor self, int[1]? dim, bool unbiased=True, bool keepdim=False) -> (Tensor, Tensor)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "int64[1]?" }, @@ -17191,7 +17606,7 @@ ] }, { - "name": "aten::var_mean.names_dim", + "name": "aten::var_mean.names_dim(Tensor self, Dimname[1] dim, bool unbiased=True, bool keepdim=False) -> (Tensor, Tensor)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dim", "type": "Dimname[1]" }, @@ -17204,7 +17619,7 @@ ] }, { - "name": "aten::vdot", + "name": "aten::vdot(Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -17214,7 +17629,7 @@ ] }, { - "name": "aten::vdot.out", + "name": "aten::vdot.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -17224,7 +17639,7 @@ ] }, { - "name": "aten::view", + "name": "aten::view(Tensor(a) self, SymInt[] size) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "size", "type": "SymInt[]" } @@ -17234,7 +17649,7 @@ ] }, { - "name": "aten::view.dtype", + "name": "aten::view.dtype(Tensor(a) self, ScalarType dtype) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dtype", "type": "ScalarType" } @@ -17244,7 +17659,7 @@ ] }, { - "name": "aten::view_as", + "name": "aten::view_as(Tensor(a) self, Tensor other) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "other", "type": "Tensor" } @@ -17254,7 +17669,7 @@ ] }, { - "name": "aten::view_as_complex", + "name": "aten::view_as_complex(Tensor(a) self) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -17263,7 +17678,7 @@ ] }, { - "name": "aten::view_as_complex_copy", + "name": "aten::view_as_complex_copy(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -17272,7 +17687,7 @@ ] }, { - "name": "aten::view_as_real", + "name": "aten::view_as_real(Tensor(a) self) -> Tensor(a)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -17281,7 +17696,7 @@ ] }, { - "name": "aten::view_as_real_copy", + "name": "aten::view_as_real_copy(Tensor self) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -17290,7 +17705,7 @@ ] }, { - "name": "aten::view_copy", + "name": "aten::view_copy(Tensor self, SymInt[] size) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "size", "type": "SymInt[]" } @@ -17300,7 +17715,7 @@ ] }, { - "name": "aten::view_copy.dtype", + "name": "aten::view_copy.dtype(Tensor self, ScalarType dtype) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dtype", "type": "ScalarType" } @@ -17310,7 +17725,7 @@ ] }, { - "name": "aten::vstack", + "name": "aten::vstack(Tensor[] tensors) -> Tensor", "inputs": [ { "name": "tensors", "type": "Tensor[]" } ], @@ -17319,7 +17734,7 @@ ] }, { - "name": "aten::vstack.out", + "name": "aten::vstack.out(Tensor[] tensors, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "tensors", "type": "Tensor[]" } ], @@ -17328,7 +17743,7 @@ ] }, { - "name": "aten::wait", + "name": "aten::wait(Future(t) self) -> t", "inputs": [ { "name": "self", "type": "Future(t)" } ], @@ -17337,7 +17752,7 @@ ] }, { - "name": "aten::warn", + "name": "aten::warn(str message, int stacklevel=2) -> ()", "inputs": [ { "name": "message", "type": "string" }, { "name": "stacklevel", "type": "int64", "default": 2 } @@ -17345,7 +17760,7 @@ "outputs": [] }, { - "name": "aten::where", + "name": "aten::where(Tensor condition) -> Tensor[]", "inputs": [ { "name": "condition", "type": "Tensor" } ], @@ -17354,7 +17769,7 @@ ] }, { - "name": "aten::where.Scalar", + "name": "aten::where.Scalar(Tensor condition, Scalar self, Scalar other) -> Tensor", "inputs": [ { "name": "condition", "type": "Tensor" }, { "name": "self", "type": "Scalar" }, @@ -17365,7 +17780,7 @@ ] }, { - "name": "aten::where.ScalarOther", + "name": "aten::where.ScalarOther(Tensor condition, Tensor self, Scalar other) -> Tensor", "inputs": [ { "name": "condition", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -17376,7 +17791,7 @@ ] }, { - "name": "aten::where.ScalarSelf", + "name": "aten::where.ScalarSelf(Tensor condition, Scalar self, Tensor other) -> Tensor", "inputs": [ { "name": "condition", "type": "Tensor" }, { "name": "self", "type": "Scalar" }, @@ -17387,7 +17802,7 @@ ] }, { - "name": "aten::where.self", + "name": "aten::where.self(Tensor condition, Tensor self, Tensor other) -> Tensor", "inputs": [ { "name": "condition", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -17398,7 +17813,7 @@ ] }, { - "name": "aten::where.self_out", + "name": "aten::where.self_out(Tensor condition, Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "condition", "type": "Tensor" }, { "name": "self", "type": "Tensor" }, @@ -17409,7 +17824,7 @@ ] }, { - "name": "aten::zero_", + "name": "aten::zero_(Tensor(a!) self) -> Tensor(a!)", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -17418,7 +17833,7 @@ ] }, { - "name": "aten::zeros", + "name": "aten::zeros(SymInt[] size, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "size", "type": "SymInt[]" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -17431,7 +17846,7 @@ ] }, { - "name": "aten::zeros.names", + "name": "aten::zeros.names(int[] size, *, Dimname[]? names, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor", "inputs": [ { "name": "size", "type": "int64[]" }, { "name": "names", "type": "Dimname[]?", "kwarg_only": true }, @@ -17445,7 +17860,7 @@ ] }, { - "name": "aten::zeros.out", + "name": "aten::zeros.out(SymInt[] size, *, Tensor(a!) out) -> Tensor(a!)", "inputs": [ { "name": "size", "type": "SymInt[]" } ], @@ -17454,7 +17869,7 @@ ] }, { - "name": "aten::zeros_like", + "name": "aten::zeros_like(Tensor self, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None, MemoryFormat? memory_format=None) -> Tensor", "inputs": [ { "name": "self", "type": "Tensor" }, { "name": "dtype", "type": "ScalarType?", "default": null, "kwarg_only": true }, @@ -17468,17 +17883,17 @@ ] }, { - "name": "neuron::forward_v2_1", + "name": "neuron::forward_v2_1(Tensor[] _0, __torch__.torch.classes.neuron.Model _1) -> (Tensor _0)", "inputs": [ - { "name": "input", "type": "Tensor[]" }, - { "name": "model" } + { "name": "_0", "type": "Tensor[]" }, + { "name": "_1", "type": "__torch__.torch.classes.neuron.Model" } ], "outputs": [ - { "type": "Tensor" } + { "name": "_0", "type": "Tensor" } ] }, { - "name": "prepacked::conv2d_clamp_prepack", + "name": "prepacked::conv2d_clamp_prepack(Tensor W, Tensor? B, int[2] stride, int[2] padding, int[2] dilation, int groups, Scalar? output_min=None, Scalar? output_max=None) -> __torch__.torch.classes.xnnpack.Conv2dOpContext", "inputs": [ { "name": "W", "type": "Tensor" }, { "name": "B", "type": "Tensor?" }, @@ -17494,7 +17909,7 @@ ] }, { - "name": "prepacked::conv2d_clamp_run", + "name": "prepacked::conv2d_clamp_run(Tensor X, __torch__.torch.classes.xnnpack.Conv2dOpContext W_prepack) -> Tensor Y", "category": "Layer", "inputs": [ { "name": "X", "type": "Tensor" }, @@ -17505,7 +17920,7 @@ ] }, { - "name": "prepacked::conv2d_transpose_clamp_run", + "name": "prepacked::conv2d_transpose_clamp_run(Tensor X, __torch__.torch.classes.xnnpack.TransposeConv2dOpContext W_prepack) -> Tensor Y", "category": "Layer", "inputs": [ { "name": "X", "type": "Tensor" }, @@ -17516,7 +17931,7 @@ ] }, { - "name": "prepacked::linear_clamp_prepack", + "name": "prepacked::linear_clamp_prepack(Tensor W, Tensor? B=None, Scalar? output_min=None, Scalar? output_max=None) -> __torch__.torch.classes.xnnpack.LinearOpContext", "inputs": [ { "name": "W", "type": "Tensor" }, { "name": "B", "type": "Tensor?", "default": null }, @@ -17528,7 +17943,7 @@ ] }, { - "name": "prepacked::linear_clamp_run", + "name": "prepacked::linear_clamp_run(Tensor X, __torch__.torch.classes.xnnpack.LinearOpContext W_prepack) -> Tensor Y", "category": "Layer", "inputs": [ { "name": "X", "type": "Tensor" }, @@ -17539,7 +17954,7 @@ ] }, { - "name": "prim::AutogradAdd", + "name": "prim::AutogradAdd(Any a, Any b) -> Any", "inputs": [ { "name": "a", "type": "Any" }, { "name": "b", "type": "Any" } @@ -17549,42 +17964,42 @@ ] }, { - "name": "prim::AutogradAllNonZero", + "name": "prim::AutogradAllNonZero(...) -> bool", "inputs": [], "outputs": [ { "type": "boolean" } ] }, { - "name": "prim::AutogradAllZero", + "name": "prim::AutogradAllZero(...) -> bool", "inputs": [], "outputs": [ { "type": "boolean" } ] }, { - "name": "prim::AutogradAnyNonZero", + "name": "prim::AutogradAnyNonZero(...) -> bool", "inputs": [], "outputs": [ { "type": "boolean" } ] }, { - "name": "prim::AutogradZero", + "name": "prim::AutogradZero() -> Tensor", "inputs": [], "outputs": [ { "type": "Tensor" } ] }, { - "name": "prim::BroadcastSizes", + "name": "prim::BroadcastSizes(...) -> int[]", "inputs": [], "outputs": [ { "type": "int64[]" } ] }, { - "name": "prim::EnumName", + "name": "prim::EnumName(AnyEnumType enum) -> str", "inputs": [ { "name": "enum", "type": "AnyEnumType" } ], @@ -17593,7 +18008,7 @@ ] }, { - "name": "prim::EnumValue.float", + "name": "prim::EnumValue.float(AnyEnumType enum) -> float", "inputs": [ { "name": "enum", "type": "AnyEnumType" } ], @@ -17602,7 +18017,7 @@ ] }, { - "name": "prim::EnumValue.int", + "name": "prim::EnumValue.int(AnyEnumType enum) -> int", "inputs": [ { "name": "enum", "type": "AnyEnumType" } ], @@ -17611,7 +18026,7 @@ ] }, { - "name": "prim::EnumValue.str", + "name": "prim::EnumValue.str(AnyEnumType enum) -> str", "inputs": [ { "name": "enum", "type": "AnyEnumType" } ], @@ -17620,7 +18035,7 @@ ] }, { - "name": "prim::IfThenElse", + "name": "prim::IfThenElse(bool cond, Any(a) x, Any(b) y) -> Any(a|b)", "inputs": [ { "name": "cond", "type": "boolean" }, { "name": "x", "type": "Any" }, @@ -17631,7 +18046,7 @@ ] }, { - "name": "prim::ModuleContainerIndex.list", + "name": "prim::ModuleContainerIndex.list(Any self, int ind) -> Any", "inputs": [ { "name": "self", "type": "Any" }, { "name": "ind", "type": "int64" } @@ -17641,7 +18056,7 @@ ] }, { - "name": "prim::NumToTensor.Scalar", + "name": "prim::NumToTensor.Scalar(Scalar a) -> Tensor", "inputs": [ { "name": "a", "type": "Scalar" } ], @@ -17650,7 +18065,7 @@ ] }, { - "name": "prim::NumToTensor.bool", + "name": "prim::NumToTensor.bool(bool a) -> Tensor", "inputs": [ { "name": "a", "type": "boolean" } ], @@ -17659,12 +18074,12 @@ ] }, { - "name": "prim::Print", + "name": "prim::Print(...) -> ()", "inputs": [], "outputs": [] }, { - "name": "prim::RaiseException", + "name": "prim::RaiseException(str msg, str? cls=None) -> ()", "inputs": [ { "name": "msg", "type": "string" }, { "name": "cls", "type": "string?", "default": null } @@ -17672,7 +18087,7 @@ "outputs": [] }, { - "name": "prim::ReductionSizes", + "name": "prim::ReductionSizes(int[] size, int[] red_axes, bool keepdim = False) -> int[]", "inputs": [ { "name": "size", "type": "int64[]" }, { "name": "red_axes", "type": "int64[]" }, @@ -17683,7 +18098,7 @@ ] }, { - "name": "prim::StringIndex", + "name": "prim::StringIndex(str string, int index) -> str", "inputs": [ { "name": "string", "type": "string" }, { "name": "index", "type": "int64" } @@ -17693,7 +18108,7 @@ ] }, { - "name": "prim::TupleIndex", + "name": "prim::TupleIndex(Any tup, int i) -> Any", "inputs": [ { "name": "tup", "type": "Any" }, { "name": "i", "type": "int64" } @@ -17703,35 +18118,35 @@ ] }, { - "name": "prim::TupleUnpack", + "name": "prim::TupleUnpack(Any tup) -> ...", "inputs": [ { "name": "tup", "type": "Any" } ], "outputs": [] }, { - "name": "prim::Uninitialized", + "name": "prim::Uninitialized() -> Any", "inputs": [], "outputs": [ { "type": "Any" } ] }, { - "name": "prim::VarConcat", + "name": "prim::VarConcat(...) -> Tensor", "inputs": [], "outputs": [ { "type": "Tensor" } ] }, { - "name": "prim::VarStack", + "name": "prim::VarStack(...) -> Tensor", "inputs": [], "outputs": [ { "type": "Tensor" } ] }, { - "name": "prim::abs", + "name": "prim::abs(Tensor x) -> Tensor", "inputs": [ { "name": "x", "type": "Tensor" } ], @@ -17740,7 +18155,7 @@ ] }, { - "name": "prim::abs.Scalar", + "name": "prim::abs.Scalar(Scalar a) -> Scalar", "inputs": [ { "name": "a", "type": "Scalar" } ], @@ -17749,7 +18164,7 @@ ] }, { - "name": "prim::abs.complex", + "name": "prim::abs.complex(complex a) -> float", "inputs": [ { "name": "a", "type": "complex" } ], @@ -17758,7 +18173,7 @@ ] }, { - "name": "prim::abs.float", + "name": "prim::abs.float(float a) -> float", "inputs": [ { "name": "a", "type": "float32" } ], @@ -17767,7 +18182,7 @@ ] }, { - "name": "prim::abs.int", + "name": "prim::abs.int(int a) -> int", "inputs": [ { "name": "a", "type": "int64" } ], @@ -17776,7 +18191,7 @@ ] }, { - "name": "prim::data", + "name": "prim::data(Tensor(a) a) -> Tensor(a)", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -17785,7 +18200,7 @@ ] }, { - "name": "prim::device", + "name": "prim::device(Tensor a) -> Device", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -17794,7 +18209,7 @@ ] }, { - "name": "prim::dtype", + "name": "prim::dtype(Tensor a) -> int", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -17803,7 +18218,7 @@ ] }, { - "name": "prim::grad", + "name": "prim::grad(Tensor a) -> Tensor(*)", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -17812,7 +18227,7 @@ ] }, { - "name": "prim::id", + "name": "prim::id(AnyClassType? x) -> int", "inputs": [ { "name": "x", "type": "AnyClassType?" } ], @@ -17821,7 +18236,7 @@ ] }, { - "name": "prim::index", + "name": "prim::index(Device self) -> int?", "inputs": [ { "name": "self", "type": "Device" } ], @@ -17830,7 +18245,7 @@ ] }, { - "name": "prim::is_cpu", + "name": "prim::is_cpu(Tensor a) -> bool", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -17839,7 +18254,7 @@ ] }, { - "name": "prim::is_cuda", + "name": "prim::is_cuda(Tensor a) -> bool", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -17848,7 +18263,7 @@ ] }, { - "name": "prim::is_ipu", + "name": "prim::is_ipu(Tensor a) -> bool", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -17857,7 +18272,7 @@ ] }, { - "name": "prim::is_maia", + "name": "prim::is_maia(Tensor a) -> bool", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -17866,7 +18281,7 @@ ] }, { - "name": "prim::is_meta", + "name": "prim::is_meta(Tensor a) -> bool", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -17875,7 +18290,7 @@ ] }, { - "name": "prim::is_mkldnn", + "name": "prim::is_mkldnn(Tensor a) -> bool", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -17884,7 +18299,7 @@ ] }, { - "name": "prim::is_mps", + "name": "prim::is_mps(Tensor a) -> bool", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -17893,7 +18308,7 @@ ] }, { - "name": "prim::is_mtia", + "name": "prim::is_mtia(Tensor a) -> bool", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -17902,7 +18317,7 @@ ] }, { - "name": "prim::is_nested", + "name": "prim::is_nested(Tensor a) -> bool", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -17911,7 +18326,7 @@ ] }, { - "name": "prim::is_quantized", + "name": "prim::is_quantized(Tensor a) -> bool", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -17920,7 +18335,7 @@ ] }, { - "name": "prim::is_sparse", + "name": "prim::is_sparse(Tensor a) -> bool", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -17929,7 +18344,7 @@ ] }, { - "name": "prim::is_sparse_csr", + "name": "prim::is_sparse_csr(Tensor a) -> bool", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -17938,7 +18353,7 @@ ] }, { - "name": "prim::is_vulkan", + "name": "prim::is_vulkan(Tensor a) -> bool", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -17947,7 +18362,7 @@ ] }, { - "name": "prim::is_xla", + "name": "prim::is_xla(Tensor a) -> bool", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -17956,7 +18371,7 @@ ] }, { - "name": "prim::is_xpu", + "name": "prim::is_xpu(Tensor a) -> bool", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -17965,7 +18380,7 @@ ] }, { - "name": "prim::isinstance", + "name": "prim::isinstance(Any to_check) -> bool", "inputs": [ { "name": "to_check", "type": "Any" } ], @@ -17974,7 +18389,7 @@ ] }, { - "name": "prim::itemsize", + "name": "prim::itemsize(Tensor a) -> int", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -17983,7 +18398,7 @@ ] }, { - "name": "prim::layout", + "name": "prim::layout(Tensor a) -> Layout", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -17992,7 +18407,7 @@ ] }, { - "name": "prim::max", + "name": "prim::max(Scalar a, Scalar b) -> Scalar", "inputs": [ { "name": "a", "type": "Scalar" }, { "name": "b", "type": "Scalar" } @@ -18002,7 +18417,7 @@ ] }, { - "name": "prim::max.bool_list", + "name": "prim::max.bool_list(bool[] l, bool[] r) -> bool[]", "inputs": [ { "name": "l", "type": "boolean[]" }, { "name": "r", "type": "boolean[]" } @@ -18012,7 +18427,7 @@ ] }, { - "name": "prim::max.float", + "name": "prim::max.float(float a, float b) -> float", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "float32" } @@ -18022,7 +18437,7 @@ ] }, { - "name": "prim::max.float_int", + "name": "prim::max.float_int(float a, int b) -> float", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "int64" } @@ -18032,7 +18447,7 @@ ] }, { - "name": "prim::max.float_list", + "name": "prim::max.float_list(float[] l, float[] r) -> float[]", "inputs": [ { "name": "l", "type": "float32[]" }, { "name": "r", "type": "float32[]" } @@ -18042,7 +18457,7 @@ ] }, { - "name": "prim::max.int", + "name": "prim::max.int(int a, int b) -> int", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "int64" } @@ -18052,7 +18467,7 @@ ] }, { - "name": "prim::max.int_float", + "name": "prim::max.int_float(int a, float b) -> float", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "float32" } @@ -18062,7 +18477,7 @@ ] }, { - "name": "prim::max.int_list", + "name": "prim::max.int_list(int[] l, int[] r) -> int[]", "inputs": [ { "name": "l", "type": "int64[]" }, { "name": "r", "type": "int64[]" } @@ -18072,7 +18487,7 @@ ] }, { - "name": "prim::max.self_bool", + "name": "prim::max.self_bool(bool[] self) -> bool", "inputs": [ { "name": "self", "type": "boolean[]" } ], @@ -18081,7 +18496,7 @@ ] }, { - "name": "prim::max.self_float", + "name": "prim::max.self_float(float[] self) -> float", "inputs": [ { "name": "self", "type": "float32[]" } ], @@ -18090,7 +18505,7 @@ ] }, { - "name": "prim::max.self_int", + "name": "prim::max.self_int(int[] self) -> int", "inputs": [ { "name": "self", "type": "int64[]" } ], @@ -18099,7 +18514,7 @@ ] }, { - "name": "prim::min", + "name": "prim::min(Scalar a, Scalar b) -> Scalar", "inputs": [ { "name": "a", "type": "Scalar" }, { "name": "b", "type": "Scalar" } @@ -18109,7 +18524,7 @@ ] }, { - "name": "prim::min.bool_list", + "name": "prim::min.bool_list(bool[] l, bool[] r) -> bool[]", "inputs": [ { "name": "l", "type": "boolean[]" }, { "name": "r", "type": "boolean[]" } @@ -18119,7 +18534,7 @@ ] }, { - "name": "prim::min.float", + "name": "prim::min.float(float a, float b) -> float", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "float32" } @@ -18129,7 +18544,7 @@ ] }, { - "name": "prim::min.float_int", + "name": "prim::min.float_int(float a, int b) -> float", "inputs": [ { "name": "a", "type": "float32" }, { "name": "b", "type": "int64" } @@ -18139,7 +18554,7 @@ ] }, { - "name": "prim::min.float_list", + "name": "prim::min.float_list(float[] l, float[] r) -> float[]", "inputs": [ { "name": "l", "type": "float32[]" }, { "name": "r", "type": "float32[]" } @@ -18149,7 +18564,7 @@ ] }, { - "name": "prim::min.int", + "name": "prim::min.int(int a, int b) -> int", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "int64" } @@ -18159,7 +18574,7 @@ ] }, { - "name": "prim::min.int_float", + "name": "prim::min.int_float(int a, float b) -> float", "inputs": [ { "name": "a", "type": "int64" }, { "name": "b", "type": "float32" } @@ -18169,7 +18584,7 @@ ] }, { - "name": "prim::min.int_list", + "name": "prim::min.int_list(int[] l, int[] r) -> int[]", "inputs": [ { "name": "l", "type": "int64[]" }, { "name": "r", "type": "int64[]" } @@ -18179,7 +18594,7 @@ ] }, { - "name": "prim::min.self_bool", + "name": "prim::min.self_bool(bool[] self) -> bool", "inputs": [ { "name": "self", "type": "boolean[]" } ], @@ -18188,7 +18603,7 @@ ] }, { - "name": "prim::min.self_float", + "name": "prim::min.self_float(float[] self) -> float", "inputs": [ { "name": "self", "type": "float32[]" } ], @@ -18197,7 +18612,7 @@ ] }, { - "name": "prim::min.self_int", + "name": "prim::min.self_int(int[] self) -> int", "inputs": [ { "name": "self", "type": "int64[]" } ], @@ -18206,7 +18621,7 @@ ] }, { - "name": "prim::name", + "name": "prim::name(Tensor a) -> str?", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -18215,7 +18630,7 @@ ] }, { - "name": "prim::nbytes", + "name": "prim::nbytes(Tensor a) -> int", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -18224,7 +18639,7 @@ ] }, { - "name": "prim::rangelist", + "name": "prim::rangelist(int n) -> int[]", "inputs": [ { "name": "n", "type": "int64" } ], @@ -18233,7 +18648,7 @@ ] }, { - "name": "prim::requires_grad", + "name": "prim::requires_grad(Tensor a) -> bool", "inputs": [ { "name": "a", "type": "Tensor" } ], @@ -18242,7 +18657,7 @@ ] }, { - "name": "prim::shape", + "name": "prim::shape(Tensor self) -> int[]", "inputs": [ { "name": "self", "type": "Tensor" } ], @@ -18251,7 +18666,7 @@ ] }, { - "name": "prim::type", + "name": "prim::type(Device self) -> str", "inputs": [ { "name": "self", "type": "Device" } ], @@ -18260,7 +18675,7 @@ ] }, { - "name": "prim::unchecked_cast", + "name": "prim::unchecked_cast(t x) -> t", "inputs": [ { "name": "x", "type": "t" } ], @@ -18269,7 +18684,7 @@ ] }, { - "name": "prim::unchecked_unwrap_optional", + "name": "prim::unchecked_unwrap_optional(t(a)? optional) -> t(a)", "inputs": [ { "name": "optional", "type": "t?" } ], @@ -18278,7 +18693,7 @@ ] }, { - "name": "quantized::add", + "name": "quantized::add(Tensor qa, Tensor qb, float scale, int zero_point) -> Tensor qc", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "qb", "type": "Tensor" }, @@ -18290,7 +18705,7 @@ ] }, { - "name": "quantized::add.Scalar", + "name": "quantized::add.Scalar(Tensor qa, Scalar b) -> Tensor qc", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Scalar" } @@ -18300,7 +18715,7 @@ ] }, { - "name": "quantized::add.Scalar2", + "name": "quantized::add.Scalar2(Scalar b, Tensor qa) -> Tensor qc", "inputs": [ { "name": "b", "type": "Scalar" }, { "name": "qa", "type": "Tensor" } @@ -18310,7 +18725,7 @@ ] }, { - "name": "quantized::add.Scalar_out", + "name": "quantized::add.Scalar_out(Tensor qa, Scalar b, Tensor(a!) out) -> Tensor(a!) out", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Scalar" }, @@ -18321,7 +18736,7 @@ ] }, { - "name": "quantized::add.out", + "name": "quantized::add.out(Tensor qa, Tensor qb, Tensor(a!) out) -> Tensor(a!) out", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "qb", "type": "Tensor" }, @@ -18332,7 +18747,7 @@ ] }, { - "name": "quantized::add_out", + "name": "quantized::add_out(Tensor qa, Tensor qb, Tensor(a!) out) -> Tensor(a!) out", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "qb", "type": "Tensor" }, @@ -18343,7 +18758,7 @@ ] }, { - "name": "quantized::add_relu", + "name": "quantized::add_relu(Tensor qa, Tensor qb, float scale, int zero_point) -> Tensor qc", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "qb", "type": "Tensor" }, @@ -18355,7 +18770,7 @@ ] }, { - "name": "quantized::add_relu.Scalar", + "name": "quantized::add_relu.Scalar(Tensor qa, Scalar b) -> Tensor qc", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Scalar" } @@ -18365,7 +18780,7 @@ ] }, { - "name": "quantized::add_relu.Scalar2", + "name": "quantized::add_relu.Scalar2(Scalar b, Tensor qa) -> Tensor qc", "inputs": [ { "name": "b", "type": "Scalar" }, { "name": "qa", "type": "Tensor" } @@ -18375,7 +18790,7 @@ ] }, { - "name": "quantized::add_relu.Scalar_out", + "name": "quantized::add_relu.Scalar_out(Tensor qa, Scalar b, Tensor(a!) out) -> Tensor(a!) out", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Scalar" }, @@ -18386,7 +18801,7 @@ ] }, { - "name": "quantized::add_relu.out", + "name": "quantized::add_relu.out(Tensor qa, Tensor qb, Tensor(a!) out) -> Tensor(a!) out", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "qb", "type": "Tensor" }, @@ -18397,7 +18812,7 @@ ] }, { - "name": "quantized::add_relu_out", + "name": "quantized::add_relu_out(Tensor qa, Tensor qb, Tensor(a!) out) -> Tensor(a!) out", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "qb", "type": "Tensor" }, @@ -18408,7 +18823,7 @@ ] }, { - "name": "quantized::add_scalar", + "name": "quantized::add_scalar(Tensor qa, Scalar b) -> Tensor qc", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Scalar" } @@ -18418,7 +18833,7 @@ ] }, { - "name": "quantized::add_scalar.Tensor", + "name": "quantized::add_scalar.Tensor(Tensor qa, Tensor b) -> Tensor qc", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Tensor" } @@ -18428,7 +18843,7 @@ ] }, { - "name": "quantized::add_scalar_out", + "name": "quantized::add_scalar_out(Tensor qa, Scalar b, Tensor(a!) out) -> Tensor(a!) out", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Scalar" }, @@ -18439,7 +18854,7 @@ ] }, { - "name": "quantized::add_scalar_out.Tensor", + "name": "quantized::add_scalar_out.Tensor(Tensor qa, Tensor b, Tensor(a!) out) -> Tensor(a!) out", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Tensor" }, @@ -18450,7 +18865,7 @@ ] }, { - "name": "quantized::add_scalar_relu", + "name": "quantized::add_scalar_relu(Tensor qa, Scalar b) -> Tensor qc", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Scalar" } @@ -18460,7 +18875,7 @@ ] }, { - "name": "quantized::add_scalar_relu.Tensor", + "name": "quantized::add_scalar_relu.Tensor(Tensor qa, Tensor b) -> Tensor qc", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Tensor" } @@ -18470,7 +18885,7 @@ ] }, { - "name": "quantized::add_scalar_relu_out", + "name": "quantized::add_scalar_relu_out(Tensor qa, Scalar b, Tensor(a!) out) -> Tensor(a!) out", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Scalar" }, @@ -18481,7 +18896,7 @@ ] }, { - "name": "quantized::add_scalar_relu_out.Tensor", + "name": "quantized::add_scalar_relu_out.Tensor(Tensor qa, Tensor b, Tensor(a!) out) -> Tensor(a!) out", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Tensor" }, @@ -18492,7 +18907,7 @@ ] }, { - "name": "quantized::batch_norm", + "name": "quantized::batch_norm(Tensor qx, Tensor? weight, Tensor? bias, Tensor mean, Tensor var, float eps, float output_scale, int output_zero_point) -> Tensor", "inputs": [ { "name": "qx", "type": "Tensor" }, { "name": "weight", "type": "Tensor?" }, @@ -18508,7 +18923,7 @@ ] }, { - "name": "quantized::batch_norm1d", + "name": "quantized::batch_norm1d(Tensor qx, Tensor? weight, Tensor? bias, Tensor mean, Tensor var, float eps, float output_scale, int output_zero_point) -> Tensor", "category": "Normalization", "inputs": [ { "name": "qx", "type": "Tensor" }, @@ -18525,7 +18940,7 @@ ] }, { - "name": "quantized::batch_norm1d_relu", + "name": "quantized::batch_norm1d_relu(Tensor qx, Tensor? weight, Tensor? bias, Tensor mean, Tensor var, float eps, float output_scale, int output_zero_point) -> Tensor", "category": "Normalization", "inputs": [ { "name": "qx", "type": "Tensor" }, @@ -18542,7 +18957,7 @@ ] }, { - "name": "quantized::batch_norm2d", + "name": "quantized::batch_norm2d(Tensor qx, Tensor? weight, Tensor? bias, Tensor mean, Tensor var, float eps, float output_scale, int output_zero_point) -> Tensor", "category": "Normalization", "inputs": [ { "name": "qx", "type": "Tensor" }, @@ -18559,7 +18974,7 @@ ] }, { - "name": "quantized::batch_norm2d_relu", + "name": "quantized::batch_norm2d_relu(Tensor qx, Tensor? weight, Tensor? bias, Tensor mean, Tensor var, float eps, float output_scale, int output_zero_point) -> Tensor", "category": "Normalization", "inputs": [ { "name": "qx", "type": "Tensor" }, @@ -18576,7 +18991,7 @@ ] }, { - "name": "quantized::batch_norm3d", + "name": "quantized::batch_norm3d(Tensor qx, Tensor? weight, Tensor? bias, Tensor mean, Tensor var, float eps, float output_scale, int output_zero_point) -> Tensor", "category": "Normalization", "inputs": [ { "name": "qx", "type": "Tensor" }, @@ -18593,7 +19008,7 @@ ] }, { - "name": "quantized::batch_norm3d_relu", + "name": "quantized::batch_norm3d_relu(Tensor qx, Tensor? weight, Tensor? bias, Tensor mean, Tensor var, float eps, float output_scale, int output_zero_point) -> Tensor", "category": "Normalization", "inputs": [ { "name": "qx", "type": "Tensor" }, @@ -18610,7 +19025,7 @@ ] }, { - "name": "quantized::batch_norm_relu", + "name": "quantized::batch_norm_relu(Tensor qx, Tensor? weight, Tensor? bias, Tensor mean, Tensor var, float eps, float output_scale, int output_zero_point) -> Tensor", "category": "Normalization", "inputs": [ { "name": "qx", "type": "Tensor" }, @@ -18627,7 +19042,7 @@ ] }, { - "name": "quantized::cat", + "name": "quantized::cat(Tensor[] qx, int dim, float? scale, int? zero_point) -> Tensor", "category": "Tensor", "inputs": [ { "name": "qx", "type": "Tensor[]" }, @@ -18640,7 +19055,7 @@ ] }, { - "name": "quantized::cat_relu", + "name": "quantized::cat_relu(Tensor[] qx, int dim, float? scale, int? zero_point) -> Tensor", "category": "Tensor", "inputs": [ { "name": "qx", "type": "Tensor[]" }, @@ -18653,7 +19068,7 @@ ] }, { - "name": "quantized::celu", + "name": "quantized::celu(Tensor self, float output_scale, int output_zero_point, Scalar alpha=1) -> Tensor", "category": "Activation", "inputs": [ { "name": "self", "type": "Tensor" }, @@ -18666,7 +19081,7 @@ ] }, { - "name": "quantized::conv1d", + "name": "quantized::conv1d(Tensor qx, __torch__.torch.classes.quantized.Conv2dPackedParamsBase packed_weight, float output_scale, int output_zero_point) -> Tensor", "category": "Layer", "inputs": [ { "name": "qx", "type": "Tensor" }, @@ -18679,7 +19094,7 @@ ] }, { - "name": "quantized::conv1d_prepack", + "name": "quantized::conv1d_prepack(Tensor weight, Tensor? bias, int[] stride, int[] padding, int[] dilation, int groups) -> __torch__.torch.classes.quantized.Conv2dPackedParamsBase", "inputs": [ { "name": "weight", "type": "Tensor" }, { "name": "bias", "type": "Tensor?" }, @@ -18693,7 +19108,7 @@ ] }, { - "name": "quantized::conv1d_relu", + "name": "quantized::conv1d_relu(Tensor qx, __torch__.torch.classes.quantized.Conv2dPackedParamsBase packed_weight, float output_scale, int output_zero_point) -> Tensor", "category": "Layer", "inputs": [ { "name": "qx", "type": "Tensor" }, @@ -18706,7 +19121,7 @@ ] }, { - "name": "quantized::conv2d", + "name": "quantized::conv2d(Tensor qx, __torch__.torch.classes.quantized.Conv2dPackedParamsBase weight, int[] stride, int[] padding, int[] dilation, int groups, float output_scale, int output_zero_point) -> Tensor", "category": "Layer", "inputs": [ { "name": "qx", "type": "Tensor" }, @@ -18723,7 +19138,7 @@ ] }, { - "name": "quantized::conv2d.new", + "name": "quantized::conv2d.new(Tensor qx, __torch__.torch.classes.quantized.Conv2dPackedParamsBase packed_weight, float output_scale, int output_zero_point) -> Tensor", "category": "Layer", "inputs": [ { "name": "qx", "type": "Tensor" }, @@ -18736,7 +19151,7 @@ ] }, { - "name": "quantized::conv2d_dilation", + "name": "quantized::conv2d_dilation(__torch__.torch.classes.quantized.Conv2dPackedParamsBase packed_weights) -> int[]", "inputs": [ { "name": "packed_weights", "type": "__torch__.torch.classes.quantized.Conv2dPackedParamsBase" } ], @@ -18745,7 +19160,7 @@ ] }, { - "name": "quantized::conv2d_dynamic", + "name": "quantized::conv2d_dynamic(Tensor qx, __torch__.torch.classes.quantized.Conv2dPackedParamsBase packed_weight, bool reduce_range=False) -> Tensor", "inputs": [ { "name": "qx", "type": "Tensor" }, { "name": "packed_weight", "type": "__torch__.torch.classes.quantized.Conv2dPackedParamsBase" }, @@ -18756,7 +19171,7 @@ ] }, { - "name": "quantized::conv2d_groups", + "name": "quantized::conv2d_groups(__torch__.torch.classes.quantized.Conv2dPackedParamsBase packed_weights) -> int", "inputs": [ { "name": "packed_weights", "type": "__torch__.torch.classes.quantized.Conv2dPackedParamsBase" } ], @@ -18765,7 +19180,7 @@ ] }, { - "name": "quantized::conv2d_output_padding", + "name": "quantized::conv2d_output_padding(__torch__.torch.classes.quantized.Conv2dPackedParamsBase packed_weights) -> int[]", "inputs": [ { "name": "packed_weights", "type": "__torch__.torch.classes.quantized.Conv2dPackedParamsBase" } ], @@ -18774,7 +19189,7 @@ ] }, { - "name": "quantized::conv2d_padding", + "name": "quantized::conv2d_padding(__torch__.torch.classes.quantized.Conv2dPackedParamsBase packed_weights) -> int[]", "inputs": [ { "name": "packed_weights", "type": "__torch__.torch.classes.quantized.Conv2dPackedParamsBase" } ], @@ -18783,7 +19198,7 @@ ] }, { - "name": "quantized::conv2d_prepack", + "name": "quantized::conv2d_prepack(Tensor weight, Tensor? bias, int[] stride, int[] padding, int[] dilation, int groups) -> __torch__.torch.classes.quantized.Conv2dPackedParamsBase", "inputs": [ { "name": "weight", "type": "Tensor" }, { "name": "bias", "type": "Tensor?" }, @@ -18797,7 +19212,7 @@ ] }, { - "name": "quantized::conv2d_relu", + "name": "quantized::conv2d_relu(Tensor qx, __torch__.torch.classes.quantized.Conv2dPackedParamsBase weight, int[] stride, int[] padding, int[] dilation, int groups, float output_scale, int output_zero_point) -> Tensor", "category": "Layer", "inputs": [ { "name": "qx", "type": "Tensor" }, @@ -18814,7 +19229,7 @@ ] }, { - "name": "quantized::conv2d_relu.new", + "name": "quantized::conv2d_relu.new(Tensor qx, __torch__.torch.classes.quantized.Conv2dPackedParamsBase packed_weight, float output_scale, int output_zero_point) -> Tensor", "category": "Layer", "inputs": [ { "name": "qx", "type": "Tensor" }, @@ -18827,7 +19242,7 @@ ] }, { - "name": "quantized::conv2d_stride", + "name": "quantized::conv2d_stride(__torch__.torch.classes.quantized.Conv2dPackedParamsBase packed_weights) -> int[]", "inputs": [ { "name": "packed_weights", "type": "__torch__.torch.classes.quantized.Conv2dPackedParamsBase" } ], @@ -18836,7 +19251,7 @@ ] }, { - "name": "quantized::conv2d_transpose", + "name": "quantized::conv2d_transpose(__torch__.torch.classes.quantized.Conv2dPackedParamsBase packed_weights) -> int", "inputs": [ { "name": "packed_weights", "type": "__torch__.torch.classes.quantized.Conv2dPackedParamsBase" } ], @@ -18845,7 +19260,7 @@ ] }, { - "name": "quantized::conv2d_unpack", + "name": "quantized::conv2d_unpack(__torch__.torch.classes.quantized.Conv2dPackedParamsBase packed_weights) -> (Tensor unpacked_weights, Tensor? B_origin)", "inputs": [ { "name": "packed_weights", "type": "__torch__.torch.classes.quantized.Conv2dPackedParamsBase" } ], @@ -18855,7 +19270,7 @@ ] }, { - "name": "quantized::conv2d_unpack_sizes", + "name": "quantized::conv2d_unpack_sizes(Any packed_weights) -> (Any)", "inputs": [ { "name": "packed_weights", "type": "Any" } ], @@ -18864,7 +19279,7 @@ ] }, { - "name": "quantized::conv3d", + "name": "quantized::conv3d(Tensor qx, __torch__.torch.classes.quantized.Conv3dPackedParamsBase weight, int[] stride, int[] padding, int[] dilation, int groups, float output_scale, int output_zero_point) -> Tensor", "category": "Layer", "inputs": [ { "name": "qx", "type": "Tensor" }, @@ -18881,7 +19296,7 @@ ] }, { - "name": "quantized::conv3d.new", + "name": "quantized::conv3d.new(Tensor qx, __torch__.torch.classes.quantized.Conv3dPackedParamsBase packed_weight, float output_scale, int output_zero_point) -> Tensor", "category": "Layer", "inputs": [ { "name": "qx", "type": "Tensor" }, @@ -18894,7 +19309,7 @@ ] }, { - "name": "quantized::conv3d_prepack", + "name": "quantized::conv3d_prepack(Tensor weight, Tensor? bias, int[] stride, int[] padding, int[] dilation, int groups) -> __torch__.torch.classes.quantized.Conv3dPackedParamsBase", "inputs": [ { "name": "weight", "type": "Tensor" }, { "name": "bias", "type": "Tensor?" }, @@ -18908,7 +19323,7 @@ ] }, { - "name": "quantized::conv3d_relu", + "name": "quantized::conv3d_relu(Tensor qx, __torch__.torch.classes.quantized.Conv3dPackedParamsBase weight, int[] stride, int[] padding, int[] dilation, int groups, float output_scale, int output_zero_point) -> Tensor", "category": "Layer", "inputs": [ { "name": "qx", "type": "Tensor" }, @@ -18925,7 +19340,7 @@ ] }, { - "name": "quantized::conv3d_relu.new", + "name": "quantized::conv3d_relu.new(Tensor qx, __torch__.torch.classes.quantized.Conv3dPackedParamsBase packed_weight, float output_scale, int output_zero_point) -> Tensor", "category": "Layer", "inputs": [ { "name": "qx", "type": "Tensor" }, @@ -18938,7 +19353,7 @@ ] }, { - "name": "quantized::conv_prepack", + "name": "quantized::conv_prepack(Tensor weight, Tensor? bias, int[] stride, int[] padding, int[] dilation, int groups) -> __torch__.torch.classes.quantized.Conv2dPackedParamsBase", "inputs": [ { "name": "weight", "type": "Tensor" }, { "name": "bias", "type": "Tensor?" }, @@ -18952,7 +19367,7 @@ ] }, { - "name": "quantized::conv_transpose1d_prepack", + "name": "quantized::conv_transpose1d_prepack(Tensor weight, Tensor? bias, int[] stride, int[] padding, int[] output_padding, int[] dilation, int groups) -> __torch__.torch.classes.quantized.Conv2dPackedParamsBase", "inputs": [ { "name": "weight", "type": "Tensor" }, { "name": "bias", "type": "Tensor?" }, @@ -18967,7 +19382,7 @@ ] }, { - "name": "quantized::conv_transpose2d", + "name": "quantized::conv_transpose2d(Tensor qx, __torch__.torch.classes.quantized.Conv2dPackedParamsBase packed_weight, float output_scale, int output_zero_point) -> Tensor", "inputs": [ { "name": "qx", "type": "Tensor" }, { "name": "packed_weight", "type": "__torch__.torch.classes.quantized.Conv2dPackedParamsBase" }, @@ -18979,7 +19394,7 @@ ] }, { - "name": "quantized::conv_transpose2d_dilation", + "name": "quantized::conv_transpose2d_dilation(__torch__.torch.classes.quantized.Conv2dPackedParamsBase packed_weights) -> int[]", "inputs": [ { "name": "packed_weights", "type": "__torch__.torch.classes.quantized.Conv2dPackedParamsBase" } ], @@ -18988,7 +19403,7 @@ ] }, { - "name": "quantized::conv_transpose2d_dynamic", + "name": "quantized::conv_transpose2d_dynamic(Tensor qx, __torch__.torch.classes.quantized.Conv2dPackedParamsBase packed_weight, bool reduce_range=False) -> Tensor", "inputs": [ { "name": "qx", "type": "Tensor" }, { "name": "packed_weight", "type": "__torch__.torch.classes.quantized.Conv2dPackedParamsBase" }, @@ -18999,7 +19414,7 @@ ] }, { - "name": "quantized::conv_transpose2d_groups", + "name": "quantized::conv_transpose2d_groups(__torch__.torch.classes.quantized.Conv2dPackedParamsBase packed_weights) -> int", "inputs": [ { "name": "packed_weights", "type": "__torch__.torch.classes.quantized.Conv2dPackedParamsBase" } ], @@ -19008,7 +19423,7 @@ ] }, { - "name": "quantized::conv_transpose2d_output_padding", + "name": "quantized::conv_transpose2d_output_padding(__torch__.torch.classes.quantized.Conv2dPackedParamsBase packed_weights) -> int[]", "inputs": [ { "name": "packed_weights", "type": "__torch__.torch.classes.quantized.Conv2dPackedParamsBase" } ], @@ -19017,7 +19432,7 @@ ] }, { - "name": "quantized::conv_transpose2d_padding", + "name": "quantized::conv_transpose2d_padding(__torch__.torch.classes.quantized.Conv2dPackedParamsBase packed_weights) -> int[]", "inputs": [ { "name": "packed_weights", "type": "__torch__.torch.classes.quantized.Conv2dPackedParamsBase" } ], @@ -19026,7 +19441,7 @@ ] }, { - "name": "quantized::conv_transpose2d_prepack", + "name": "quantized::conv_transpose2d_prepack(Tensor weight, Tensor? bias, int[] stride, int[] padding, int[] output_padding, int[] dilation, int groups) -> __torch__.torch.classes.quantized.Conv2dPackedParamsBase", "inputs": [ { "name": "weight", "type": "Tensor" }, { "name": "bias", "type": "Tensor?" }, @@ -19041,7 +19456,7 @@ ] }, { - "name": "quantized::conv_transpose2d_stride", + "name": "quantized::conv_transpose2d_stride(__torch__.torch.classes.quantized.Conv2dPackedParamsBase packed_weights) -> int[]", "inputs": [ { "name": "packed_weights", "type": "__torch__.torch.classes.quantized.Conv2dPackedParamsBase" } ], @@ -19050,7 +19465,7 @@ ] }, { - "name": "quantized::conv_transpose2d_transpose", + "name": "quantized::conv_transpose2d_transpose(__torch__.torch.classes.quantized.Conv2dPackedParamsBase packed_weights) -> int", "inputs": [ { "name": "packed_weights", "type": "__torch__.torch.classes.quantized.Conv2dPackedParamsBase" } ], @@ -19059,7 +19474,7 @@ ] }, { - "name": "quantized::conv_transpose2d_unpack", + "name": "quantized::conv_transpose2d_unpack(__torch__.torch.classes.quantized.Conv2dPackedParamsBase packed_weights) -> (Tensor unpacked_weights, Tensor? B_origin)", "inputs": [ { "name": "packed_weights", "type": "__torch__.torch.classes.quantized.Conv2dPackedParamsBase" } ], @@ -19069,7 +19484,7 @@ ] }, { - "name": "quantized::embedding_bag_4bit_rowwise_offsets", + "name": "quantized::embedding_bag_4bit_rowwise_offsets(Tensor weight, Tensor indices, Tensor? offsets=None, bool scale_grad_by_freq=False, int mode=0, bool pruned_weights=False, Tensor? per_sample_weights=None, Tensor? compressed_indices_mapping=None, bool include_last_offset=False) -> Tensor", "inputs": [ { "name": "weight", "type": "Tensor" }, { "name": "indices", "type": "Tensor" }, @@ -19086,7 +19501,7 @@ ] }, { - "name": "quantized::embedding_bag_byte_rowwise_offsets", + "name": "quantized::embedding_bag_byte_rowwise_offsets(Tensor weight, Tensor indices, Tensor? offsets=None, bool scale_grad_by_freq=False, int mode=0, bool pruned_weights=False, Tensor? per_sample_weights=None, Tensor? compressed_indices_mapping=None, bool include_last_offset=False) -> Tensor", "inputs": [ { "name": "weight", "type": "Tensor" }, { "name": "indices", "type": "Tensor" }, @@ -19103,7 +19518,7 @@ ] }, { - "name": "quantized::embedding_byte", + "name": "quantized::embedding_byte(__torch__.torch.classes.quantized.EmbeddingPackedParamsBase weight, Tensor indices, bool pruned_weights=False) -> Tensor", "inputs": [ { "name": "weight", "type": "__torch__.torch.classes.quantized.EmbeddingPackedParamsBase" }, { "name": "indices", "type": "Tensor" }, @@ -19114,7 +19529,7 @@ ] }, { - "name": "quantized::hardswish", + "name": "quantized::hardswish(Tensor input, float output_scale, int output_zero_point) -> Tensor", "category": "Activation", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -19126,7 +19541,7 @@ ] }, { - "name": "quantized::instance_norm", + "name": "quantized::instance_norm(Tensor input, Tensor? weight, Tensor? bias, float eps, float output_scale, int output_zero_point) -> Tensor", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "weight", "type": "Tensor?" }, @@ -19140,7 +19555,7 @@ ] }, { - "name": "quantized::layer_norm", + "name": "quantized::layer_norm(Tensor input, int[] normalized_shape, Tensor? weight, Tensor? bias, float eps, float output_scale, int output_zero_point) -> Tensor", "category": "Normalization", "inputs": [ { "name": "input", "type": "Tensor" }, @@ -19156,7 +19571,7 @@ ] }, { - "name": "quantized::leaky_relu", + "name": "quantized::leaky_relu(Tensor qx, Scalar negative_slope, bool inplace, float output_scale, int output_zero_point) -> Tensor", "category": "Activation", "inputs": [ { "name": "qx", "type": "Tensor" }, @@ -19170,7 +19585,7 @@ ] }, { - "name": "quantized::linear", + "name": "quantized::linear(Tensor X, __torch__.torch.classes.quantized.LinearPackedParamsBase W_prepack, float Y_scale_i, int Y_zero_point_i) -> Tensor Y", "category": "Layer", "inputs": [ { "name": "X", "type": "Tensor" }, @@ -19183,7 +19598,7 @@ ] }, { - "name": "quantized::linear_dynamic", + "name": "quantized::linear_dynamic(Tensor X, __torch__.torch.classes.quantized.LinearPackedParamsBase W_prepack, bool reduce_range=False) -> Tensor Y", "category": "Layer", "inputs": [ { "name": "X", "type": "Tensor" }, @@ -19195,7 +19610,7 @@ ] }, { - "name": "quantized::linear_prepack_fp16", + "name": "quantized::linear_prepack_fp16(Tensor W, Tensor? B=None) -> __torch__.torch.classes.quantized.LinearPackedParamsBase W_prepack", "inputs": [ { "name": "W", "type": "Tensor" }, { "name": "B", "type": "Tensor?", "default": null } @@ -19205,7 +19620,7 @@ ] }, { - "name": "quantized::linear_prepack_fp16_legacy", + "name": "quantized::linear_prepack_fp16_legacy(Tensor W, Tensor? B=None) -> Tensor W_prepack", "inputs": [ { "name": "W", "type": "Tensor" }, { "name": "B", "type": "Tensor?", "default": null } @@ -19215,7 +19630,7 @@ ] }, { - "name": "quantized::linear_relu", + "name": "quantized::linear_relu(Tensor X, __torch__.torch.classes.quantized.LinearPackedParamsBase W_prepack, float Y_scale_i, int Y_zero_point_i) -> Tensor Y", "category": "Layer", "inputs": [ { "name": "X", "type": "Tensor" }, @@ -19228,7 +19643,7 @@ ] }, { - "name": "quantized::linear_relu_dynamic", + "name": "quantized::linear_relu_dynamic(Tensor X, __torch__.torch.classes.quantized.LinearPackedParamsBase W_prepack, bool reduce_range=False) -> Tensor Y", "category": "Layer", "inputs": [ { "name": "X", "type": "Tensor" }, @@ -19240,7 +19655,7 @@ ] }, { - "name": "quantized::make_quantized_cell_params", + "name": "quantized::make_quantized_cell_params(Tensor w_ih, Tensor w_hh, Tensor b_ih, Tensor b_hh) -> __torch__.torch.classes.rnn.CellParamsBase", "inputs": [ { "name": "w_ih", "type": "Tensor" }, { "name": "w_hh", "type": "Tensor" }, @@ -19252,7 +19667,7 @@ ] }, { - "name": "quantized::make_quantized_cell_params_dynamic", + "name": "quantized::make_quantized_cell_params_dynamic(__torch__.torch.classes.quantized.LinearPackedParamsBase w_ih, __torch__.torch.classes.quantized.LinearPackedParamsBase w_hh, Tensor bias_ih, Tensor bias_hh, bool reduce_range=False) -> __torch__.torch.classes.rnn.CellParamsBase", "inputs": [ { "name": "w_ih", "type": "__torch__.torch.classes.quantized.LinearPackedParamsBase" }, { "name": "w_hh", "type": "__torch__.torch.classes.quantized.LinearPackedParamsBase" }, @@ -19265,7 +19680,7 @@ ] }, { - "name": "quantized::make_quantized_cell_params_fp16", + "name": "quantized::make_quantized_cell_params_fp16(__torch__.torch.classes.quantized.LinearPackedParamsBase w_ih, __torch__.torch.classes.quantized.LinearPackedParamsBase w_hh) -> __torch__.torch.classes.rnn.CellParamsBase", "inputs": [ { "name": "w_ih", "type": "__torch__.torch.classes.quantized.LinearPackedParamsBase" }, { "name": "w_hh", "type": "__torch__.torch.classes.quantized.LinearPackedParamsBase" } @@ -19275,7 +19690,7 @@ ] }, { - "name": "quantized::mul", + "name": "quantized::mul(Tensor qa, Tensor qb, float scale, int zero_point)-> Tensor qc", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "qb", "type": "Tensor" }, @@ -19287,7 +19702,7 @@ ] }, { - "name": "quantized::mul.Scalar", + "name": "quantized::mul.Scalar(Tensor qa, Scalar b)-> Tensor qc", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Scalar" } @@ -19297,7 +19712,7 @@ ] }, { - "name": "quantized::mul.Scalar2", + "name": "quantized::mul.Scalar2(Scalar b, Tensor qa)-> Tensor qc", "inputs": [ { "name": "b", "type": "Scalar" }, { "name": "qa", "type": "Tensor" } @@ -19307,7 +19722,7 @@ ] }, { - "name": "quantized::mul.Scalar_out", + "name": "quantized::mul.Scalar_out(Tensor qa, Scalar b, Tensor(a!) out)-> Tensor(a!) out", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Scalar" }, @@ -19318,7 +19733,7 @@ ] }, { - "name": "quantized::mul.out", + "name": "quantized::mul.out(Tensor qa, Tensor qb, Tensor(a!) out)-> Tensor(a!) out", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "qb", "type": "Tensor" }, @@ -19329,7 +19744,7 @@ ] }, { - "name": "quantized::mul_out", + "name": "quantized::mul_out(Tensor qa, Tensor qb, Tensor(a!) out)-> Tensor(a!) out", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "qb", "type": "Tensor" }, @@ -19340,7 +19755,7 @@ ] }, { - "name": "quantized::mul_relu", + "name": "quantized::mul_relu(Tensor qa, Tensor qb, float scale, int zero_point)-> Tensor qc", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "qb", "type": "Tensor" }, @@ -19352,7 +19767,7 @@ ] }, { - "name": "quantized::mul_relu.Scalar", + "name": "quantized::mul_relu.Scalar(Tensor qa, Scalar b)-> Tensor qc", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Scalar" } @@ -19362,7 +19777,7 @@ ] }, { - "name": "quantized::mul_relu.Scalar2", + "name": "quantized::mul_relu.Scalar2(Scalar b, Tensor qa)-> Tensor qc", "inputs": [ { "name": "b", "type": "Scalar" }, { "name": "qa", "type": "Tensor" } @@ -19372,7 +19787,7 @@ ] }, { - "name": "quantized::mul_relu.Scalar_out", + "name": "quantized::mul_relu.Scalar_out(Tensor qa, Scalar b, Tensor(a!) out)-> Tensor(a!) out", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Scalar" }, @@ -19383,7 +19798,7 @@ ] }, { - "name": "quantized::mul_relu.out", + "name": "quantized::mul_relu.out(Tensor qa, Tensor qb, Tensor(a!) out)-> Tensor(a!) out", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "qb", "type": "Tensor" }, @@ -19394,7 +19809,7 @@ ] }, { - "name": "quantized::mul_relu_out", + "name": "quantized::mul_relu_out(Tensor qa, Tensor qb, Tensor(a!) out)-> Tensor(a!) out", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "qb", "type": "Tensor" }, @@ -19405,7 +19820,7 @@ ] }, { - "name": "quantized::mul_scalar", + "name": "quantized::mul_scalar(Tensor qa, Scalar b)-> Tensor qc", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Scalar" } @@ -19415,7 +19830,7 @@ ] }, { - "name": "quantized::mul_scalar.Tensor", + "name": "quantized::mul_scalar.Tensor(Tensor qa, Tensor b)-> Tensor qc", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Tensor" } @@ -19425,7 +19840,7 @@ ] }, { - "name": "quantized::mul_scalar_out", + "name": "quantized::mul_scalar_out(Tensor qa, Scalar b, Tensor(a!) out)-> Tensor(a!) out", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Scalar" }, @@ -19436,7 +19851,7 @@ ] }, { - "name": "quantized::mul_scalar_out.Tensor", + "name": "quantized::mul_scalar_out.Tensor(Tensor qa, Tensor b, Tensor(a!) out)-> Tensor(a!) out", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Tensor" }, @@ -19447,7 +19862,7 @@ ] }, { - "name": "quantized::mul_scalar_relu", + "name": "quantized::mul_scalar_relu(Tensor qa, Scalar b)-> Tensor qc", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Scalar" } @@ -19457,7 +19872,7 @@ ] }, { - "name": "quantized::mul_scalar_relu.Tensor", + "name": "quantized::mul_scalar_relu.Tensor(Tensor qa, Tensor b)-> Tensor qc", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Tensor" } @@ -19467,7 +19882,7 @@ ] }, { - "name": "quantized::mul_scalar_relu_out", + "name": "quantized::mul_scalar_relu_out(Tensor qa, Scalar b, Tensor(a!) out)-> Tensor(a!) out", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Scalar" }, @@ -19478,7 +19893,7 @@ ] }, { - "name": "quantized::mul_scalar_relu_out.Tensor", + "name": "quantized::mul_scalar_relu_out.Tensor(Tensor qa, Tensor b, Tensor(a!) out)-> Tensor(a!) out", "inputs": [ { "name": "qa", "type": "Tensor" }, { "name": "b", "type": "Tensor" }, @@ -19489,7 +19904,7 @@ ] }, { - "name": "quantized::prelu", + "name": "quantized::prelu(Tensor qx, Tensor weight, float output_scale, int output_zero_point) -> Tensor", "inputs": [ { "name": "qx", "type": "Tensor" }, { "name": "weight", "type": "Tensor" }, @@ -19501,7 +19916,7 @@ ] }, { - "name": "quantized::quantized_gru_cell_dynamic", + "name": "quantized::quantized_gru_cell_dynamic(Tensor input, Tensor hx, __torch__.torch.classes.quantized.LinearPackedParamsBase w_ih, __torch__.torch.classes.quantized.LinearPackedParamsBase w_hh, Tensor b_ih, Tensor b_hh) -> Tensor", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "hx", "type": "Tensor" }, @@ -19515,7 +19930,7 @@ ] }, { - "name": "quantized::quantized_lstm_cell_dynamic", + "name": "quantized::quantized_lstm_cell_dynamic(Tensor input, Tensor[] hx, __torch__.torch.classes.quantized.LinearPackedParamsBase w_ih, __torch__.torch.classes.quantized.LinearPackedParamsBase w_hh, Tensor bias_ih, Tensor bias_hh) -> (Tensor, Tensor)", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "hx", "type": "Tensor[]" }, @@ -19530,7 +19945,7 @@ ] }, { - "name": "quantized::quantized_rnn_relu_cell_dynamic", + "name": "quantized::quantized_rnn_relu_cell_dynamic(Tensor input, Tensor hx, __torch__.torch.classes.quantized.LinearPackedParamsBase w_ih, __torch__.torch.classes.quantized.LinearPackedParamsBase w_hh, Tensor b_ih, Tensor b_hh) -> Tensor", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "hx", "type": "Tensor" }, @@ -19544,7 +19959,7 @@ ] }, { - "name": "quantized::quantized_rnn_tanh_cell_dynamic", + "name": "quantized::quantized_rnn_tanh_cell_dynamic(Tensor input, Tensor hx, __torch__.torch.classes.quantized.LinearPackedParamsBase w_ih, __torch__.torch.classes.quantized.LinearPackedParamsBase w_hh, Tensor b_ih, Tensor b_hh) -> Tensor", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "hx", "type": "Tensor" }, @@ -19558,7 +19973,7 @@ ] }, { - "name": "quantized::relu6", + "name": "quantized::relu6(Tensor qx, bool inplace=False) -> Tensor", "category": "Activation", "inputs": [ { "name": "qx", "type": "Tensor" }, @@ -19569,7 +19984,7 @@ ] }, { - "name": "quantized::sigmoid", + "name": "quantized::sigmoid(Tensor qx, float output_scale, int output_zero_point) -> Tensor", "category": "Activation", "inputs": [ { "name": "qx", "type": "Tensor" }, @@ -19581,7 +19996,7 @@ ] }, { - "name": "quantized::softmax", + "name": "quantized::softmax(Tensor qx, int dim, float output_scale, int output_zero_point) -> Tensor", "inputs": [ { "name": "qx", "type": "Tensor" }, { "name": "dim", "type": "int64" }, @@ -19967,7 +20382,7 @@ "category": "Data" }, { - "name": "torchaudio::sox_effects_apply_effects_tensor", + "name": "torchaudio::sox_effects_apply_effects_tensor(Tensor tensor, int sample_rate, str[][] effects, bool channels_first=True) -> (Tensor, int64)", "inputs": [ { "name": "tensor", "type": "Tensor" }, { "name": "sample_rate", "type": "int64" }, @@ -19976,11 +20391,11 @@ ], "outputs": [ { "type": "Tensor" }, - { "name": "?", "type": "Tensor" } + { "type": "int64" } ] }, { - "name": "torchvision::nms", + "name": "torchvision::nms(Tensor dets, Tensor scores, float iou_threshold) -> Tensor", "inputs": [ { "name": "dets", "type": "Tensor" }, { "name": "scores", "type": "Tensor" }, @@ -19991,7 +20406,7 @@ ] }, { - "name": "torchvision::roi_align", + "name": "torchvision::roi_align(Tensor input, Tensor rois, float spatial_scale, int pooled_height, int pooled_width, int sampling_ratio, bool aligned) -> Tensor", "inputs": [ { "name": "input", "type": "Tensor" }, { "name": "rois", "type": "Tensor" }, diff --git a/source/pytorch.js b/source/pytorch.js index 2cae8ecce9..c4edb4416d 100644 --- a/source/pytorch.js +++ b/source/pytorch.js @@ -364,7 +364,8 @@ pytorch.Node = class { const value = metadata.type(key); const type = value ? { ...value } : { name }; type.identifier = name; - type.name = type.name.indexOf('::') === -1 ? type.name : type.name.split('::').pop().split('.')[0]; + [name] = type.name.split('('); + type.name = name.indexOf('::') === -1 ? name : name.split('::').pop().split('.')[0]; return type; }; const createAttribute = (metadata, name, value) => { @@ -397,6 +398,7 @@ pytorch.Node = class { let module = null; if (pytorch.Utility.isInstance(obj, 'torch.Node')) { const node = obj; + // const schema = node.schema(); this.type = createType(metadata, node.kind()); for (const name of node.attributeNames()) { const kind = node.kindOf(name); @@ -1213,6 +1215,7 @@ pytorch.Container.Zip = class extends pytorch.Container { execution.trace = false; const module = torch.jit.load(reader); execution.trace = true; + execution.registerMetadata(metadata); if (module.data && module.data.forward) { this.module = module; } else { @@ -1279,9 +1282,10 @@ pytorch.Container.ModelJson = class extends pytorch.Container { this.producer = this._model.producerName + (this._model.producerVersion ? ` v${this._model.producerVersion}` : ''); } this.format = reader.has_record('attributes.pkl') ? 'TorchScript v1.1' : 'TorchScript v1.0'; - execution.false = true; + execution.trace = false; const module = torch.jit.load(reader); execution.trace = true; + execution.registerMetadata(metadata); if (module.data && module.data.forward) { this.module = module; } else { @@ -1690,63 +1694,9 @@ pytorch.Execution = class extends python.Execution { const args = [this.data]; // self if (this.data.forward.__code__ && this.data.forward.__code__.args) { for (const arg of this.data.forward.__code__.args) { - const defaultValue = (type, name) => { - if (type.type === 'type' && type.name.type) { - switch (type.name.value) { - case 'Tensor': { - const tensor = execution.invoke('torch.Tensor', []); - tensor.__variable__ = name; - tensor.__origin__ = 'graph-input'; - const value = execution.variable(tensor, execution.graph.param_node()); - if (value && name) { - value.setDebugName(name); - } - return tensor; - } - case 'Tuple': { - return type.arguments.map((type, index) => defaultValue(type, `${name}[${index}]`)); - } - case 'List': { - return type.arguments.map((type, index) => defaultValue(type, `${name}[${index}]`)); - } - case 'Dict': { - if (type.arguments[1].name.value === 'Tensor') { - const Dict = class extends Map { - get(key) { - if (!super.has(key)) { - super.set(key, defaultValue(type.arguments[1], `${name}:${key}`)); - } - return super.get(key); - } - }; - return new Dict(); - } - return new Map(); - } - case 'int': { - return 0; - } - case 'float': { - return 0.0; - } - case 'bool': { - return false; - } - case 'Optional': { - return undefined; - } - case 'str': - return ''; - default: { - break; - } - } - } - throw new pytorch.Error(`Unsupported parameter type '${JSON.stringify(type)}'.`); - }; if (arg.name !== 'self') { - const type = arg.parameterType; - const value = defaultValue(type, arg.name); + const value = execution.graph.addInput(arg.name); + value.setType(execution.type(arg.parameterType)); if (pytorch.Utility.isTensor(value)) { value.__variable__ = arg.name; value.__origin__ = 'graph-input'; @@ -1888,92 +1838,93 @@ pytorch.Execution = class extends python.Execution { const torch = this.torch; switch (type) { case undefined: return null; - case 'boolean': return new torch.BoolType(); + case 'boolean': return torch.BoolType.get(); case 'boolean[3]': - case 'boolean[]': return new torch.ListType(new torch.BoolType()); - case 'boolean?': return new torch.OptionalType(new torch.BoolType()); - case 'int64': return new torch.IntType(); + case 'boolean[]': return torch.ListType.get(torch.BoolType.get()); + case 'boolean?': return torch.OptionalType.get(torch.BoolType.get()); + case 'int64': return torch.IntType.get(); case 'int64[1]': case 'int64[2]': case 'int64[3]': - case 'int64[]': return new torch.ListType(new torch.IntType()); - case 'int64?': return new torch.OptionalType(new torch.IntType()); + case 'int64[]': return torch.ListType.get(torch.IntType.get()); + case 'int64?': return torch.OptionalType.get(torch.IntType.get()); case 'int64[1]?': case 'int64[2]?': - case 'int64[]?': return new torch.OptionalType(new torch.ListType(new torch.IntType())); - case 'SymInt': return new torch.SymIntType(); + case 'int64[]?': return torch.OptionalType.get(torch.ListType.get(torch.IntType.get())); + case 'SymInt': return torch.SymIntType.get(); case 'SymInt[1]': case 'SymInt[2]': case 'SymInt[3]': case 'SymInt[4]': case 'SymInt[5]': case 'SymInt[6]': - case 'SymInt[]': return new torch.ListType(new torch.SymIntType()); - case 'SymInt?': return new torch.OptionalType(new torch.SymIntType()); + case 'SymInt[]': return torch.ListType.get(torch.SymIntType.get()); + case 'SymInt?': return torch.OptionalType.get(torch.SymIntType.get()); case 'SymInt[1]?': - case 'SymInt[]?': return new torch.OptionalType(new torch.ListType(new torch.SymIntType())); - case 'float32': return new torch.FloatType(); - case 'float32[]': return new torch.ListType(new torch.FloatType()); - case 'float32?': return new torch.OptionalType(new torch.FloatType()); - case 'float32[]?': return new torch.OptionalType(new torch.ListType(new torch.FloatType())); - case 'string': return new torch.StringType(); - case 'string[]': return new torch.ListType(new torch.StringType()); - case 'string[][]': return new torch.ListType(new torch.ListType(new torch.StringType())); - case 'string?': return new torch.OptionalType(new torch.StringType()); - case 'complex': return new torch.ComplexType(); - case 'complex[]': return new torch.ListType(new torch.ComplexType()); - case 'Tensor': return new torch.TensorType(); - case 'Tensor[]': return new torch.ListType(new torch.TensorType()); - case 'Tensor?': return new torch.OptionalType(new torch.TensorType()); - case 'Tensor?[]': return new torch.ListType(new torch.OptionalType(new torch.TensorType())); - case 'Scalar': return new torch.NumberType(); - case 'Scalar?': return new torch.OptionalType(new torch.NumberType()); - case 'ScalarType': return new torch.Type('ScalarTypeType'); - case 'ScalarType?': return new torch.OptionalType(new torch.Type('ScalarTypeType')); - case 'Layout': return new torch.Type('Layout'); - case 'Layout?': return new torch.OptionalType(new torch.Type('Layout')); - case 'Device': return new torch.DeviceObjType(); - case 'Device?': return new torch.OptionalType(new torch.DeviceObjType()); - case 'MemoryFormat': return new torch.Type('MemoryFormat'); - case 'MemoryFormat?': return new torch.OptionalType(new torch.Type('MemoryFormat')); - case 'Dimname': return new torch.StringType(); + case 'SymInt[]?': return torch.OptionalType.get(torch.ListType.get(torch.SymIntType.get())); + case 'float32': return torch.FloatType.get(); + case 'float32[]': return torch.ListType.get(torch.FloatType.get()); + case 'float32?': return torch.OptionalType.get(torch.FloatType.get()); + case 'float32[]?': return torch.OptionalType.get(torch.ListType.get(torch.FloatType.get())); + case 'string': return torch.StringType.get(); + case 'string[1]': + case 'string[]': return torch.ListType.get(torch.StringType.get()); + case 'string[][]': return torch.ListType.get(torch.ListType.get(torch.StringType.get())); + case 'string?': return torch.OptionalType.get(torch.StringType.get()); + case 'complex': return torch.ComplexType.get(); + case 'complex[]': return torch.ListType.get(torch.ComplexType.get()); + case 'Tensor': return torch.TensorType.get(); + case 'Tensor[]': return torch.ListType.get(torch.TensorType.get()); + case 'Tensor?': return torch.OptionalType.get(torch.TensorType.get()); + case 'Tensor?[]': return torch.ListType.get(torch.OptionalType.get(torch.TensorType.get())); + case 'Scalar': return torch.NumberType.get(); + case 'Scalar?': return torch.OptionalType.get(torch.NumberType.get()); + case 'ScalarType': return torch.Type.get('ScalarTypeType'); + case 'ScalarType?': return torch.OptionalType.get(torch.Type.get('ScalarTypeType')); + case 'Layout': return torch.Type.get('Layout'); + case 'Layout?': return torch.OptionalType.get(torch.Type.get('Layout')); + case 'Device': return torch.DeviceObjType.get(); + case 'Device?': return torch.OptionalType.get(torch.DeviceObjType.get()); + case 'MemoryFormat': return torch.Type.get('MemoryFormat'); + case 'MemoryFormat?': return torch.OptionalType.get(torch.Type.get('MemoryFormat')); + case 'Dimname': return torch.StringType.get(); case 'Dimname[1]': - case 'Dimname[]': return new torch.ListType(new torch.StringType()); - case 'Dimname[]?': return new torch.OptionalType(new torch.ListType(new torch.StringType())); + case 'Dimname[]': return torch.ListType.get(torch.StringType.get()); + case 'Dimname[]?': return torch.OptionalType.get(torch.ListType.get(torch.StringType.get())); case 'Generator': return new torch._C._GeneratorType(); - case 'Generator?': return new torch.OptionalType(new torch._C._GeneratorType()); - case 'Any': return new torch.AnyType(); - case 'Any[]': return new torch.ListType(new torch.AnyType()); - case 'AnyEnumType': return new torch.Type('AnyEnumType'); - case 't[]': return new torch.ListType(new torch.Type('VarType', 't')); - case 't?': return new torch.OptionalType(new torch.Type('VarType', 't')); + case 'Generator?': return torch.OptionalType.get(new torch._C._GeneratorType()); + case 'Any': return torch.AnyType.get(); + case 'Any[]': return torch.ListType.get(torch.AnyType.get()); + case 'AnyEnumType': return torch.Type.get('AnyEnumType'); + case 't[]': return torch.ListType.get(torch.Type.get('VarType', 't')); + case 't?': return torch.OptionalType.get(torch.Type.get('VarType', 't')); case 't': case 't1': case 't2': case 'tVal': { if (types && types.has(type)) { return types.get(type); } - return new torch.Type('VarType', type); + return torch.Type.get('VarType', type); } - case 'Future(t)': return new torch.FutureType(new torch.Type('VarType', 't')); + case 'Future(t)': return torch.FutureType.get(torch.Type.get('VarType', 't')); case 'AnyClassType': return null; - case 'AnyClassType?': return new torch.OptionalType(null); - case 'Dict(string, tVal)': return new torch.DictType(new torch.StringType(), new torch.Type('VarType', 'tVal')); - case 'Dict(string, t)': return new torch.DictType(new torch.StringType(), new torch.Type('VarType', 't')); - case 'Dict(string, Tensor)': return new torch.DictType(new torch.StringType(), new torch.TensorType('VarType')); - case 'Dict(int64, tVal)': return new torch.DictType(new torch.IntType(), new torch.Type('VarType', 'tVal')); - case 'Dict(int64, t)': return new torch.DictType(new torch.IntType(), new torch.Type('VarType', 't')); - case 'Dict(float32, tVal)': return new torch.DictType(new torch.FloatType(), new torch.Type('VarType', 'tVal')); - case 'Dict(float32, t)': return new torch.DictType(new torch.FloatType(), new torch.Type('VarType', 't')); - case 'Dict(boolean, tVal)': return new torch.DictType(new torch.BoolType(), new torch.Type('VarType', 'tVal')); - case 'Dict(boolean, t)': return new torch.DictType(new torch.BoolType(), new torch.Type('VarType', 't')); - case 'Dict(complex, tVal)': return new torch.DictType(new torch.ComplexType(), new torch.Type('VarType', 'tVal')); - case 'Dict(complex, t)': return new torch.DictType(new torch.ComplexType(), new torch.Type('VarType', 't')); - case 'Dict(Tensor, t)': return new torch.DictType(new torch.TensorType(), new torch.Type('VarType', 't')); - case 'Dict(Tensor, tVal)': return new torch.DictType(new torch.TensorType(), new torch.Type('VarType', 'tVal')); - case '(boolean, t)[]': return new torch.ListType(new torch.TupleType([new torch.BoolType(), new torch.Type('VarType', 't')])); - case '(boolean, tVal)[]': return new torch.ListType(new torch.TupleType([new torch.BoolType(), new torch.Type('VarType', 'Val')])); - case '(float32, t)[]': return new torch.ListType(new torch.TupleType([new torch.FloatType(), new torch.Type('VarType', 't')])); - case '(float32, tVal)[]': return new torch.ListType(new torch.TupleType([new torch.FloatType(), new torch.Type('VarType', 'tVal')])); - case '(int64, t)[]': return new torch.ListType(new torch.TupleType([new torch.IntType(), new torch.Type('VarType', 't')])); - case '(int64, tVal)[]': return new torch.ListType(new torch.TupleType([new torch.IntType(), new torch.Type('VarType', 'tVal')])); - case '(string, t)[]': return new torch.ListType(new torch.TupleType([new torch.StringType(), new torch.Type('VarType', 't')])); - case '(string, tVal)[]': return new torch.ListType(new torch.TupleType([new torch.StringType(), new torch.Type('VarType', 'tVal')])); - case '(complex, t)[]': return new torch.ListType(new torch.TupleType([new torch.ComplexType(), new torch.Type('VarType', 't')])); - case '(complex, tVal)[]': return new torch.ListType(new torch.TupleType([new torch.ComplexType(), new torch.Type('VarType', 'tVal')])); - case '(Tensor, t)[]': return new torch.ListType(new torch.TupleType([new torch.TensorType(), new torch.Type('VarType', 't')])); - case '(Tensor, tVal)[]': return new torch.ListType(new torch.TupleType([new torch.TensorType(), new torch.Type('VarType', 'tVal')])); + case 'AnyClassType?': return torch.OptionalType.get(null); + case 'Dict(string, tVal)': return torch.DictType.get(torch.StringType.get(), torch.Type.get('VarType', 'tVal')); + case 'Dict(string, t)': return torch.DictType.get(torch.StringType.get(), torch.Type.get('VarType', 't')); + case 'Dict(string, Tensor)': return torch.DictType.get(torch.StringType.get(), torch.TensorType.get()); + case 'Dict(int64, tVal)': return torch.DictType.get(torch.IntType.get(), torch.Type.get('VarType', 'tVal')); + case 'Dict(int64, t)': return torch.DictType.get(torch.IntType.get(), torch.Type.get('VarType', 't')); + case 'Dict(float32, tVal)': return torch.DictType.get(torch.FloatType.get(), torch.Type.get('VarType', 'tVal')); + case 'Dict(float32, t)': return torch.DictType.get(torch.FloatType.get(), torch.Type.get('VarType', 't')); + case 'Dict(boolean, tVal)': return torch.DictType.get(torch.BoolType.get(), torch.Type.get('VarType', 'tVal')); + case 'Dict(boolean, t)': return torch.DictType.get(torch.BoolType.get(), torch.Type.get('VarType', 't')); + case 'Dict(complex, tVal)': return torch.DictType.get(torch.ComplexType.get(), torch.Type.get('VarType', 'tVal')); + case 'Dict(complex, t)': return torch.DictType.get(torch.ComplexType.get(), torch.Type.get('VarType', 't')); + case 'Dict(Tensor, t)': return torch.DictType.get(torch.TensorType.get(), torch.Type.get('VarType', 't')); + case 'Dict(Tensor, tVal)': return torch.DictType.get(torch.TensorType.get(), torch.Type.get('VarType', 'tVal')); + case '(boolean, t)[]': return torch.ListType.get(torch.TupleType.get([torch.BoolType.get(), torch.Type.get('VarType', 't')])); + case '(boolean, tVal)[]': return torch.ListType.get(torch.TupleType.get([torch.BoolType.get(), torch.Type.get('VarType', 'Val')])); + case '(float32, t)[]': return torch.ListType.get(torch.TupleType.get([torch.FloatType.get(), torch.Type.get('VarType', 't')])); + case '(float32, tVal)[]': return torch.ListType.get(torch.TupleType.get([torch.FloatType.get(), torch.Type.get('VarType', 'tVal')])); + case '(int64, t)[]': return torch.ListType.get(torch.TupleType.get([torch.IntType.get(), torch.Type.get('VarType', 't')])); + case '(int64, tVal)[]': return torch.ListType.get(torch.TupleType.get([torch.IntType.get(), torch.Type.get('VarType', 'tVal')])); + case '(string, t)[]': return torch.ListType.get(torch.TupleType.get([torch.StringType.get(), torch.Type.get('VarType', 't')])); + case '(string, tVal)[]': return torch.ListType.get(torch.TupleType.get([torch.StringType.get(), torch.Type.get('VarType', 'tVal')])); + case '(complex, t)[]': return torch.ListType.get(torch.TupleType.get([torch.ComplexType.get(), torch.Type.get('VarType', 't')])); + case '(complex, tVal)[]': return torch.ListType.get(torch.TupleType.get([torch.ComplexType.get(), torch.Type.get('VarType', 'tVal')])); + case '(Tensor, t)[]': return torch.ListType.get(torch.TupleType.get([torch.TensorType.get(), torch.Type.get('VarType', 't')])); + case '(Tensor, tVal)[]': return torch.ListType.get(torch.TupleType.get([torch.TensorType.get(), torch.Type.get('VarType', 'tVal')])); default: { if (type.startsWith('__torch__.')) { return new torch.ClassType(type); @@ -1990,11 +1941,17 @@ pytorch.Execution = class extends python.Execution { const modules = new Set(); for (const [name, type] of metadata._types) { if (name.indexOf('::') !== -1) { - const [name, overload_name] = type.name.split('.'); - const args = type.inputs.map((arg) => new torch.Argument(arg.name, null, this.toType(arg.type), null, arg.default, arg.kwarg_only || false, arg.alias_info)); - const returns = type.outputs.map((arg) => new torch.Argument(arg.name, null, this.toType(arg.type), null, arg.default, arg.kwarg_only || false, arg.alias_info)); - const schema = new torch.FunctionSchema(name, overload_name || '', args, returns, type.is_vararg || false, type.is_varret || false); - // console.log(schema.__str__()); + // const [name, overload_name] = type.name.split('.'); + // const args = type.inputs.map((arg) => new torch.Argument(arg.name, null, this.toType(arg.type), null, arg.default, arg.kwarg_only || false, arg.alias_info)); + // const returns = type.outputs.map((arg) => new torch.Argument(arg.name, null, this.toType(arg.type), null, arg.default, arg.kwarg_only || false, arg.alias_info)); + // let schema = new torch.FunctionSchema(name, overload_name || '', args, returns, type.is_vararg || false, type.is_varret || false); + // if (type.name.indexOf('(') === -1) { + // console.log(schema.__str__()); + // } + const schema = torch.FunctionSchema.parse(type.name); + if (type.category) { + schema.category = type.category; + } const op = new torch._C.Operator(schema); registry.registerOperator(op); modules.add(type.name.split('::')[0]); @@ -2138,6 +2095,9 @@ pytorch.jit.Execution = class extends pytorch.Execution { return_node() { return this._block.return_node(); } + addInput(name) { + return this._block.addInput(name); + } }); this.registerType('torch.Block', class { constructor(graph) { @@ -2181,6 +2141,23 @@ pytorch.jit.Execution = class extends pytorch.Execution { kind() { return this._kind; } + schema() { + if (this._op === undefined) { + this._op = null; + const registry = torch._C._get_registry(); + const index = this._kind.indexOf('.'); + const name = index === -1 ? this._kind : this._kind.substring(0, index); + const overload_name = index === -1 ? '' : this._kind.substring(index + 1); + const candidates = registry.getAllOperatorsFor(name); + for (const candidate of candidates) { + if (candidate.schema().overload_name === overload_name) { + this._op = candidate; + break; + } + } + } + return this._op ? this._op.schema() : null; + } inputs() { return this._inputs; } @@ -2289,8 +2266,7 @@ pytorch.jit.Execution = class extends pytorch.Execution { }); this._metadata = metadata; this._types = new Map(); - for (const [, value] of this._metadata._types) { - const name = value.name; + for (const [name, value] of this._metadata._types) { if (name.indexOf('::') !== -1) { const index = name.lastIndexOf('.'); const key = index === -1 ? name : name.substring(0, index); @@ -2324,24 +2300,24 @@ pytorch.jit.Execution = class extends pytorch.Execution { let type = null; if (value === null) { node.ival_('value', value); - type = new torch.NoneType(); + type = torch.NoneType.get(); } else if (typeof value === 'string') { node.s_('value', value); - type = new torch.StringType(); + type = torch.StringType.get(); } else if (Array.isArray(value) && value.every((item) => typeof item === 'string')) { node.ss_('value', value); - type = new torch.ListType(new torch.StringType()); + type = torch.ListType.get(torch.StringType.get()); } else if (typeof value === 'boolean') { // return value; node.i_('value', value === true ? 1 : 0); - type = new torch.BoolType(); + type = torch.BoolType.get(); } else if (Number.isInteger(value)) { node.i_('value', value); - type = new torch.IntType(); + type = torch.IntType.get(); } else if (typeof value === 'number') { // return value; node.f_('value', value); - type = new torch.FloatType(); + type = torch.FloatType.get(); } else { throw new pytorch.Error(`Unsupported value type '${typeof value}'.`); } @@ -2367,7 +2343,7 @@ pytorch.jit.Execution = class extends pytorch.Execution { } if (pytorch.Utility.isTensor(obj)) { value.value = obj; - value.setType(new torch.TensorType()); + value.setType(torch.TensorType.get()); if (typeof obj !== 'string' && typeof obj !== 'number') { this._values.set(obj, value); } @@ -2377,7 +2353,7 @@ pytorch.jit.Execution = class extends pytorch.Execution { } if (typeof obj === 'string') { value.value = obj; - value.setType(new torch.StringType()); + value.setType(torch.StringType.get()); } return value; } @@ -2548,12 +2524,10 @@ pytorch.jit.Execution = class extends pytorch.Execution { } if (expression.target.type === 'id' && expression.target.value === 'uninitialized') { const type = this.type(expression.args[0], context); - // let value = this.expression(expression.args[0], context); const node = this._graph.create('prim::Uninitialized'); const value = node.addOutput(); value.setType(type); return value; - // throw new pytorch.Error(`Unsupported uninitialized type '${type.kind()}'.`); } if (expression.target.type === 'id' && expression.target.value === 'unchecked_cast') { let value = this.expression(expression.args[1], context); @@ -2570,7 +2544,7 @@ pytorch.jit.Execution = class extends pytorch.Execution { const node = this._graph.create('prim::isinstance'); node.addInput(this.variable(value)); value = node.addOutput(); - value.setType(new torch.BoolType()); + value.setType(torch.BoolType.get()); return value; } /* @@ -2613,9 +2587,11 @@ pytorch.jit.Execution = class extends pytorch.Execution { const node = this._graph.create('aten::__getitem__.t'); node.addInput(target); if (target.type().getKeyType() instanceof torch.StringType && typeof key === 'string') { - const value = this.invoke('torch.Value', [node]); + const value = new torch.Value(node); value.value = key; key = value; + } else if (target.type().getKeyType() instanceof torch.StringType && key.type() instanceof torch.StringType) { + // continue } else { throw new pytorch.Error(`Unsupported dictionary key type.`); } @@ -2665,23 +2641,23 @@ pytorch.jit.Execution = class extends pytorch.Execution { for (const item of list) { if (item instanceof torch.Value) { node.addInput(item); - output.setType(new torch.ListType(item.type())); + output.setType(torch.ListType.get(item.type())); } else if (Number.isInteger(item)) { const value = new torch.Value(node); value.value = item; - value.setType(new torch.IntType()); + value.setType(torch.IntType.get()); node.addInput(value); - output.setType(new torch.ListType(new torch.IntType())); + output.setType(torch.ListType.get(torch.IntType.get())); } else if (typeof item === 'string') { const value = new torch.Value(node); value.value = item; - value.setType(new torch.StringType()); + value.setType(torch.StringType.get()); node.addInput(value); - output.setType(new torch.ListType(new torch.StringType())); + output.setType(torch.ListType.get(torch.StringType.get())); } else if (pytorch.Utility.isTensor(item)) { const value = this.variable(item, node); node.addInput(value); - output.setType(new torch.ListType(new torch.TensorType())); + output.setType(torch.ListType.get(torch.TensorType.get())); } else { const value = new torch.Value(node); value.value = item; @@ -2706,37 +2682,37 @@ pytorch.jit.Execution = class extends pytorch.Execution { const value = this.variable(item, node); node.addInput(value); // value.value = item; - // value.setType(new torch.TensorType()); + // value.setType(torch.TensorType.get()); types.push(value.type()); elements.push(item); } else if (Number.isInteger(item)) { const value = new torch.Value(node); value.value = item; - types.push(new torch.IntType()); + types.push(torch.IntType.get()); elements.push(item); } else if (typeof item === 'boolean') { const value = new torch.Value(node); value.value = item; node.addInput(value); - types.push(new torch.BoolType()); + types.push(torch.BoolType.get()); elements.push(item); } else if (item === null) { const value = new torch.Value(node); value.value = item; node.addInput(value); - types.push(new torch.NoneType()); + types.push(torch.NoneType.get()); elements.push(item); } else { const value = new torch.Value(node); value.value = item; node.addInput(value); - types.push(new torch.Type()); + types.push(torch.Type.get()); elements.push(item); } } const value = node.addOutput(); value.value = elements; - value.setType(new torch.TupleType(types)); + value.setType(torch.TupleType.get(types)); return value; } case 'dict': { @@ -2757,7 +2733,7 @@ pytorch.jit.Execution = class extends pytorch.Execution { } const output = node.addOutput(); if (keyType && valueType) { - output.setType(new torch.DictType(keyType, valueType)); + output.setType(torch.DictType.get(keyType, valueType)); } return output; } @@ -2787,6 +2763,18 @@ pytorch.jit.Execution = class extends pytorch.Execution { value.__type__ = type; return undefined; } + case 'if': { + if (!this.traceIf) { + return super.statement(statement, context); + } + /* + const test = this.expression(statement.test, context); + const n = this._graph.create('prim::If'); + const true_block = n.addBlock(); + const false_block = n.addBlock(); + */ + return undefined; + } default: { break; } @@ -2800,20 +2788,20 @@ pytorch.jit.Execution = class extends pytorch.Execution { switch (expression.target.value) { case 'List': { const elementType = this.type(expression.arguments.value[0]); - return new torch.ListType(elementType); + return torch.ListType.get(elementType); } case 'Optional': { const elementType = this.type(expression.arguments.value[0]); - return new torch.OptionalType(elementType); + return torch.OptionalType.get(elementType); } case 'Tuple': { const elements = expression.arguments.value.map((expression) => this.type(expression)); - return new torch.TupleType(elements); + return torch.TupleType.get(elements); } case 'Dict': { const key = this.type(expression.arguments.value[0]); const value = this.type(expression.arguments.value[1]); - return new torch.DictType(key, value); + return torch.DictType.get(key, value); } case 'Final': { return this.type(expression.arguments.value[0]); @@ -2825,14 +2813,14 @@ pytorch.jit.Execution = class extends pytorch.Execution { } if (expression.type === 'id') { switch (expression.value) { - case 'Tensor': return new torch.TensorType(); - case 'int': return new torch.IntType(); - case 'str': return new torch.StringType(); - case 'float': return new torch.FloatType(); - case 'number': return new torch.NumberType(); - case 'bool': return new torch.BoolType(); - case 'None': return new torch.NoneType(); - case 'NoneType': return new torch.NoneType(); + case 'Tensor': return torch.TensorType.get(); + case 'int': return torch.IntType.get(); + case 'str': return torch.StringType.get(); + case 'float': return torch.FloatType.get(); + case 'number': return torch.NumberType.get(); + case 'bool': return torch.BoolType.get(); + case 'None': return torch.NoneType.get(); + case 'NoneType': return torch.NoneType.get(); default: throw new pytorch.Error(`Unsupported type expression '${expression.value}'.`); } } @@ -2896,8 +2884,9 @@ pytorch.jit.Execution = class extends pytorch.Execution { const [schema, evalArgs] = overload; const copyArgs = Array.prototype.slice.call(args); const copyEvalArgs = Array.prototype.slice.call(evalArgs); - const node = this._graph.create(schema.name); - node.schema = schema; + const index = schema.name.indexOf('('); + const op = index === -1 ? schema.name : schema.name.substring(0, index); + const node = this._graph.create(op); const referencedParameters = []; const parameters = Array.prototype.slice.call(schema.inputs || []).concat(Array.prototype.slice.call(schema.attributes || [])); const types = new Map(); @@ -2970,11 +2959,11 @@ pytorch.jit.Execution = class extends pytorch.Execution { tensor.__count__ = (tensor.__count__ || 0) + 1; } const value = this.variable(tensor); - value.setType(new torch.TensorType()); + value.setType(torch.TensorType.get()); list.addInput(value); } const output = list.addOutput(); - output.setType(new torch.ListType(new torch.TensorType())); + output.setType(torch.ListType.get(torch.TensorType.get())); input = output; match = true; } @@ -3014,7 +3003,7 @@ pytorch.jit.Execution = class extends pytorch.Execution { if (input instanceof torch.Value && input.type()) { types.set(key, input.type()); } else if (input instanceof torch.Value && Number.isInteger(input.value)) { - types.set(key, new torch.IntType()); + types.set(key, torch.IntType.get()); } // throw new pytorch.Error("Unknown value type 't'."); } @@ -3025,11 +3014,11 @@ pytorch.jit.Execution = class extends pytorch.Execution { } else if (input instanceof torch.Value && input.type() instanceof torch.ListType) { types.set(key, input.type().getElementType()); } else if (Array.isArray(input) && input.length > 0 && input.every((item) => Number.isInteger(item))) { - types.set(key, new torch.IntType()); + types.set(key, torch.IntType.get()); } else if (input.value && Array.isArray(input.value) && input.value.length > 0 && input.value.every((item) => Number.isInteger(item) || isNaN(item))) { - types.set(key, new torch.IntType()); + types.set(key, torch.IntType.get()); } else if (input.value && Array.isArray(input.value) && input.value.length > 0 && input.value.every((item) => pytorch.Utility.isTensor(item))) { - types.set(key, new torch.TensorType()); + types.set(key, torch.TensorType.get()); } else { // throw new pytorch.Error("Unknown value type 't'."); continue; @@ -3293,7 +3282,7 @@ pytorch.jit.Execution = class extends pytorch.Execution { } const value = node.addOutput(); - value.setType(new torch.ListType(new torch.TensorType())); + value.setType(torch.ListType.get(torch.TensorType.get())); result.push(value); /* @@ -3327,35 +3316,35 @@ pytorch.jit.Execution = class extends pytorch.Execution { case 'Scalar': { const value = this.variable(null, node); value.__origin__ = schema.name; - value.setType(new torch.NumberType()); + value.setType(torch.NumberType.get()); result.push(value); break; } case 'boolean': { const value = this.variable(null, node); value.__origin__ = schema.name; - value.setType(new torch.BoolType()); + value.setType(torch.BoolType.get()); result.push(value); break; } case 'boolean[]': { const value = this.variable(null, node); value.__origin__ = schema.name; - value.setType(new torch.ListType(new torch.BoolType())); + value.setType(torch.ListType.get(torch.BoolType.get())); result.push(value); break; } case 'string[]': { const value = this.variable(null, node); value.__origin__ = schema.name; - value.setType(new torch.ListType(new torch.StringType())); + value.setType(torch.ListType.get(torch.StringType.get())); result.push(value); break; } case 'int64': { const value = this.variable(null, node); value.__origin__ = schema.name; - value.setType(new torch.IntType()); + value.setType(torch.IntType.get()); switch (schema.name) { case 'aten::div.int': value.value = torch.div(evalArgs[0], evalArgs[1]); break; case 'aten::dim': value.value = torch.dim(evalArgs[0]); break; @@ -3369,7 +3358,7 @@ pytorch.jit.Execution = class extends pytorch.Execution { case 'int64[]': { const value = this.variable(null, node); value.__origin__ = schema.name; - value.setType(new torch.ListType(new torch.IntType())); + value.setType(torch.ListType.get(torch.IntType.get())); switch (schema.name) { // case 'aten::size': value.value = torch.size(evalArgs[0], evalArgs[1]); break; default: break; @@ -3380,42 +3369,42 @@ pytorch.jit.Execution = class extends pytorch.Execution { case 'float32': { const value = this.variable(null, node); value.__origin__ = schema.name; - value.setType(new torch.FloatType()); + value.setType(torch.FloatType.get()); result.push(value); break; } case 'float32[]': { const value = this.variable(null, node); value.__origin__ = schema.name; - value.setType(new torch.ListType(new torch.FloatType())); + value.setType(torch.ListType.get(torch.FloatType.get())); result.push(value); break; } case 'complex': { const value = this.variable(null, node); value.__origin__ = schema.name; - value.setType(new torch.ComplexType()); + value.setType(torch.ComplexType.get()); result.push(value); break; } case 'string': { const value = this.variable(null, node); value.__origin__ = schema.name; - value.setType(new torch.StringType()); + value.setType(torch.StringType.get()); result.push(value); break; } case 'Device': { const value = this.variable(null, node); value.__origin__ = schema.name; - value.setType(new torch.DeviceObjType()); + value.setType(torch.DeviceObjType.get()); result.push(value); break; } case 'Dict(string, Tensor)': { const value = this.variable(null, node); value.__origin__ = schema.name; - value.setType(new torch.DictType(new torch.StringType(), new torch.TensorType())); + value.setType(torch.DictType.get(torch.StringType.get(), torch.TensorType.get())); result.push(value); break; } @@ -3437,7 +3426,7 @@ pytorch.jit.Execution = class extends pytorch.Execution { if (!t) { throw new pytorch.Error(); } - value.setType(new torch.ListType(t)); + value.setType(torch.ListType.get(t)); result.push(value); break; } @@ -3448,7 +3437,7 @@ pytorch.jit.Execution = class extends pytorch.Execution { value.__origin__ = schema.name; const keyType = this.toType(dict[1], types); const valueType = this.toType(dict[2], types); - value.setType(new torch.DictType(keyType, valueType)); + value.setType(torch.DictType.get(keyType, valueType)); result.push(value); break; } @@ -3458,7 +3447,7 @@ pytorch.jit.Execution = class extends pytorch.Execution { value.__origin__ = schema.name; const keyType = this.toType(tuple[1], types); const valueType = this.toType(tuple[2], types); - value.setType(new torch.ListType(new torch.TupleType([keyType, valueType]))); + value.setType(torch.ListType.get(torch.TupleType.get([keyType, valueType]))); result.push(value); break; } @@ -3623,7 +3612,6 @@ pytorch.jit.Execution = class extends pytorch.Execution { } const torch = this.torch; const type = name ? `${moduleName}.${name}` : moduleName; - // const outputTypes = null; // https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/native_functions.yaml let overloads = null; if (type.startsWith('torch.')) { @@ -3850,10 +3838,6 @@ pytorch.jit.Execution = class extends pytorch.Execution { break; default: { throw new pytorch.Error(`Unknown return type '${parameter.type}'.`); - // if (!outputTypes || schema.outputs.length !== 1 || schema.outputs[0].type !== outputTypes[0]) { - // next = true; - // } - // break; } } } @@ -3881,6 +3865,9 @@ pytorch.jit.Execution = class extends pytorch.Execution { } block(statements, context) { + if (!this.traceIf) { + return super.block(statements, context); + } statements = Array.prototype.slice.call(statements); while (statements.length > 0) { if (statements.length > 1) { @@ -4721,6 +4708,8 @@ pytorch.Utility = class { return 'int64[]'; } else if (Array.isArray(value) && value.every((item) => Number(item) === item)) { return 'float32[]'; + } else if (pytorch.Utility.isInstance(value, 'torch.Value')) { + return pytorch.Utility.toType(value.type()); } const text = (JSON.stringify(value) || '(undefined)').substring(0, 10); throw new pytorch.Error(`Unsupported ops argument type '${text}'.`); @@ -5303,7 +5292,9 @@ pytorch.Metadata = class { if (data) { const items = JSON.parse(data); for (const item of items) { - this._types.set(item.name, item); + const index = item.name.indexOf('('); + const key = index === -1 ? item.name : item.name.substring(0, index); + this._types.set(key, item); } } } diff --git a/source/pytorch.py b/source/pytorch.py index c4b0c70aea..35af7c3ef1 100644 --- a/source/pytorch.py +++ b/source/pytorch.py @@ -305,6 +305,7 @@ def _argument(self, argument, value): class Schema: # pylint: disable=too-few-public-methods,missing-class-docstring def __init__(self, value): + self.value = value lexer = Schema.Lexer(value) lexer.whitespace(0) self._parse_name(lexer) diff --git a/test/models.json b/test/models.json index 6a444f67c9..48de312d80 100644 --- a/test/models.json +++ b/test/models.json @@ -5691,7 +5691,6 @@ "type": "pytorch", "target": "module_000007.pt", "source": "https://github.com/lutzroeder/netron/files/9075631/module_000007.pt.zip[module_000007.pt]", - "error": "Unknown function 'torch.size'.", "link": "https://github.com/lutzroeder/netron/issues/842" }, { @@ -5783,7 +5782,6 @@ "type": "pytorch", "target": "netron_issue_547_1.pt", "source": "https://github.com/lutzroeder/netron/files/5137393/netron_issue_547_1.zip[netron_issue_547_1.pt]", - "error": "Unknown function 'torch.add'.", "link": "https://github.com/lutzroeder/netron/issues/547" }, { diff --git a/tools/pytorch b/tools/pytorch index fef92cad31..d10d5dce04 100755 --- a/tools/pytorch +++ b/tools/pytorch @@ -8,6 +8,16 @@ case "${OSTYPE}" in *) python="python";; esac +venv() { + env_dir=./third_party/env/pytorch + [ -d "${env_dir}" ] || ${python} -m venv ${env_dir} + case "${OSTYPE}" in + msys*) source ${env_dir}/Scripts/activate;; + *) source ${env_dir}/bin/activate;; + esac + ${python} -m pip install --quiet --upgrade pip requests +} + clean() { echo "pytorch clean" rm -rf "./third_party/source/pytorch" @@ -23,6 +33,16 @@ sync() { curl --silent --location --output "./third_party/source/executorch/schema/program.fbs" "https://github.com/pytorch/executorch/raw/main/schema/program.fbs" } +install() { + echo "pytorch install" + venv + ${python} -m pip install --quiet --upgrade wheel + ${python} -m pip install --quiet --upgrade --pre torch torchvision torchaudio --index-url https://download.pytorch.org/whl/nightly/cpu + # ${python} -m pip install --quiet --upgrade torch-neuron --index-url https://pip.repos.neuron.amazonaws.com + deactivate +} + + schema() { [[ $(grep -U $'\x0D' ./source/pytorch-schema.js) ]] && crlf=1 echo "pytorch schema" @@ -45,6 +65,7 @@ while [ "$#" != 0 ]; do command="$1" && shift case "${command}" in "clean") clean;; + "install") install;; "sync") sync;; "schema") schema;; "metadata") metadata;; diff --git a/tools/pytorch_script.py b/tools/pytorch_script.py index 3992a347c8..d4609cef3b 100644 --- a/tools/pytorch_script.py +++ b/tools/pytorch_script.py @@ -28,6 +28,7 @@ def _read_metadata(): metadata = {} for value in json.loads(_read(metadata_file)): key = value['name'] + key = key.split("(")[0] if key in metadata: raise ValueError(f"Duplicate key '{key}'") metadata[key] = value @@ -64,6 +65,7 @@ def _write_metadata(value): re.compile(r'(prim::.*->\s*.*)"', re.MULTILINE)), ] +# pylint: disable=line-too-long known_schema_definitions = [ 'aten::__and__.bool(bool a, bool b) -> bool', 'aten::__and__.int(int a, int b) -> int', @@ -77,12 +79,40 @@ def _write_metadata(value): 'aten::__getitem__.Dict_Tensor(Dict(Tensor, t) self, Tensor key) -> t(*)', 'aten::__getitem__.str(str s, int index) -> str', 'aten::__getitem__.t(t[](a) list, int idx) -> t(*)', + 'aten::_native_batch_norm_legit(Tensor input, Tensor? weight, Tensor? bias, Tensor(a!) running_mean, Tensor(b!) running_var, bool training, float momentum, float eps) -> (Tensor, Tensor, Tensor)', + 'aten::_native_batch_norm_legit.no_stats(Tensor input, Tensor? weight, Tensor? bias, bool training, float momentum, float eps) -> (Tensor, Tensor, Tensor)', + 'aten::_native_batch_norm_legit.no_stats_out(Tensor input, Tensor? weight, Tensor? bias, bool training, float momentum, float eps, *, Tensor(a!) out, Tensor(b!) save_mean, Tensor(c!) save_invstd) -> (Tensor(a!), Tensor(b!), Tensor(c!))', + 'aten::_native_batch_norm_legit.out(Tensor input, Tensor? weight, Tensor? bias, Tensor(a!) running_mean, Tensor(b!) running_var, bool training, float momentum, float eps, *, Tensor(d!) out, Tensor(e!) save_mean, Tensor(f!) save_invstd) -> (Tensor(d!), Tensor(e!), Tensor(f!))', + 'aten::_native_batch_norm_legit_functional(Tensor input, Tensor? weight, Tensor? bias, Tensor running_mean, Tensor running_var, bool training, float momentum, float eps) -> (Tensor, Tensor, Tensor, Tensor running_mean_out, Tensor running_var_out)', + 'aten::_native_batch_norm_legit_no_training(Tensor input, Tensor? weight, Tensor? bias, Tensor running_mean, Tensor running_var, float momentum, float eps) -> (Tensor, Tensor, Tensor)', + 'aten::_native_batch_norm_legit_no_training.out(Tensor input, Tensor? weight, Tensor? bias, Tensor running_mean, Tensor running_var, float momentum, float eps, *, Tensor(a!) out0, Tensor(b!) out1, Tensor(c!) out2) -> (Tensor(a!), Tensor(b!), Tensor(c!))', + 'aten::_native_multi_head_attention(Tensor query, Tensor key, Tensor value, int embed_dim, int num_head, Tensor qkv_weight, Tensor qkv_bias, Tensor proj_weight, Tensor proj_bias, Tensor? mask=None, bool need_weights=True, bool average_attn_weights=True, int? mask_type=None) -> (Tensor, Tensor)', + 'aten::_native_multi_head_attention.out(Tensor query, Tensor key, Tensor value, int embed_dim, int num_head, Tensor qkv_weight, Tensor qkv_bias, Tensor proj_weight, Tensor proj_bias, Tensor? mask=None, bool need_weights=True, bool average_attn_weights=True, int? mask_type=None, *, Tensor(a!) out0, Tensor(b!) out1) -> (Tensor(a!), Tensor(b!))', + 'aten::add(Scalar a, Scalar b) -> Scalar', + 'aten::add.Scalar(Tensor self, Scalar other, Scalar alpha=1) -> Tensor', + 'aten::add.Scalar_out(Tensor self, Scalar other, Scalar alpha=1, *, Tensor(a!) out) -> Tensor(a!)', + 'aten::add.Tensor(Tensor self, Tensor other, *, Scalar alpha=1) -> Tensor', + 'aten::add.complex(complex a, complex b) -> complex', + 'aten::add.complex_float(complex a, float b) -> complex', + 'aten::add.complex_int(complex a, int b) -> complex', + 'aten::add.float(float a, float b) -> float', + 'aten::add.float_complex(float a, complex b) -> complex', + 'aten::add.float_int(float a, int b) -> float', + 'aten::add.int(int a, int b) -> int', + 'aten::add.int_complex(int a, complex b) -> complex', + 'aten::add.int_float(int a, float b) -> float', + 'aten::add.out(Tensor self, Tensor other, *, Scalar alpha=1, Tensor(a!) out) -> Tensor(a!)', + 'aten::add.str(str a, str b) -> str', + 'aten::add.t(t[] a, t[] b) -> t[]', + 'aten::add_.Scalar(Tensor(a!) self, Scalar other, Scalar alpha=1) -> Tensor(a!)', + 'aten::add_.Tensor(Tensor(a!) self, Tensor other, *, Scalar alpha=1) -> Tensor(a!)', + 'aten::add_.t(t[](a!) self, t[] b) -> t[]', 'aten::any.all_out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)', 'aten::any.bool(bool[] self) -> bool', 'aten::any.dim(Tensor self, int dim, bool keepdim=False) -> Tensor', - 'aten::any.dimname_out(Tensor self, str dim, bool keepdim=False, *, Tensor(a!) out) -> Tensor(a!)', # pylint: disable=line-too-long + 'aten::any.dimname_out(Tensor self, str dim, bool keepdim=False, *, Tensor(a!) out) -> Tensor(a!)', 'aten::any.dimname(Tensor self, str dim, bool keepdim=False) -> Tensor', - 'aten::any.dims_out(Tensor self, int[]? dim=None, bool keepdim=False, *, Tensor(a!) out) -> Tensor(a!)', # pylint: disable=line-too-long + 'aten::any.dims_out(Tensor self, int[]? dim=None, bool keepdim=False, *, Tensor(a!) out) -> Tensor(a!)', 'aten::any.dims(Tensor self, int[]? dim=None, bool keepdim=False) -> Tensor', 'aten::any.float(float[] self) -> bool', 'aten::any.int(int[] self) -> bool', @@ -94,20 +124,20 @@ def _write_metadata(value): 'aten::as_tensor.float(float t, *, ScalarType? dtype=None, Device? device=None) -> Tensor', 'aten::as_tensor.int(int t, *, ScalarType? dtype=None, Device? device=None) -> Tensor', 'aten::as_tensor.list(t[] data, *, ScalarType? dtype=None, Device? device=None) -> Tensor', - 'aten::as_tensor(Tensor(a) data, *, ScalarType? dtype=None, Device? device=None) -> Tensor(b|a)', # pylint: disable=line-too-long + 'aten::as_tensor(Tensor(a) data, *, ScalarType? dtype=None, Device? device=None) -> Tensor(b|a)', 'aten::bitwise_and.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)', 'aten::bitwise_and.Scalar_Tensor(Scalar self, Tensor other) -> Tensor', - 'aten::bitwise_and.Scalar_Tensor_out(Scalar self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', # pylint: disable=line-too-long + 'aten::bitwise_and.Scalar_Tensor_out(Scalar self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', 'aten::bitwise_and.Tensor(Tensor self, Tensor other) -> Tensor', 'aten::bitwise_and.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', 'aten::bitwise_and_.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)', 'aten::bitwise_and_.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)', 'aten::bitwise_left_shift.Scalar_Tensor(Scalar self, Tensor other) -> Tensor', - 'aten::bitwise_left_shift.Scalar_Tensor_out(Scalar self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', # pylint: disable=line-too-long + 'aten::bitwise_left_shift.Scalar_Tensor_out(Scalar self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', 'aten::bitwise_left_shift.Tensor(Tensor self, Tensor other) -> Tensor', - 'aten::bitwise_left_shift.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', # pylint: disable=line-too-long + 'aten::bitwise_left_shift.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', 'aten::bitwise_left_shift.Tensor_Scalar(Tensor self, Scalar other) -> Tensor', - 'aten::bitwise_left_shift.Tensor_Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)', # pylint: disable=line-too-long + 'aten::bitwise_left_shift.Tensor_Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)', 'aten::bitwise_left_shift_.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)', 'aten::bitwise_left_shift_.Tensor_Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)', 'aten::bitwise_not(Tensor self) -> Tensor', @@ -116,23 +146,23 @@ def _write_metadata(value): 'aten::bitwise_or.Scalar(Tensor self, Scalar other) -> Tensor', 'aten::bitwise_or.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)', 'aten::bitwise_or.Scalar_Tensor(Scalar self, Tensor other) -> Tensor', - 'aten::bitwise_or.Scalar_Tensor_out(Scalar self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', # pylint: disable=line-too-long + 'aten::bitwise_or.Scalar_Tensor_out(Scalar self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', 'aten::bitwise_or.Tensor(Tensor self, Tensor other) -> Tensor', 'aten::bitwise_or.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', 'aten::bitwise_or_.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)', 'aten::bitwise_or_.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)', 'aten::bitwise_right_shift.Scalar_Tensor(Scalar self, Tensor other) -> Tensor', - 'aten::bitwise_right_shift.Scalar_Tensor_out(Scalar self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', # pylint: disable=line-too-long + 'aten::bitwise_right_shift.Scalar_Tensor_out(Scalar self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', 'aten::bitwise_right_shift.Tensor(Tensor self, Tensor other) -> Tensor', - 'aten::bitwise_right_shift.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', # pylint: disable=line-too-long + 'aten::bitwise_right_shift.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', 'aten::bitwise_right_shift.Tensor_Scalar(Tensor self, Scalar other) -> Tensor', - 'aten::bitwise_right_shift.Tensor_Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)', # pylint: disable=line-too-long + 'aten::bitwise_right_shift.Tensor_Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)', 'aten::bitwise_right_shift_.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)', 'aten::bitwise_right_shift_.Tensor_Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)', 'aten::bitwise_xor.Scalar(Tensor self, Scalar other) -> Tensor', 'aten::bitwise_xor.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)', 'aten::bitwise_xor.Scalar_Tensor(Scalar self, Tensor other) -> Tensor', - 'aten::bitwise_xor.Scalar_Tensor_out(Scalar self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', # pylint: disable=line-too-long + 'aten::bitwise_xor.Scalar_Tensor_out(Scalar self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', 'aten::bitwise_xor.Tensor(Tensor self, Tensor other) -> Tensor', 'aten::bitwise_xor.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', 'aten::bitwise_xor_.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)', @@ -181,9 +211,9 @@ def _write_metadata(value): 'aten::div.complex(complex a, complex b) -> complex', 'aten::div.float(float a, float b) -> float', 'aten::div.int(int a, int b) -> float', - 'aten::div.out_mode(Tensor self, Tensor other, *, str? rounding_mode, Tensor(a!) out) -> Tensor(a!)', # pylint: disable=line-too-long + 'aten::div.out_mode(Tensor self, Tensor other, *, str? rounding_mode, Tensor(a!) out) -> Tensor(a!)', 'aten::div.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', - 'aten::div.Scalar_mode_out(Tensor self, Scalar other, *, str? rounding_mode, Tensor(a!) out) -> Tensor(a!)', # pylint: disable=line-too-long + 'aten::div.Scalar_mode_out(Tensor self, Scalar other, *, str? rounding_mode, Tensor(a!) out) -> Tensor(a!)', 'aten::div.Scalar_mode(Tensor self, Scalar other, *, str? rounding_mode) -> Tensor', 'aten::div.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)', 'aten::div.Scalar(Tensor self, Scalar other) -> Tensor', @@ -220,6 +250,38 @@ def _write_metadata(value): 'aten::Float.Scalar(Scalar a) -> float', 'aten::Float.str(str a) -> float', 'aten::Float.Tensor(Tensor a) -> float', + 'aten::floor(Tensor self) -> Tensor', + 'aten::floor.Scalar(Scalar a) -> Scalar', + 'aten::floor.float(float a) -> int', + 'aten::floor.int(int a) -> int', + 'aten::floor.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)', + 'aten::floor_(Tensor(a!) self) -> Tensor(a!)', + 'aten::floor_divide(Tensor self, Tensor other) -> Tensor', + 'aten::floor_divide.Scalar(Tensor self, Scalar other) -> Tensor', + 'aten::floor_divide.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)', + 'aten::floor_divide.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', + 'aten::floor_divide_.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)', + 'aten::floor_divide_.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)', + 'aten::floordiv(Scalar a, Scalar b) -> Scalar', + 'aten::floordiv.float(float a, float b) -> float', + 'aten::floordiv.float_int(float a, int b) -> float', + 'aten::floordiv.int(int a, int b) -> int', + 'aten::floordiv.int_float(int a, float b) -> float', + 'aten::fmax(Tensor self, Tensor other) -> Tensor', + 'aten::fmax.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', + 'aten::fmin(Tensor self, Tensor other) -> Tensor', + 'aten::fmin.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', + 'aten::fmod(Scalar a, Scalar b) -> float', + 'aten::fmod.Scalar(Tensor self, Scalar other) -> Tensor', + 'aten::fmod.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)', + 'aten::fmod.Tensor(Tensor self, Tensor other) -> Tensor', + 'aten::fmod.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', + 'aten::fmod.float(float a, float b) -> float', + 'aten::fmod.float_int(float a, int b) -> float', + 'aten::fmod.int(int a, int b) -> float', + 'aten::fmod.int_float(int a, float b) -> float', + 'aten::fmod_.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)', + 'aten::fmod_.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)', 'aten::get.bool(Dict(bool, t) self, bool key) -> t(*)?', 'aten::get.complex(Dict(complex, t) self, complex key) -> t(*)?', 'aten::get.default_bool(Dict(bool, t) self, bool key, t default_value) -> t(*)', @@ -280,10 +342,10 @@ def _write_metadata(value): 'aten::le.Tensor(Tensor self, Tensor other) -> Tensor', 'aten::le(Scalar a, Scalar b) -> bool', 'aten::leaky_relu(Tensor self, Scalar negative_slope=0.01) -> Tensor', - 'aten::leaky_relu.out(Tensor self, Scalar negative_slope=0.01, *, Tensor(a!) out) -> Tensor(a!)', # pylint: disable=line-too-long + 'aten::leaky_relu.out(Tensor self, Scalar negative_slope=0.01, *, Tensor(a!) out) -> Tensor(a!)', 'aten::leaky_relu_(Tensor(a!) self, Scalar negative_slope=0.01) -> Tensor(a!)', - 'aten::leaky_relu_backward(Tensor grad_output, Tensor self, Scalar negative_slope, bool self_is_result) -> Tensor', # pylint: disable=line-too-long - 'aten::leaky_relu_backward.grad_input(Tensor grad_output, Tensor self, Scalar negative_slope, bool self_is_result, *, Tensor(a!) grad_input) -> Tensor(a!)', # pylint: disable=line-too-long + 'aten::leaky_relu_backward(Tensor grad_output, Tensor self, Scalar negative_slope, bool self_is_result) -> Tensor', + 'aten::leaky_relu_backward.grad_input(Tensor grad_output, Tensor self, Scalar negative_slope, bool self_is_result, *, Tensor(a!) grad_input) -> Tensor(a!)', 'aten::len.any(Any[] a) -> int', 'aten::len.Dict_bool(Dict(bool, t) self) -> int', 'aten::len.Dict_complex(Dict(complex, t) self) -> int', @@ -295,9 +357,9 @@ def _write_metadata(value): 'aten::len.t(t[] a) -> int', 'aten::len.Tensor(Tensor t) -> int', 'aten::lerp.Scalar(Tensor self, Tensor end, Scalar weight) -> Tensor', - 'aten::lerp.Scalar_out(Tensor self, Tensor end, Scalar weight, *, Tensor(a!) out) -> Tensor(a!)', # pylint: disable=line-too-long + 'aten::lerp.Scalar_out(Tensor self, Tensor end, Scalar weight, *, Tensor(a!) out) -> Tensor(a!)', 'aten::lerp.Tensor(Tensor self, Tensor end, Tensor weight) -> Tensor', - 'aten::lerp.Tensor_out(Tensor self, Tensor end, Tensor weight, *, Tensor(a!) out) -> Tensor(a!)', # pylint: disable=line-too-long + 'aten::lerp.Tensor_out(Tensor self, Tensor end, Tensor weight, *, Tensor(a!) out) -> Tensor(a!)', 'aten::lerp_.Scalar(Tensor(a!) self, Tensor end, Scalar weight) -> Tensor(a!)', 'aten::lerp_.Tensor(Tensor(a!) self, Tensor end, Tensor weight) -> Tensor(a!)', 'aten::less.Scalar(Tensor self, Scalar other) -> Tensor', @@ -330,6 +392,58 @@ def _write_metadata(value): 'aten::lt.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', 'aten::lt.Tensor(Tensor self, Tensor other) -> Tensor', 'aten::lt(Scalar a, Scalar b) -> bool', + 'aten::mul(Scalar a, Scalar b) -> Scalar', + 'aten::mul.Scalar(Tensor self, Scalar other) -> Tensor', + 'aten::mul.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)', + 'aten::mul.Tensor(Tensor self, Tensor other) -> Tensor', + 'aten::mul.complex(complex a, complex b) -> complex', + 'aten::mul.complex_float(complex a, float b) -> complex', + 'aten::mul.complex_int(complex a, int b) -> complex', + 'aten::mul.float(float a, float b) -> float', + 'aten::mul.float_complex(float a, complex b) -> complex', + 'aten::mul.float_int(float a, int b) -> float', + 'aten::mul.int(int a, int b) -> int', + 'aten::mul.int_complex(int a, complex b) -> complex', + 'aten::mul.int_float(int a, float b) -> float', + 'aten::mul.left_t(t[] l, int n) -> t[]', + 'aten::mul.out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', + 'aten::mul.right_(int n, t[] l) -> t[]', + 'aten::mul_.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)', + 'aten::mul_.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)', + 'aten::mul_.t(t[](a!) l, int n) -> t[](a!)', + 'aten::ne(Scalar a, Scalar b) -> bool', + 'aten::ne.Scalar(Tensor self, Scalar other) -> Tensor', + 'aten::ne.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)', + 'aten::ne.Tensor(Tensor self, Tensor other) -> Tensor', + 'aten::ne.Tensor_list(Tensor[] a, Tensor[] b) -> bool', + 'aten::ne.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)', + 'aten::ne.bool(bool a, bool b) -> bool', + 'aten::ne.bool_list(bool[] a, bool[] b) -> bool', + 'aten::ne.complex(complex a, complex b) -> bool', + 'aten::ne.complex_float(complex a, float b) -> bool', + 'aten::ne.device(Device a, Device b) -> bool', + 'aten::ne.enum(AnyEnumType a, AnyEnumType b) -> bool', + 'aten::ne.float(float a, float b) -> bool', + 'aten::ne.float_complex(float a, complex b) -> bool', + 'aten::ne.float_int(float a, int b) -> bool', + 'aten::ne.float_list(float[] a, float[] b) -> bool', + 'aten::ne.int(int a, int b) -> bool', + 'aten::ne.int_float(int a, float b) -> bool', + 'aten::ne.int_list(int[] a, int[] b) -> bool', + 'aten::ne.str(str a, str b) -> bool', + 'aten::ne.str_list(str[] a, str[] b) -> bool', + 'aten::ne_.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)', + 'aten::ne_.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)', + 'aten::neg(Tensor self) -> Tensor', + 'aten::neg.Scalar(Scalar a) -> Scalar', + 'aten::neg.complex(complex a) -> complex', + 'aten::neg.float(float a) -> float', + 'aten::neg.int(int a) -> int', + 'aten::neg.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)', + 'aten::neg_(Tensor(a!) self) -> Tensor(a!)', + 'aten::negative(Tensor self) -> Tensor', + 'aten::negative.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)', + 'aten::negative_(Tensor(a!) self) -> Tensor(a!)', 'aten::pow.complex(complex a, complex b) -> complex', 'aten::pow.complex_float(complex a, float b) -> complex', 'aten::pow.float(float a, float b) -> float', @@ -360,10 +474,10 @@ def _write_metadata(value): 'aten::remainder(Scalar a, Scalar b) -> Scalar', 'aten::replace(str self, str old, str new, int max=-1) -> str', 'aten::ScalarImplicit(Tensor a) -> Scalar', - 'aten::searchsorted.Scalar_out(Tensor sorted_sequence, Scalar self, *, bool out_int32=False, bool right=False, str? side=None, Tensor? sorter=None, Tensor(a!) out) -> Tensor(a!)', # pylint: disable=line-too-long - 'aten::searchsorted.Scalar(Tensor sorted_sequence, Scalar self, *, bool out_int32=False, bool right=False, str? side=None, Tensor? sorter=None) -> Tensor', # pylint: disable=line-too-long - 'aten::searchsorted.Tensor_out(Tensor sorted_sequence, Tensor self, *, bool out_int32=False, bool right=False, str? side=None, Tensor? sorter=None, Tensor(a!) out) -> Tensor(a!)', # pylint: disable=line-too-long - 'aten::searchsorted.Tensor(Tensor sorted_sequence, Tensor self, *, bool out_int32=False, bool right=False, str? side=None, Tensor? sorter=None) -> Tensor', # pylint: disable=line-too-long + 'aten::searchsorted.Scalar_out(Tensor sorted_sequence, Scalar self, *, bool out_int32=False, bool right=False, str? side=None, Tensor? sorter=None, Tensor(a!) out) -> Tensor(a!)', + 'aten::searchsorted.Scalar(Tensor sorted_sequence, Scalar self, *, bool out_int32=False, bool right=False, str? side=None, Tensor? sorter=None) -> Tensor', + 'aten::searchsorted.Tensor_out(Tensor sorted_sequence, Tensor self, *, bool out_int32=False, bool right=False, str? side=None, Tensor? sorter=None, Tensor(a!) out) -> Tensor(a!)', + 'aten::searchsorted.Tensor(Tensor sorted_sequence, Tensor self, *, bool out_int32=False, bool right=False, str? side=None, Tensor? sorter=None) -> Tensor', 'aten::sqrt.complex(complex a) -> complex', 'aten::sqrt.float(float a) -> float', 'aten::sqrt.int(int a) -> float', @@ -371,6 +485,48 @@ def _write_metadata(value): 'aten::sqrt.Scalar(Scalar a) -> Scalar', 'aten::sqrt(Tensor self) -> Tensor', 'aten::str(t elem) -> str', + 'aten::sub(Scalar a, Scalar b) -> Scalar', + 'aten::sub.Scalar(Tensor self, Scalar other, Scalar alpha=1) -> Tensor', + 'aten::sub.Scalar_out(Tensor self, Scalar other, Scalar alpha=1, *, Tensor(a!) out) -> Tensor(a!)', + 'aten::sub.Tensor(Tensor self, Tensor other, *, Scalar alpha=1) -> Tensor', + 'aten::sub.complex(complex a, complex b) -> complex', + 'aten::sub.complex_float(complex a, float b) -> complex', + 'aten::sub.complex_int(complex a, int b) -> complex', + 'aten::sub.float(float a, float b) -> float', + 'aten::sub.float_complex(float a, complex b) -> complex', + 'aten::sub.float_int(float a, int b) -> float', + 'aten::sub.int(int a, int b) -> int', + 'aten::sub.int_complex(int a, complex b) -> complex', + 'aten::sub.int_float(int a, float b) -> float', + 'aten::sub.out(Tensor self, Tensor other, *, Scalar alpha=1, Tensor(a!) out) -> Tensor(a!)', + 'aten::sub_.Scalar(Tensor(a!) self, Scalar other, Scalar alpha=1) -> Tensor(a!)', + 'aten::sub_.Tensor(Tensor(a!) self, Tensor other, *, Scalar alpha=1) -> Tensor(a!)', + 'aten::subtract.Scalar(Tensor self, Scalar other, Scalar alpha=1) -> Tensor', + 'aten::subtract.Tensor(Tensor self, Tensor other, *, Scalar alpha=1) -> Tensor', + 'aten::subtract.out(Tensor self, Tensor other, *, Scalar alpha=1, Tensor(a!) out) -> Tensor(a!)', + 'aten::subtract_.Scalar(Tensor(a!) self, Scalar other, Scalar alpha=1) -> Tensor(a!)', + 'aten::subtract_.Tensor(Tensor(a!) self, Tensor other, *, Scalar alpha=1) -> Tensor(a!)', + 'aten::sum(Tensor self, *, ScalarType? dtype=None) -> Tensor', + 'aten::sum.DimnameList_out(Tensor self, str[1] dim, bool keepdim=False, *, ScalarType? dtype=None, Tensor(a!) out) -> Tensor(a!)', + 'aten::sum.IntList_out(Tensor self, int[1]? dim, bool keepdim=False, *, ScalarType? dtype=None, Tensor(a!) out) -> Tensor(a!)', + 'aten::sum.bool(bool[] self) -> int', + 'aten::sum.complex(complex[] self) -> complex', + 'aten::sum.dim_DimnameList(Tensor self, str[1] dim, bool keepdim=False, *, ScalarType? dtype=None) -> Tensor', + 'aten::sum.dim_IntList(Tensor self, int[1]? dim, bool keepdim=False, *, ScalarType? dtype=None) -> Tensor', + 'aten::sum.float(float[] self) -> float', + 'aten::sum.int(int[] self) -> int', + 'aten::sum.out(Tensor self, *, ScalarType? dtype=None, Tensor(a!) out) -> Tensor(a!)', + 'aten::sum_to_size(Tensor self, SymInt[] size) -> Tensor', + 'aten::tensor(t[] data, *, ScalarType? dtype=None, Device? device=None, bool requires_grad=False) -> Tensor', + 'aten::tensor.bool(bool t, *, ScalarType? dtype=None, Device? device=None, bool requires_grad=False) -> Tensor', + 'aten::tensor.complex(complex t, *, ScalarType? dtype=None, Device? device=None, bool requires_grad=False) -> Tensor', + 'aten::tensor.float(float t, *, ScalarType? dtype=None, Device? device=None, bool requires_grad=False) -> Tensor', + 'aten::tensor.int(int t, *, ScalarType? dtype=None, Device? device=None, bool requires_grad=False) -> Tensor', + 'aten::tensor_split.indices(Tensor(a -> *) self, SymInt[] indices, int dim=0) -> Tensor(a)[]', + 'aten::tensor_split.sections(Tensor(a -> *) self, SymInt sections, int dim=0) -> Tensor(a)[]', + 'aten::tensor_split.tensor_indices_or_sections(Tensor(a -> *) self, Tensor tensor_indices_or_sections, int dim=0) -> Tensor(a)[]', + 'aten::tensordot(Tensor self, Tensor other, int[] dims_self, int[] dims_other) -> Tensor', + 'aten::tensordot.out(Tensor self, Tensor other, int[] dims_self, int[] dims_other, *, Tensor(a!) out) -> Tensor(a!)', 'aten::values.bool(Dict(bool, t) self) -> t[](*)', 'aten::values.complex(Dict(complex, t) self) -> t[](*)', 'aten::values.float(Dict(float, t) self) -> t[](*)', @@ -425,6 +581,29 @@ def _write_metadata(value): 'prim::min(Scalar a, Scalar b) -> Scalar', ] +known_legacy_schema_definitions = [ + '_caffe2::BBoxTransform(Tensor rois, Tensor deltas, Tensor im_info, float[] weights, bool apply_scale, bool rotated, bool angle_bound_on, int angle_bound_lo, int angle_bound_hi, float clip_angle_thresh, bool legacy_plus_one) -> (Tensor output_0, Tensor output_1)', + '_caffe2::BatchPermutation(Tensor X, Tensor indices) -> Tensor', + '_caffe2::BoxWithNMSLimit(Tensor scores, Tensor boxes, Tensor batch_splits, float score_thresh, float nms, int detections_per_im, bool soft_nms_enabled, str soft_nms_method, float soft_nms_sigma, float soft_nms_min_score_thres, bool rotated, bool cls_agnostic_bbox_reg, bool input_boxes_include_bg_cls, bool output_classes_include_bg_cls, bool legacy_plus_one) -> (Tensor scores, Tensor boxes, Tensor classes, Tensor batch_splits, Tensor keeps, Tensor keeps_size)', + '_caffe2::CollectAndDistributeFpnRpnProposals(Tensor[] input_list, int roi_canonical_scale, int roi_canonical_level, int roi_max_level, int roi_min_level, int rpn_max_level, int rpn_min_level, int rpn_post_nms_topN, bool legacy_plus_one) -> (Tensor rois, Tensor rois_fpn2, Tensor rois_fpn3, Tensor rois_fpn4, Tensor rois_fpn5, Tensor rois_idx_restore_int32)', + '_caffe2::CollectRpnProposals(Tensor[] input_list, int rpn_max_level, int rpn_min_level, int rpn_post_nms_topN) -> (Tensor rois)', + '_caffe2::CopyCPUToGPU(Tensor input) -> Tensor', + '_caffe2::CopyGPUToCPU(Tensor input) -> Tensor', + '_caffe2::DistributeFpnProposals(Tensor rois, int roi_canonical_scale, int roi_canonical_level, int roi_max_level, int roi_min_level, bool legacy_plus_one) -> (Tensor rois_fpn2, Tensor rois_fpn3, Tensor rois_fpn4, Tensor rois_fpn5, Tensor rois_idx_restore_int32)', + '_caffe2::GenerateProposals(Tensor scores, Tensor bbox_deltas, Tensor im_info, Tensor anchors, float spatial_scale, int pre_nms_topN, int post_nms_topN, float nms_thresh, float min_size, bool angle_bound_on, int angle_bound_lo, int angle_bound_hi, float clip_angle_thresh, bool legacy_plus_one) -> (Tensor output_0, Tensor output_1)', + '_caffe2::RoIAlign(Tensor features, Tensor rois, str order, float spatial_scale, int pooled_h, int pooled_w, int sampling_ratio, bool aligned) -> Tensor', + 'aten::arange.start_out_(Scalar start, Scalar end) -> Tensor', + 'aten::fft(Tensor self, int signal_ndim, bool normalized=False) -> Tensor', + 'aten::grid_sampler.legacy(Tensor input, Tensor grid, int interpolation_mode, int padding_mode) -> Tensor', + 'neuron::forward_v2_1(Tensor[] _0, __torch__.torch.classes.neuron.Model _1) -> (Tensor _0)', + 'prim::shape(Tensor self) -> int[]', + 'torchaudio::sox_effects_apply_effects_tensor(Tensor tensor, int sample_rate, str[][] effects, bool channels_first=True) -> (Tensor, int64)', + 'torchvision::nms(Tensor dets, Tensor scores, float iou_threshold) -> Tensor', + 'torchvision::roi_align(Tensor input, Tensor rois, float spatial_scale, int pooled_height, int pooled_width, int sampling_ratio, bool aligned) -> Tensor', +] + +# pylint: enable=line-too-long + def _parse_schemas(): schemas = {} definitions = set() @@ -443,8 +622,11 @@ def _parse_schemas(): if schema.name in schemas: raise KeyError(schema.name) schemas[schema.name] = schema - for definition in known_schema_definitions: - schema = pytorch.Schema(definition) + for value in known_legacy_schema_definitions: + schema = pytorch.Schema(value) + schemas[schema.name] = schema + for value in known_schema_definitions: + schema = pytorch.Schema(value) schemas[schema.name] = schema return schemas @@ -455,9 +637,9 @@ def _filter_schemas(schemas, types): for key in keys: if schema.name == key or schema.name.startswith(key + '.'): filtered_schemas.add(schema.name) - for schema in schemas.values(): - if schema.name.startswith('aten::pop'): - filtered_schemas.add(schema.name) + # for schema in schemas.values(): + # if schema.name.startswith('aten::pop'): + # filtered_schemas.add(schema.name) # filtered_schemas = set(types.keys()) # content = _read('list.csv') # regex = re.compile(r'Unsupported function \'(.*)\' in', re.MULTILINE) @@ -504,41 +686,7 @@ def _check_types(types, schemas): if key.startswith('_caffe2::'): types.pop(key) known_keys = [ - 'aten::_native_batch_norm_legit_functional', - 'aten::add.float', - 'aten::add.int', - 'aten::add.str', - 'aten::arange.start_out_', - 'aten::classes._nnapi.Compilation', - 'aten::fft', - 'aten::floor.float', - 'aten::floor.int', - 'aten::floor.Scalar', - 'aten::floordiv.float_int', - 'aten::floordiv.float', - 'aten::floordiv.int_float', - 'aten::floordiv.int', - 'aten::floordiv.Scalar', - 'aten::grid_sampler.legacy', - 'aten::mul.float_int', - 'aten::mul.int_float', - 'aten::mul.int', - 'aten::mul.ScalarT', - 'aten::mul', - 'aten::ne.float', - 'aten::ne.int', - 'aten::ne.str', - 'aten::neg.complex', - 'aten::neg.float', - 'aten::neg.int', - 'aten::neg.Scalar', - 'aten::sub.float', - 'aten::sub.int', - 'aten::sub.str', - 'aten::tensor.bool', - 'aten::tensor.float', - 'aten::tensor.int', - 'prim::shape', + 'aten::classes._nnapi.Compilation' ] for key in known_keys: types.pop(key) @@ -553,7 +701,8 @@ def _metadata(): filtered_schemas = _filter_schemas(schemas, types) metadata = pytorch.Metadata(types) for schema in filtered_schemas.values(): - metadata.type(schema) + value = metadata.type(schema) + value['name'] = schema.value _write_metadata(types) def main(): # pylint: disable=missing-function-docstring