Skip to content

Commit

Permalink
[GH-PAGES] Updated website
Browse files Browse the repository at this point in the history
  • Loading branch information
gh-actions-bot authored and gh-actions-bot committed Sep 7, 2024
1 parent 3efb63a commit b8d4b68
Show file tree
Hide file tree
Showing 192 changed files with 1,050 additions and 426 deletions.
Binary file modified main/.doctrees/environment.pickle
Binary file not shown.
Binary file modified main/.doctrees/getting-started/tutorials/01-vector-add.doctree
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file modified main/.doctrees/getting-started/tutorials/05-layer-norm.doctree
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file modified main/.doctrees/getting-started/tutorials/08-grouped-gemm.doctree
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file modified main/.doctrees/index.doctree
Binary file not shown.
Binary file not shown.
Binary file modified main/.doctrees/sg_execution_times.doctree
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file modified main/_images/sphx_glr_01-vector-add_001.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file modified main/_images/sphx_glr_01-vector-add_thumb.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file modified main/_images/sphx_glr_02-fused-softmax_001.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file modified main/_images/sphx_glr_02-fused-softmax_thumb.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file modified main/_images/sphx_glr_03-matrix-multiplication_001.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file modified main/_images/sphx_glr_03-matrix-multiplication_002.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file modified main/_images/sphx_glr_03-matrix-multiplication_thumb.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file modified main/_images/sphx_glr_05-layer-norm_001.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file modified main/_images/sphx_glr_05-layer-norm_thumb.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file modified main/_images/sphx_glr_06-fused-attention_001.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file modified main/_images/sphx_glr_06-fused-attention_002.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file modified main/_images/sphx_glr_06-fused-attention_003.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file modified main/_images/sphx_glr_06-fused-attention_thumb.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file modified main/_images/sphx_glr_08-grouped-gemm_001.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file modified main/_images/sphx_glr_08-grouped-gemm_thumb.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
12 changes: 6 additions & 6 deletions main/_sources/getting-started/tutorials/01-vector-add.rst.txt
Original file line number Diff line number Diff line change
Expand Up @@ -230,30 +230,30 @@ We can now run the decorated function above. Pass `print_data=True` to see the p

vector-add-performance:
size Triton Torch
0 4096.0 8.000000 9.600000
1 8192.0 19.200000 15.999999
0 4096.0 8.000000 8.000000
1 8192.0 15.999999 15.999999
2 16384.0 31.999999 31.999999
3 32768.0 63.999998 63.999998
4 65536.0 127.999995 127.999995
5 131072.0 219.428568 219.428568
6 262144.0 384.000001 384.000001
7 524288.0 614.400016 614.400016
8 1048576.0 819.200021 819.200021
9 2097152.0 1068.521715 1023.999964
10 4194304.0 1260.307736 1228.800031
9 2097152.0 1023.999964 1023.999964
10 4194304.0 1228.800031 1228.800031
11 8388608.0 1424.695621 1404.342820
12 16777216.0 1560.380965 1560.380965
13 33554432.0 1631.601649 1624.859540
14 67108864.0 1669.706983 1662.646960
15 134217728.0 1684.008546 1680.410210
15 134217728.0 1684.008546 1678.616907





.. rst-class:: sphx-glr-timing

**Total running time of the script:** (0 minutes 9.279 seconds)
**Total running time of the script:** (0 minutes 6.723 seconds)


.. _sphx_glr_download_getting-started_tutorials_01-vector-add.py:
Expand Down
198 changes: 99 additions & 99 deletions main/_sources/getting-started/tutorials/02-fused-softmax.rst.txt
Original file line number Diff line number Diff line change
Expand Up @@ -330,104 +330,104 @@ We will then compare its performance against (1) :code:`torch.softmax` and (2) t
softmax-performance:
N Triton Torch
0 256.0 481.787233 705.724062
1 384.0 609.399830 823.145287
2 512.0 752.761241 938.715075
3 640.0 795.207191 963.687580
4 768.0 883.995555 1018.055429
5 896.0 941.363201 1065.589231
6 1024.0 996.608845 1125.695241
7 1152.0 1103.490588 611.122499
8 1280.0 1142.952994 665.994091
9 1408.0 1155.953111 725.551440
10 1536.0 1197.167063 783.601216
11 1664.0 1210.498079 816.986410
12 1792.0 1234.906613 856.984714
13 1920.0 1248.669602 904.797934
14 2048.0 1280.858942 958.774182
15 2176.0 1265.941230 975.427679
16 2304.0 1266.526337 1012.759684
17 2432.0 1294.710144 1052.632580
18 2560.0 1308.203417 1083.349826
19 2688.0 1318.264003 1100.249220
20 2816.0 1324.348610 1133.009231
21 2944.0 1327.509189 1167.599136
22 3072.0 1353.069577 1187.693105
23 3200.0 1356.729265 1193.932573
24 3328.0 1357.470861 1223.652415
25 3456.0 1373.285487 1246.906963
26 3584.0 1381.106364 1257.079873
27 3712.0 1381.131247 1271.832414
28 3840.0 1390.550268 1302.754271
29 3968.0 1388.395296 1312.913114
30 4096.0 1397.371451 1326.814149
31 4224.0 1331.338224 1157.624645
32 4352.0 1335.823557 1176.092779
33 4480.0 1353.712464 1180.856354
34 4608.0 1360.907820 1194.417818
35 4736.0 1356.910643 1198.416683
36 4864.0 1378.410223 1220.112777
37 4992.0 1367.971261 1234.844611
38 5120.0 1378.023837 1254.763490
39 5248.0 1382.237297 1256.700207
40 5376.0 1378.864400 1284.430985
41 5504.0 1381.684706 1296.449600
42 5632.0 1387.100673 1317.723281
43 5760.0 1393.815076 1326.507313
44 5888.0 1391.070778 1345.476948
45 6016.0 1400.424754 1356.097935
46 6144.0 1405.875576 1376.014650
47 6272.0 1412.176572 1375.807404
48 6400.0 1416.333048 1386.887641
49 6528.0 1416.098886 1393.563170
50 6656.0 1421.801245 1402.747032
51 6784.0 1410.664260 1414.210188
52 6912.0 1431.320343 1422.900040
53 7040.0 1424.599719 1432.120361
54 7168.0 1430.183890 1433.223061
55 7296.0 1431.681316 1443.362338
56 7424.0 1432.122579 1448.012865
57 7552.0 1425.955935 1451.492379
58 7680.0 1435.178599 1458.149471
59 7808.0 1436.457737 1465.244224
60 7936.0 1433.572496 1469.084612
61 8064.0 1435.587133 1472.240271
62 8192.0 1441.404069 1486.112245
63 8320.0 1390.356704 1400.074679
64 8448.0 1383.194856 1406.393498
65 8576.0 1396.077284 1397.507024
66 8704.0 1384.904001 1398.360599
67 8832.0 1382.374048 1402.216758
68 8960.0 1401.063692 1414.321424
69 9088.0 1412.011352 1418.236176
70 9216.0 1404.785396 1423.952059
71 9344.0 1404.486259 1425.256286
72 9472.0 1397.758869 1434.454872
73 9600.0 1396.091948 1436.841632
74 9728.0 1400.572799 1441.794405
75 9856.0 1414.065249 1441.533230
76 9984.0 1401.525312 1451.348798
77 10112.0 1412.226149 1458.403595
78 10240.0 1421.336225 1467.180651
79 10368.0 1412.411351 1463.230173
80 10496.0 1411.883268 1466.183783
81 10624.0 1408.763024 1468.983152
82 10752.0 1406.810485 1471.994477
83 10880.0 1400.275815 1479.533197
84 11008.0 1419.242557 1480.439116
85 11136.0 1422.770073 1485.121957
86 11264.0 1427.136672 1488.408851
87 11392.0 1416.667586 1492.681639
88 11520.0 1422.613285 1495.312730
89 11648.0 1424.640271 1497.273837
90 11776.0 1433.258123 1500.081728
91 11904.0 1442.979869 1505.822134
92 12032.0 1421.861382 1510.145765
93 12160.0 1421.744228 1510.685070
94 12288.0 1439.575488 1392.381115
95 12416.0 1449.718932 1391.712238
96 12544.0 1442.290831 1393.949866
97 12672.0 1446.398421 1392.661624
0 256.0 479.591087 686.780777
1 384.0 606.329385 811.084783
2 512.0 759.148903 917.236325
3 640.0 784.606674 963.285389
4 768.0 878.976622 1012.756900
5 896.0 936.982431 1067.047583
6 1024.0 993.348041 1109.168311
7 1152.0 1102.971819 613.547884
8 1280.0 1138.184560 671.634363
9 1408.0 1155.897226 724.090333
10 1536.0 1191.316485 778.447155
11 1664.0 1216.788990 809.628678
12 1792.0 1238.494639 858.334953
13 1920.0 1253.417042 908.037302
14 2048.0 1269.259338 953.436477
15 2176.0 1256.971745 975.439013
16 2304.0 1273.536301 1011.647880
17 2432.0 1293.589320 1052.346948
18 2560.0 1300.039855 1087.310601
19 2688.0 1313.709434 1098.697721
20 2816.0 1321.950869 1127.038728
21 2944.0 1321.396901 1167.914567
22 3072.0 1352.850905 1180.985293
23 3200.0 1350.322541 1194.869292
24 3328.0 1356.696780 1220.645876
25 3456.0 1370.160685 1247.851511
26 3584.0 1376.749220 1261.037051
27 3712.0 1387.322807 1265.337871
28 3840.0 1384.630822 1298.935824
29 3968.0 1392.280764 1312.133859
30 4096.0 1395.284741 1327.113740
31 4224.0 1333.870054 1157.573503
32 4352.0 1334.753963 1177.066793
33 4480.0 1354.228198 1183.200907
34 4608.0 1361.654589 1196.523793
35 4736.0 1356.343469 1199.647767
36 4864.0 1376.871385 1222.899944
37 4992.0 1368.337785 1238.186815
38 5120.0 1375.114021 1252.122355
39 5248.0 1371.741881 1254.659695
40 5376.0 1377.132400 1286.350160
41 5504.0 1374.163119 1296.429048
42 5632.0 1383.639491 1315.019705
43 5760.0 1394.006823 1323.923430
44 5888.0 1395.175890 1339.341988
45 6016.0 1398.239767 1356.552649
46 6144.0 1407.111343 1375.300442
47 6272.0 1414.286270 1376.879531
48 6400.0 1416.124408 1390.245667
49 6528.0 1413.135628 1396.274773
50 6656.0 1421.462948 1400.787411
51 6784.0 1405.418965 1413.893064
52 6912.0 1429.922242 1425.265663
53 7040.0 1420.268859 1431.970540
54 7168.0 1426.904234 1436.504632
55 7296.0 1428.848471 1439.903310
56 7424.0 1429.889980 1442.603892
57 7552.0 1427.233875 1454.240622
58 7680.0 1433.454680 1460.214321
59 7808.0 1429.409941 1466.211424
60 7936.0 1439.817454 1466.702774
61 8064.0 1435.910645 1471.880026
62 8192.0 1440.488795 1482.469613
63 8320.0 1385.196502 1403.558392
64 8448.0 1381.196296 1405.175146
65 8576.0 1394.654387 1392.750587
66 8704.0 1389.283047 1400.915735
67 8832.0 1381.447123 1404.645393
68 8960.0 1400.210307 1411.632670
69 9088.0 1409.877714 1415.073481
70 9216.0 1406.550586 1426.026518
71 9344.0 1398.570404 1423.363447
72 9472.0 1394.784036 1431.838719
73 9600.0 1396.219370 1432.961948
74 9728.0 1400.115561 1443.202285
75 9856.0 1413.348391 1442.315923
76 9984.0 1401.335040 1451.377582
77 10112.0 1412.219870 1452.109506
78 10240.0 1420.010313 1467.536831
79 10368.0 1413.298436 1463.078903
80 10496.0 1413.837933 1469.054667
81 10624.0 1410.845223 1465.886907
82 10752.0 1401.779613 1470.154689
83 10880.0 1399.075895 1483.569163
84 11008.0 1415.474194 1475.722004
85 11136.0 1422.763130 1482.443433
86 11264.0 1428.716004 1486.668993
87 11392.0 1416.737929 1486.493588
88 11520.0 1421.482944 1496.738310
89 11648.0 1423.347027 1500.292973
90 11776.0 1430.859961 1501.073525
91 11904.0 1443.636709 1505.693946
92 12032.0 1429.536656 1508.197570
93 12160.0 1420.532604 1508.366363
94 12288.0 1439.085584 1392.058470
95 12416.0 1446.501558 1390.692311
96 12544.0 1441.391276 1391.394575
97 12672.0 1446.591091 1393.094678
Expand All @@ -442,7 +442,7 @@ In the above plot, we can see that:

.. rst-class:: sphx-glr-timing

**Total running time of the script:** (0 minutes 23.201 seconds)
**Total running time of the script:** (0 minutes 23.295 seconds)


.. _sphx_glr_download_getting-started_tutorials_02-fused-softmax.py:
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -573,33 +573,33 @@ but feel free to arrange this script as you wish to benchmark any other matrix s
1 384.0 384.0 384.0 12.288000 12.288000
2 512.0 512.0 512.0 26.214401 26.214401
3 640.0 640.0 640.0 42.666665 42.666665
4 768.0 768.0 768.0 63.195428 68.056616
4 768.0 768.0 768.0 63.195428 63.195428
5 896.0 896.0 896.0 78.051553 87.808000
6 1024.0 1024.0 1024.0 104.857603 99.864382
6 1024.0 1024.0 1024.0 110.376426 99.864382
7 1152.0 1152.0 1152.0 135.726544 129.825388
8 1280.0 1280.0 1280.0 157.538463 163.840004
9 1408.0 1408.0 1408.0 155.765024 132.970149
10 1536.0 1536.0 1536.0 181.484314 157.286398
10 1536.0 1536.0 1536.0 176.947204 157.286398
11 1664.0 1664.0 1664.0 183.651271 179.978245
12 1792.0 1792.0 1792.0 172.914215 208.137481
13 1920.0 1920.0 1920.0 200.347822 168.585369
12 1792.0 1792.0 1792.0 172.914215 204.353162
13 1920.0 1920.0 1920.0 200.347822 166.554219
14 2048.0 2048.0 2048.0 226.719125 192.841562
15 2176.0 2176.0 2176.0 209.621326 214.081356
16 2304.0 2304.0 2304.0 231.921091 231.921091
17 2432.0 2432.0 2432.0 206.576938 202.118452
18 2560.0 2560.0 2560.0 224.438347 219.919464
19 2688.0 2688.0 2688.0 200.704002 201.771569
20 2816.0 2816.0 2816.0 214.848312 213.795141
21 2944.0 2944.0 2944.0 223.479969 224.486628
22 3072.0 3072.0 3072.0 212.071554 212.868821
23 3200.0 3200.0 3200.0 217.687077 216.949149
15 2176.0 2176.0 2176.0 211.827867 211.827867
16 2304.0 2304.0 2304.0 231.921091 229.691080
17 2432.0 2432.0 2432.0 205.069087 205.069087
18 2560.0 2560.0 2560.0 224.438347 222.911566
19 2688.0 2688.0 2688.0 199.647657 200.704002
20 2816.0 2816.0 2816.0 212.752230 210.696652
21 2944.0 2944.0 2944.0 221.493479 223.479969
22 3072.0 3072.0 3072.0 208.941345 212.868821
23 3200.0 3200.0 3200.0 213.333323 220.689658
24 3328.0 3328.0 3328.0 209.277023 209.887165
25 3456.0 3456.0 3456.0 221.487820 220.880999
26 3584.0 3584.0 3584.0 215.624440 215.624440
27 3712.0 3712.0 3712.0 213.000737 218.116474
28 3840.0 3840.0 3840.0 210.250955 209.851994
29 3968.0 3968.0 3968.0 212.585252 217.511464
30 4096.0 4096.0 4096.0 222.214781 221.847481
25 3456.0 3456.0 3456.0 214.419058 220.880999
26 3584.0 3584.0 3584.0 215.624440 213.069643
27 3712.0 3712.0 3712.0 210.310194 217.168134
28 3840.0 3840.0 3840.0 209.851994 209.454544
29 3968.0 3968.0 3968.0 210.749463 217.511464
30 4096.0 4096.0 4096.0 219.668951 220.029067
matmul-performance-fp8:
M N K Triton
0 256.0 256.0 256.0 3.276800
Expand All @@ -608,39 +608,39 @@ but feel free to arrange this script as you wish to benchmark any other matrix s
3 640.0 640.0 640.0 32.000000
4 768.0 768.0 768.0 42.130286
5 896.0 896.0 896.0 58.538665
6 1024.0 1024.0 1024.0 63.550060
6 1024.0 1024.0 1024.0 61.680940
7 1152.0 1152.0 1152.0 80.702267
8 1280.0 1280.0 1280.0 102.400003
8 1280.0 1280.0 1280.0 99.902441
9 1408.0 1408.0 1408.0 81.369790
10 1536.0 1536.0 1536.0 98.303997
11 1664.0 1664.0 1664.0 115.370671
12 1792.0 1792.0 1792.0 133.802668
13 1920.0 1920.0 1920.0 100.173911
14 2048.0 2048.0 2048.0 114.130722
15 2176.0 2176.0 2176.0 120.500882
16 2304.0 2304.0 2304.0 134.201527
17 2432.0 2432.0 2432.0 133.149115
18 2560.0 2560.0 2560.0 146.941707
19 2688.0 2688.0 2688.0 117.804519
20 2816.0 2816.0 2816.0 128.277083
16 2304.0 2304.0 2304.0 134.959733
17 2432.0 2432.0 2432.0 132.521057
18 2560.0 2560.0 2560.0 146.285712
19 2688.0 2688.0 2688.0 117.439807
20 2816.0 2816.0 2816.0 128.655484
21 2944.0 2944.0 2944.0 139.988852
22 3072.0 3072.0 3072.0 143.713461
23 3200.0 3200.0 3200.0 140.350874
22 3072.0 3072.0 3072.0 144.079147
23 3200.0 3200.0 3200.0 139.737993
24 3328.0 3328.0 3328.0 131.852184
25 3456.0 3456.0 3456.0 139.725414
26 3584.0 3584.0 3584.0 149.113421
27 3712.0 3712.0 3712.0 142.000489
26 3584.0 3584.0 3584.0 148.620481
27 3712.0 3712.0 3712.0 141.698358
28 3840.0 3840.0 3840.0 137.895263
29 3968.0 3968.0 3968.0 147.016795
30 4096.0 4096.0 4096.0 156.796411
30 4096.0 4096.0 4096.0 154.985826
.. rst-class:: sphx-glr-timing

**Total running time of the script:** (2 minutes 17.657 seconds)
**Total running time of the script:** (2 minutes 17.342 seconds)


.. _sphx_glr_download_getting-started_tutorials_03-matrix-multiplication.py:
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -114,6 +114,8 @@ Let's first take a look at the baseline implementation.

.. code-block:: none
/home/runner/_work/triton/triton/python/triton/language/semantic.py:1502: UserWarning: tl.where with a non-boolean condition is deprecated and will error out in a future triton release. Got int32
warnings.warn(
--------- ------- --------- -------- -------- -------- -------- -------- -------- --------- ---------
input 1.541 -0.293429 -2.17879 0.568431 -1.08452 -1.3986 0.403347 0.838026 -0.719258 -0.403344
keep mask 1 1 0 1 0 1 1 0 0 0
Expand Down Expand Up @@ -242,7 +244,7 @@ References
.. rst-class:: sphx-glr-timing

**Total running time of the script:** (0 minutes 0.730 seconds)
**Total running time of the script:** (0 minutes 0.703 seconds)


.. _sphx_glr_download_getting-started_tutorials_04-low-memory-dropout.py:
Expand Down
Loading

0 comments on commit b8d4b68

Please sign in to comment.