@@ -128,7 +128,6 @@ def print_memory(stage):
128
128
import os
129
129
import psutil
130
130
logger .info (f"dlrmv2-memory-usage-log: { time .time ()} , { stage } , { psutil .Process (os .getpid ()).memory_info ().rss / 1024 / 1024 / 1024 } " )
131
- # print("dlrmv2-memory-usage-log: ", time.time(), stage, psutil.Process(os.getpid()).memory_info().rss / 1024 / 1024 / 1024)
132
131
133
132
def fetch_batch (dataloader ):
134
133
try :
@@ -210,7 +209,6 @@ def convert_int8(args, model, dataloader):
210
209
print_memory ("int8 jit optimize" )
211
210
model (batch .dense_features , batch .sparse_features )
212
211
model (batch .dense_features , batch .sparse_features )
213
- # print(model.graph_for(batch.dense_features, batch.sparse_features))
214
212
return model
215
213
216
214
def ipex_optimize (args , model , optimizer , dataloader ):
@@ -382,7 +380,6 @@ def aoti_benchmark_compile(ninstances, nbatches, bs, tmp_dir, target_dir):
382
380
}
383
381
"""
384
382
)
385
- # os.system(f"cp {tmp_dir}/model.so {target_dir}/model.so")
386
383
os .system (f"ln -s { tmp_dir } /model.so { target_dir } /model.so" )
387
384
os .system (f"cp { tmp_dir } /inputs.pt { target_dir } /inputs.pt" )
388
385
model_dir = f"{ target_dir } /model.so"
@@ -474,9 +471,6 @@ def stock_pt_optimize(args, model, optimizer, dataloader):
474
471
prepared_model (dense , sparse )
475
472
converted_model = convert_pt2e (prepared_model )
476
473
torch .ao .quantization .move_exported_model_to_eval (converted_model )
477
- # print(converted_model.graph)
478
- # print("===========================")
479
- # converted_model.graph.print_tabular()
480
474
if args .ipex :
481
475
print ('[Info] Running torch.compile() with IPEX backend' )
482
476
model (dense , sparse )
0 commit comments