File tree Expand file tree Collapse file tree 2 files changed +11
-8
lines changed Expand file tree Collapse file tree 2 files changed +11
-8
lines changed Original file line number Diff line number Diff line change @@ -571,9 +571,8 @@ def do_nothing(max_batch_size, max_seq_length):
571
571
# attributes will NOT be seen on by AOTI-compiled forward
572
572
# function, e.g. calling model.setup_cache will NOT touch
573
573
# AOTI compiled and maintained model buffers such as kv_cache.
574
- from torch ._inductor .package import load_package
575
574
576
- aoti_compiled_model = load_package (
575
+ aoti_compiled_model = torch . _inductor . aoti_load_package (
577
576
str (builder_args .aoti_package_path .absolute ())
578
577
)
579
578
Original file line number Diff line number Diff line change 8
8
from typing import Optional
9
9
10
10
import torch
11
+ import torch ._inductor
11
12
import torch .nn as nn
12
13
13
14
from torch .export import Dim
14
- import torch ._inductor
15
15
16
16
from torchchat .cli .builder import (
17
17
_initialize_model ,
@@ -68,20 +68,24 @@ def export_for_server(
68
68
69
69
with torch .nn .attention .sdpa_kernel ([torch .nn .attention .SDPBackend .MATH ]):
70
70
metadata = {} # TODO: put more metadata here
71
- options = {"aot_inductor.package" : package , "aot_inductor. metadata" : metadata }
71
+ options = {"aot_inductor.metadata" : metadata }
72
72
if not package :
73
73
options = {"aot_inductor.output_path" : output_path }
74
74
75
- path = torch ._export . aot_compile (
75
+ ep = torch .export . export (
76
76
model ,
77
77
example_inputs ,
78
78
dynamic_shapes = dynamic_shapes ,
79
- options = options ,
80
79
)
81
80
82
81
if package :
83
- from torch ._inductor .package import package_aoti
84
- path = package_aoti (output_path , path )
82
+ path = torch ._inductor .aoti_compile_and_package (
83
+ ep , package_path = output_path , inductor_configs = options
84
+ )
85
+ else :
86
+ path = torch ._inductor .aot_compile (
87
+ ep .module (), example_inputs , options = options
88
+ )
85
89
86
90
print (f"The generated packaged model can be found at: { path } " )
87
91
return path
You can’t perform that action at this time.
0 commit comments