@@ -82,8 +82,6 @@ runtime.python_binary(
82
82
],
83
83
deps = [
84
84
":export_library",
85
- ":export_llama_args",
86
- ":export_llama_hydra",
87
85
"//caffe2:torch",
88
86
"//executorch/extension/pybindings:aten_lib",
89
87
],
@@ -134,6 +132,8 @@ runtime.python_library(
134
132
name = "export_library",
135
133
srcs = [
136
134
"export_llama.py",
135
+ "export_llama_args.py",
136
+ "export_llama_hydra.py",
137
137
"export_llama_lib.py",
138
138
"model.py",
139
139
],
@@ -234,40 +234,6 @@ runtime.python_library(
234
234
],
235
235
)
236
236
237
- runtime.python_library(
238
- name = "export_llama_args",
239
- srcs = [
240
- "export_llama_args.py",
241
- ],
242
- _is_external_target = True,
243
- base_module = "executorch.examples.models.llama",
244
- visibility = [
245
- "//executorch/examples/...",
246
- "@EXECUTORCH_CLIENTS",
247
- ],
248
- deps = [
249
- ":export_library",
250
- ],
251
- )
252
-
253
- runtime.python_library(
254
- name = "export_llama_hydra",
255
- srcs = [
256
- "export_llama_hydra.py",
257
- ],
258
- _is_external_target = True,
259
- base_module = "executorch.examples.models.llama",
260
- visibility = [
261
- "//executorch/examples/...",
262
- "@EXECUTORCH_CLIENTS",
263
- ],
264
- deps = [
265
- ":export_library",
266
- "//executorch/examples/models/llama/config:llm_config",
267
- "fbsource//third-party/pypi/hydra-core:hydra-core",
268
- ],
269
- )
270
-
271
237
runtime.python_test(
272
238
name = "quantized_kv_cache_test",
273
239
srcs = [
0 commit comments