@@ -1225,12 +1225,12 @@ int main(int argc, char ** argv) {
1225
1225
// struct ggml_tensor * e = cross_entropy_loss(ctx0, targets1, logits1);
1226
1226
// struct ggml_tensor * e = square_error_loss(ctx0, targets1, logits1);
1227
1227
1228
- struct ggml_tensor * e = ggml_add (ctx0,
1229
- square_error_loss (ctx0, targets1, logits1),
1230
- square_error_loss (ctx0, targets2, logits2));
1231
1228
// struct ggml_tensor * e = ggml_add(ctx0,
1232
- // cross_entropy_loss(ctx0, targets1, logits1),
1233
- // cross_entropy_loss(ctx0, targets2, logits2));
1229
+ // square_error_loss(ctx0, targets1, logits1),
1230
+ // square_error_loss(ctx0, targets2, logits2));
1231
+ struct ggml_tensor * e = ggml_add (ctx0,
1232
+ cross_entropy_loss (ctx0, targets1, logits1),
1233
+ cross_entropy_loss (ctx0, targets2, logits2));
1234
1234
// struct ggml_tensor * e = ggml_add(ctx0,
1235
1235
// ggml_add(ctx0,
1236
1236
// cross_entropy_loss(ctx0, targets1, logits1),
@@ -1258,8 +1258,8 @@ int main(int argc, char ** argv) {
1258
1258
opt_params_lbfgs.print_backward_graph = false ;
1259
1259
opt_params_adam.adam .n_iter = 16 ;
1260
1260
opt_params_lbfgs.lbfgs .n_iter = 16 ;
1261
- // ggml_opt(ctx0, opt_params_adam, e);
1262
- ggml_opt (ctx0, opt_params_lbfgs, e);
1261
+ ggml_opt (ctx0, opt_params_adam, e);
1262
+ // ggml_opt(ctx0, opt_params_lbfgs, e);
1263
1263
//
1264
1264
ggml_build_forward_expand (&gf, e);
1265
1265
ggml_graph_compute (ctx0, &gf);
0 commit comments