Skip to content

Commit d40fded

Browse files
committed
llama : fix comment for "output.weight" tensor
1 parent 2510c18 commit d40fded

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

llama.cpp

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1618,8 +1618,8 @@ static void llama_model_quantize_internal(const std::string & fname_inp, const s
16181618
// quantize only 2D tensors
16191619
quantize &= (tensor.ne.size() == 2);
16201620

1621-
// GG: uncomment this to keep the output layer in FP16
1622-
//if (tensor.name.rfind("output")) {
1621+
// uncomment this to keep the output layer in FP16
1622+
//if (tensor.name == "output.weight") {
16231623
// quantize = false;
16241624
//}
16251625

0 commit comments

Comments
 (0)