You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
hooks : setting up flake8 and pre-commit hooks (ggml-org#1681)
Small, non-functional changes were made to non-compliant files.
These include breaking up long lines, whitespace sanitation and
unused import removal.
Maximum line length in python files was set to a generous 125 chars,
in order to minimize number of changes needed in scripts and general
annoyance. The "txt" prompts directory is excluded from the checks
as it may contain oddly formatted files and strings for a good reason.
Signed-off-by: Jiri Podivin <[email protected]>
raiseException(f"Can't turn an unquantized tensor into a quantized type ({data_type})")
514
514
ifself.data_type.have_g_idx:
515
-
sys.stderr.write("Error: Input uses the newer GPTQ-for-LLaMa format (using g_idx), which is not yet natively supported by GGML. For now you can still convert this model by passing `--outtype f16` to dequantize, but that will result in a much larger output file for no quality benefit.\n")
515
+
sys.stderr.write(
516
+
"Error: Input uses the newer GPTQ-for-LLaMa format (using g_idx), "
517
+
"which is not yet natively supported by GGML. "
518
+
"For now you can still convert this model by passing `--outtype f16` to dequantize, "
519
+
"but that will result in a much larger output file for no quality benefit.\n")
0 commit comments