We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 3a69560 commit 77ccda8Copy full SHA for 77ccda8
flashinfer/jit/core.py
@@ -87,8 +87,9 @@ def load_cuda_ops(
87
extra_cuda_cflags: Optional[List[str]] = None,
88
extra_ldflags=None,
89
extra_include_paths=None,
90
- verbose=False,
91
):
+ verbose = os.environ.get("FLASHINFER_JIT_VERBOSE", "0") == "1"
92
+
93
if extra_cflags is None:
94
extra_cflags = []
95
if extra_cuda_cflags is None:
@@ -106,6 +107,14 @@ def load_cuda_ops(
106
107
"-DFLASHINFER_ENABLE_FP8_E4M3",
108
"-DFLASHINFER_ENABLE_FP8_E5M2",
109
]
110
+ if verbose:
111
+ cuda_cflags += [
112
+ "-g",
113
+ "-lineinfo",
114
+ "--ptxas-options=-v",
115
+ "--ptxas-options=--verbose,--register-usage-level=10,--warn-on-local-memory-usage",
116
+ ]
117
118
cflags += extra_cflags
119
cuda_cflags += extra_cuda_cflags
120
logger.info(f"Loading JIT ops: {name}")
0 commit comments