Spaces:
Sleeping
Sleeping
Update tools/llama/generate.py
Browse files- tools/llama/generate.py +1 -0
tools/llama/generate.py
CHANGED
@@ -22,6 +22,7 @@ torch.cuda.is_available = lambda: False
|
|
22 |
os.environ["TOKENIZERS_PARALLELISM"] = "false"
|
23 |
torch._inductor.config.coordinate_descent_tuning = True
|
24 |
torch._inductor.config.triton.unique_kernel_names = True
|
|
|
25 |
|
26 |
if hasattr(torch._inductor.config, "fx_graph_cache"):
|
27 |
# Experimental feature to reduce compilation times, will be on by default in future
|
|
|
22 |
os.environ["TOKENIZERS_PARALLELISM"] = "false"
|
23 |
torch._inductor.config.coordinate_descent_tuning = True
|
24 |
torch._inductor.config.triton.unique_kernel_names = True
|
25 |
+
device = torch.device('cpu')
|
26 |
|
27 |
if hasattr(torch._inductor.config, "fx_graph_cache"):
|
28 |
# Experimental feature to reduce compilation times, will be on by default in future
|