uto1125 commited on
Commit
77b2a9a
1 Parent(s): b264b39

Update tools/llama/generate.py

Browse files
Files changed (1) hide show
  1. tools/llama/generate.py +1 -1
tools/llama/generate.py CHANGED
@@ -22,7 +22,7 @@ torch.cuda.is_available = lambda: False
22
  os.environ["TOKENIZERS_PARALLELISM"] = "false"
23
  torch._inductor.config.coordinate_descent_tuning = True
24
  torch._inductor.config.triton.unique_kernel_names = True
25
- device = torch.device('cpu')
26
 
27
  if hasattr(torch._inductor.config, "fx_graph_cache"):
28
  # Experimental feature to reduce compilation times, will be on by default in future
 
22
  os.environ["TOKENIZERS_PARALLELISM"] = "false"
23
  torch._inductor.config.coordinate_descent_tuning = True
24
  torch._inductor.config.triton.unique_kernel_names = True
25
+
26
 
27
  if hasattr(torch._inductor.config, "fx_graph_cache"):
28
  # Experimental feature to reduce compilation times, will be on by default in future