Alex Cheema 1 anno fa
parent
commit
c691205591
1 ha cambiato i file con 0 aggiunte e 1 eliminazioni
  1. 0 1
      exo/inference/mlx/test_sharded_llama.py

+ 0 - 1
exo/inference/mlx/test_sharded_llama.py

@@ -28,7 +28,6 @@ for _ in range(max_tokens):
 
 print("full response: ", full_tokenizer.decode(full_generated_tokens))
 
-
 sharded_generated_tokens = []
 sharded_resp = prompt_tokens
 for _ in range(max_tokens):