1
0
Alex Cheema 1 жил өмнө
parent
commit
c691205591

+ 0 - 1
exo/inference/mlx/test_sharded_llama.py

@@ -28,7 +28,6 @@ for _ in range(max_tokens):
 
 print("full response: ", full_tokenizer.decode(full_generated_tokens))
 
-
 sharded_generated_tokens = []
 sharded_resp = prompt_tokens
 for _ in range(max_tokens):