Browse Source

added logger

josh 8 months ago
parent
commit
f4a5562c83
2 changed files with 6 additions and 3 deletions
  1. 5 3
      exo/main.py
  2. 1 0
      exo/networking/udp/udp_discovery.py

+ 5 - 3
exo/main.py

@@ -156,8 +156,11 @@ async def shutdown(signal, loop):
 async def select_best_inference_engine(node: StandardNode):
   supported_engines = node.get_supported_inference_engines()
   await node.broadcast_supported_engines(supported_engines)
-
+  logger.error('ABOVE and ALL')
+  logger.error("Topology inference engines pool: %s", node.get_topology_inference_engines())
+  logger.error(f'result:{node.get_topology_inference_engines()}')
   if node.get_topology_inference_engines():
+    logger.info("Topology inference engines pool: %s", node.get_topology_inference_engines())
     topology_inference_engines_pool = node.get_topology_inference_engines()
     if any("tinygrad" in engines and len(engines) == 1 for engines in topology_inference_engines_pool):
         return "tinygrad"
@@ -207,14 +210,13 @@ async def main():
     loop.add_signal_handler(s, handle_exit)
 
   await node.start(wait_for_peers=args.wait_for_peers)
-  await select_best_inference_engine(node)
-  
 
   if args.command == "run" or args.run_model:
     model_name = args.model_name or args.run_model
     if not model_name:
       print("Error: Model name is required when using 'run' command or --run-model")
       return
+    await select_best_inference_engine(node)
     await run_model_cli(node, inference_engine, model_name, args.prompt)
   else:
     asyncio.create_task(api.run(port=args.chatgpt_api_port))  # Start the API server as a non-blocking task

+ 1 - 0
exo/networking/udp/udp_discovery.py

@@ -160,6 +160,7 @@ class UDPDiscovery(Discovery):
           return
         if peer_id in self.known_peers: self.known_peers[peer_id] = (self.known_peers[peer_id][0], self.known_peers[peer_id][1], time.time(), peer_prio)
     elif message["type"] == "supported_inference_engines":
+      logger.error(f'supported_inference_engines: {message}')
       peer_id = message["node_id"]
       engines = message["engines"]
       if peer_id in self.known_peers: self.known_peers[peer_id][0].topology_inference_engines_pool.append(engines)