| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171 |
- import ctypes, collections, time, itertools
- from typing import List, Any, Dict, cast, Optional, Tuple
- from tinygrad.helpers import GraphException, init_c_var, round_up
- from tinygrad.device import Buffer, BufferOptions
- from tinygrad.device import Compiled, Device
- from tinygrad.shape.symbolic import Variable
- from tinygrad.runtime.ops_hsa import HSADevice, PROFILE, Profiler
- from tinygrad.engine.realize import ExecItem, BufferXfer, CompiledRunner
- from tinygrad.engine.jit import MultiGraphRunner
- import tinygrad.runtime.autogen.hsa as hsa
- from tinygrad.runtime.support.hsa import check, AQLQueue, AQL_PACKET_SIZE, EMPTY_SIGNAL
- def dedup_signals(signals): return [hsa.hsa_signal_t(hndl) for hndl in set([x.handle for x in signals if isinstance(x, hsa.hsa_signal_t)])]
- class VirtAQLQueue(AQLQueue):
- def __init__(self, device, sz):
- self.device = device
- self.virt_queue = (hsa.hsa_kernel_dispatch_packet_t * sz)()
- self.queue_base = self.write_addr = ctypes.addressof(self.virt_queue)
- self.packets_count = 0
- self.available_packet_slots = sz
- def _wait_queue(self, need_packets=1): assert False, f"VirtQueue is too small to handle {self.packets_count+need_packets} packets!"
- def _submit_packet(self):
- self.write_addr += AQL_PACKET_SIZE
- self.packets_count += 1
- self.available_packet_slots -= 1
- class HSAGraph(MultiGraphRunner):
- def __init__(self, jit_cache: List[ExecItem], input_rawbuffers: List[Buffer], var_vals: Dict[Variable, int]):
- super().__init__(jit_cache, input_rawbuffers, var_vals)
- # Check all jit items are compatible.
- compiled_devices = set()
- for ji in self.jit_cache:
- if isinstance(ji.prg, CompiledRunner): compiled_devices.add(ji.prg.device)
- elif isinstance(ji.prg, BufferXfer):
- for x in ji.bufs[0:2]: compiled_devices.add(Device[cast(Buffer, x).device])
- else: raise GraphException
- if any(not isinstance(d, HSADevice) for d in compiled_devices): raise GraphException
- self.devices: List[HSADevice] = list(compiled_devices) #type:ignore
- # Allocate kernel args.
- kernargs_size: Dict[Compiled, int] = collections.defaultdict(int)
- for ji in self.jit_cache:
- if isinstance(ji.prg, CompiledRunner): kernargs_size[ji.prg.device] += round_up(ctypes.sizeof(ji.prg.clprg.args_struct_t), 16)
- kernargs_ptrs: Dict[Compiled, int] = {dev:dev.allocator._alloc(sz, BufferOptions()) for dev,sz in kernargs_size.items()}
- # Fill initial arguments.
- self.ji_kargs_structs: Dict[int, ctypes.Structure] = {}
- for j,ji in enumerate(self.jit_cache):
- if not isinstance(ji.prg, CompiledRunner): continue
- self.ji_kargs_structs[j] = ji.prg.clprg.args_struct_t.from_address(kernargs_ptrs[ji.prg.device])
- kernargs_ptrs[ji.prg.device] += round_up(ctypes.sizeof(ji.prg.clprg.args_struct_t), 16)
- for i in range(len(ji.bufs)): self.ji_kargs_structs[j].__setattr__(f'f{i}', cast(Buffer, ji.bufs[i])._buf)
- for i in range(len(ji.prg.p.vars)): self.ji_kargs_structs[j].__setattr__(f'v{i}', var_vals[ji.prg.p.vars[i]])
- # Build queues.
- self.virt_aql_queues: Dict[Compiled, VirtAQLQueue] = {dev:VirtAQLQueue(dev, 2*len(self.jit_cache)+16) for dev in self.devices}
- self.packets = {}
- self.transfers = []
- self.ji_to_transfer: Dict[int, int] = {} # faster to store transfers as list and update using this mapping table.
- self.signals_to_reset: List[hsa.hsa_signal_t] = []
- self.signals_to_devices: Dict[ctypes.c_uint64, List[HSADevice]] = {}
- self.profile_info: Dict[Compiled, List[Tuple[Any, ...]]] = collections.defaultdict(list)
- # Special packet to wait for the world.
- self.kickoff_signals: Dict[HSADevice, hsa.hsa_signal_t] = {dev:self.alloc_signal(reset_on_start=True) for dev in self.devices}
- for dev in self.devices: self.virt_aql_queues[dev].submit_barrier([], self.kickoff_signals[dev])
- for j,ji in enumerate(self.jit_cache):
- if isinstance(ji.prg, CompiledRunner):
- wait_signals = self.access_resources(ji.bufs[(outs:=ji.prg.p.outcount):], ji.bufs[:outs], new_dependency=j, sync_with_aql_packets=False)
- for i in range(0, len(wait_signals), 5):
- self.virt_aql_queues[ji.prg.device].submit_barrier(wait_signals[i:i+5])
- self.packets[j] = hsa.hsa_kernel_dispatch_packet_t.from_address(self.virt_aql_queues[ji.prg.device].write_addr)
- sync_signal = self.alloc_signal(reset_on_start=True) if PROFILE else None
- self.virt_aql_queues[ji.prg.device].submit_kernel(ji.prg.clprg, *ji.prg.p.launch_dims(var_vals), #type:ignore
- ctypes.addressof(self.ji_kargs_structs[j]), completion_signal=sync_signal)
- if PROFILE: self.profile_info[ji.prg.device].append((sync_signal, ji.prg.clprg.name, False))
- elif isinstance(ji.prg, BufferXfer):
- dest, src = [cast(Buffer, x) for x in ji.bufs[0:2]]
- dest_dev, src_dev = cast(HSADevice, Device[dest.device]), cast(HSADevice, Device[src.device])
- sync_signal = self.alloc_signal(reset_on_start=True, wait_on=[dest_dev, src_dev])
- wait_signals = self.access_resources(read=[src], write=[dest], new_dependency=sync_signal, sync_with_aql_packets=True)
- self.transfers.append([dest._buf, dest_dev.agent, src._buf, src_dev.agent, dest.nbytes, len(wait_signals),
- (hsa.hsa_signal_t*len(wait_signals))(*wait_signals), sync_signal, hsa.HSA_AMD_SDMA_ENGINE_0, True])
- self.ji_to_transfer[j] = len(self.transfers) - 1
- if PROFILE: self.profile_info[src_dev].append((sync_signal, f"transfer: HSA:{src_dev.device_id} -> HSA:{dest_dev.device_id}", True))
- # Wait for all active signals to finish the graph
- wait_signals_to_finish: Dict[HSADevice, List[hsa.hsa_signal_t]] = collections.defaultdict(list)
- for v in dedup_signals(list(self.w_dependency_map.values()) + list(itertools.chain.from_iterable(self.r_dependency_map.values()))):
- for dev in self.signals_to_devices[v.handle]:
- wait_signals_to_finish[dev].append(v)
- self.finish_signal = init_c_var(hsa.hsa_signal_t(), lambda x: check(hsa.hsa_amd_signal_create(1, 0, None, 0, ctypes.byref(x))))
- for dev in self.devices:
- wait_signals = wait_signals_to_finish[dev]
- for i in range(0, max(1, len(wait_signals)), 5):
- self.virt_aql_queues[dev].submit_barrier(wait_signals[i:i+5], completion_signal=self.finish_signal if i+5>=len(wait_signals) else None)
- # Zero signals to allow graph to start and execute.
- for sig in self.signals_to_reset: hsa.hsa_signal_silent_store_relaxed(sig, 0)
- hsa.hsa_signal_silent_store_relaxed(self.finish_signal, 0)
- def __call__(self, input_rawbuffers: List[Buffer], var_vals: Dict[Variable, int], wait=False) -> Optional[float]:
- # Wait and restore signals
- hsa.hsa_signal_wait_scacquire(self.finish_signal, hsa.HSA_SIGNAL_CONDITION_LT, 1, (1 << 64) - 1, hsa.HSA_WAIT_STATE_ACTIVE)
- for sig in self.signals_to_reset: hsa.hsa_signal_silent_store_relaxed(sig, 1)
- hsa.hsa_signal_silent_store_relaxed(self.finish_signal, len(self.devices))
- # Update rawbuffers
- for (j,i),input_idx in self.input_replace.items():
- if j in self.ji_kargs_structs:
- self.ji_kargs_structs[j].__setattr__(f'f{i}', input_rawbuffers[input_idx]._buf)
- else:
- if i == 0: self.transfers[self.ji_to_transfer[j]][0] = input_rawbuffers[input_idx]._buf # dest
- elif i == 1: self.transfers[self.ji_to_transfer[j]][2] = input_rawbuffers[input_idx]._buf # src
- # Update var_vals
- for j in self.jc_idx_with_updatable_var_vals:
- for i,v in enumerate(cast(CompiledRunner, self.jit_cache[j].prg).p.vars):
- self.ji_kargs_structs[j].__setattr__(f'v{i}', var_vals[v])
- # Update launch dims
- for j in self.jc_idx_with_updatable_launch_dims:
- gl, lc = cast(CompiledRunner, self.jit_cache[j].prg).p.launch_dims(var_vals)
- self.packets[j].workgroup_size_x = lc[0]
- self.packets[j].workgroup_size_y = lc[1]
- self.packets[j].workgroup_size_z = lc[2]
- self.packets[j].grid_size_x = gl[0] * lc[0]
- self.packets[j].grid_size_y = gl[1] * lc[1]
- self.packets[j].grid_size_z = gl[2] * lc[2]
- for dev in self.devices:
- dev.flush_hdp()
- dev.hw_queue.blit_packets(self.virt_aql_queues[dev].queue_base, self.virt_aql_queues[dev].packets_count)
- for transfer_data in self.transfers:
- check(hsa.hsa_amd_memory_async_copy_on_engine(*transfer_data))
- et = None
- if wait:
- st = time.perf_counter()
- hsa.hsa_signal_wait_scacquire(self.finish_signal, hsa.HSA_SIGNAL_CONDITION_LT, 1, (1 << 64) - 1, hsa.HSA_WAIT_STATE_ACTIVE)
- et = time.perf_counter() - st
- for profdev,profdata in self.profile_info.items(): Profiler.tracked_signals[profdev] += profdata
- return et
- def alloc_signal(self, reset_on_start=False, wait_on=None):
- sync_signal = init_c_var(hsa.hsa_signal_t(), lambda x: check(hsa.hsa_amd_signal_create(1, 0, None, 0, ctypes.byref(x))))
- if reset_on_start: self.signals_to_reset.append(sync_signal)
- if wait_on is not None: self.signals_to_devices[sync_signal.handle] = wait_on
- return sync_signal
- def dependency_as_signal(self, dep, sync_with_aql_packets) -> Optional[hsa.hsa_signal_t]:
- if isinstance(dep, hsa.hsa_signal_t): return dep
- elif sync_with_aql_packets and isinstance(packet := self.packets.get(dep), hsa.hsa_kernel_dispatch_packet_t):
- if packet.completion_signal.handle == EMPTY_SIGNAL.handle: packet.completion_signal = self.alloc_signal(reset_on_start=True)
- return packet.completion_signal
- return None
- def access_resources(self, read, write, new_dependency, sync_with_aql_packets=False):
- rdeps = self._access_resources(read, write, new_dependency)
- wait_signals = [self.dependency_as_signal(dep, sync_with_aql_packets=sync_with_aql_packets) for dep in rdeps]
- if sync_with_aql_packets: wait_signals += [self.kickoff_signals[cast(HSADevice, Device[rawbuf.device])] for rawbuf in read+write]
- return dedup_signals(wait_signals)
|