
    9iJ                        d dl Z d dlZd dlmZmZmZ d dlZd dlmZ	 de j
        fdZde j
        fdZde j
        fdZdeddfd	Zde j
        fd
Zde j
        fdZde j
        fdZdee         fdZ	 	 	 	 d!dededee         dee         dee         dedeeef         fdZ G d d          Z G d d          Z	 d"deeef         deee                  deeeedf         f         fdZ	 d#dedededefd ZdS )$    N)AnyOptionalUnion)_get_device_indexreturnc                  "   t           j        dk    r.t          j        dt          j        j        d          d          } nt          j        d          } | j        | _        | j	        | _
        | j        | _        | j        | _        | j        | _        | S )Nwin32	amdhip64_r   .dllzlibamdhip64.so)sysplatformctypesCDLLtorchversionhiphipGetErrorStringcuGetErrorStringhipModuleLoadDatacuModuleLoadDatahipModuleGetFunctioncuModuleGetFunctionhipModuleLaunchKernelcuLaunchKernelhipFuncSetAttributecuFuncSetAttribute)libs    D/var/www/icac/venv/lib/python3.11/site-packages/torch/cuda/_utils.py_get_hip_runtime_libraryr      s    
|wk@em&7&:@@@AAk*++0C0C!6C2C 4CJ    c                  r    t           j        dk    rt          j        d          S t          j        d          S )Nr	   z
nvcuda.dllzlibcuda.so.1)r   r   r   r    r    r   _get_cuda_libraryr#      s/    
|w{<((({>***r    c                  \    t           j        j        rt                      S t	                      S N)r   r   r   r   r#   r"   r    r   _get_gpu_runtime_libraryr&       s'    } #'))) """r    resultc                    | dk    rd S t          j                    }t                      }|                    | t          j        |                     |j        |j                                        nd}t          d|           )Nr   Unknown CUDA errorCUDA error: )r   c_char_pr&   r   byrefvaluedecodeRuntimeError)r'   err_strlibcudaerror_messages       r   _check_cudar3   (   s    {{oG&((GVV\'%:%:;;;")-";AU  5m55
6
66r    c                     t           j        dk    r\d                    dt          j        j        d         dt          j        j        d         g          } t          j        d|  d          }nt          j        d          }|j        |_	        |j
        |_        |j        |_        |j        |_        |j        |_        |j        |_        |j        |_        |j        |_        |j        |_        |j        |_        |S )	Nr	    0r      hiprtcr   zlibhiprtc.so)r   r   joinr   r   r   r   r   hiprtcGetErrorStringnvrtcGetErrorStringhiprtcCreateProgramnvrtcCreateProgramhiprtcDestroyProgramnvrtcDestroyProgramhiprtcCompileProgramnvrtcCompileProgramhiprtcGetCodeSizenvrtcGetPTXSizehiprtcGetCodenvrtcGetPTXhiprtcGetProgramLogSizenvrtcGetProgramLogSizehiprtcGetProgramLognvrtcGetProgramLoghiprtcAddNameExpressionnvrtcAddNameExpressionhiprtcGetLoweredNamenvrtcGetLoweredName)version_strr   s     r   _get_hiprtc_libraryrO   4   s    
|wggsEM$5a$8#u}?PQR?STUUk4;44455k.)) "6C 4C!6C!6C/C'CO!$!<C 4C!$!<C!6CJr    c                  "   t          t          j        j                            d          d                   } t
          j        dk    rd|  dg}nd|  dg}|D ](}	 t          j        |          c S # t          $ r Y %w xY wt          d          )	N.r   r	   nvrtc64_z0_0.dllzlibnvrtc.so.zlibnvrtc.soz Could not find any NVRTC library)
intr   r   cudasplitr   r   r   r   OSError)major_version
nvrtc_libslib_names      r   _get_nvrtc_libraryrZ   I   s    *0055a899M
|w-}---



 +=**

   	;x((((( 	 	 	H	
4
5
55s   A22
A?>A?c                  \    t           j        j        rt                      S t	                      S r%   )r   r   r   rO   rZ   r"   r    r   _get_gpu_rtc_libraryr\   \   s)     } $"$$$!###r    c                      ddl m} m} dhfd|D             }t          j        j        r|                    |            |S )z
    Get HIPCC/NVCC flags that are compatible with NVRTC compilation.

    Returns:
        List of HIPCC/NVCC flags that can be safely used with NVRTC.
    r   )COMMON_HIPCC_FLAGSCOMMON_NVCC_FLAGSz--expt-relaxed-constexprc                     g | ]}|v|	S r"   r"   ).0flagnvrtc_unsupported_flagss     r   
<listcomp>z1_get_gpu_rtc_compatible_flags.<locals>.<listcomp>s   s+       d:Q.Q.Q.Q.Q.Qr    )torch.utils.cpp_extensionr^   r_   r   r   r   extend)r^   r_   compatible_flagsrc   s      @r   _get_gpu_rtc_compatible_flagsrh   e   s~     POOOOOOO 	#
   *   } 4 2333r    Fkernel_sourcekernel_namecompute_capabilitycuda_include_dirsnvcc_optionsauto_pchc           
       	   ddl }t                      ddt          ddffd}|                     d          }|V|j                            |j                                                  }	|j        j        r	|	j	         }n|	j
         |	j         }g }
|j        j        r+|
                    d|                                            n*|
                    d|                                            dd	lm}  |d
          }|D ],}|
                    d|                                            -|r/|D ],}|
                    d|                                            -|r@t          |j        j                  dk    s
J d            |g }|                    d           |r-|D ]*}|
                    |                    d                     +t!                      }|
                    d |D                        t%          |
          }t'          j        |z  |
 }t'          j                    } |                    t'          j        |          || d                                ddd                     |                    d          } |                    ||                                         |||          }|k    rt'          j                    }                    |t'          j        |                     t'          j        |j                  }                    ||           t?          d|j                                                    t'          j                    } |!                    |t'          j        |                               t'          j        |j                  } |"                    ||                     t'          j                    } |#                    ||t'          j        |                               |j        |j                                         }nd}$                    t'          j        |                     |j        j        r|j%        n|j        }||fS )a  
    Compiles a CUDA kernel using NVRTC and returns the PTX code.

    Args:
        kernel_source (str): The CUDA kernel source code as a string
        kernel_name (str): The name of the kernel function to compile
        compute_capability (str, None): The compute capability to target (e.g., "86").
                                           If None, will detect from current device.
        cuda_include_dirs (list, None): List of directories containing CUDA headers
        nvcc_options (list, None): Additional options to pass to NVRTC
        auto_pch (bool): Enable automatic precompiled headers (CUDA 12.8+)

    Returns:
        Tuple[bytes, str]: The compiled PTX code and mangled kernel name
    r   Nr'   r   c                     | k    rot          j                    }                    | t          j        |                     |j        |j                                        nd}t          d|           d S )Nr)   r*   )r   r+   r;   r,   r-   r.   r/   )r'   r0   r2   NVRTC_SUCCESSlibnvrtcs      r   check_nvrtcz#_nvrtc_compile.<locals>.check_nvrtc   s    ]""o''G((g1F1FGGG =, $$&&&) 
 =m==>>> #"r    utf-8z--offload-arch=z--gpu-architecture=sm_)include_pathsrT   z-Iz12.8zPCH requires CUDA 12.8+z--pchc                 8    g | ]}|                     d           S )rt   )encode)ra   rb   s     r   rd   z"_nvrtc_compile.<locals>.<listcomp>   s$    LLLTDKK((LLLr    z.cuzKernel compilation failed:
r5   )&
torch.cudar\   rS   rw   rT   get_device_propertiescurrent_devicer   r   gcnArchNamemajorminorappendre   ru   strrh   rf   lenr   r+   c_void_pr=   r,   rK   rA   c_size_trG   create_string_bufferr-   rI   r/   r.   rC   rE   rM   r?   raw)ri   rj   rk   rl   rm   rn   r   rs   source_bytespropsoptionsru   cuda_include_paths	cuda_path	directoryoptionnvrtc_compatible_flagsnum_optionsoptions_arrayprogc_kernel_namereslog_sizelogptx_sizeptxc_mangled_namemangled_name	ptx_bytesrq   rr   s                                @@r   _nvrtc_compiler   }   s   0  $%%H M	?C 	?D 	? 	? 	? 	? 	? 	? 	? !''00L !
001J1J1L1LMM= 	?$)$5!7$)K!>!>!> G} O=);==DDFFGGGGD0BDDKKMMNNN 877777&v..' 2 2	'I''..001111  6* 	6 	6INN+	++22445555  %5=%&&&0002K000LG$$$  3" 	3 	3FNN6==112222:<<NNLL5KLLLMMM g,,K_{2W=M ?DK##L&&((	
 	
	 	 	  &&w//MK//mDDEEE 
&
&t[-
H
HC m?$$''fl8.D.DEEE)(.99##D#...N#):J:J:L:LNNOOO   HK((v|H/E/EFFGGG

%hn
5
5CK$$T3//000 _&&NK$$T=&,~:V:VWW   '%+2244  d!3!3444
 !=,;#)Il""r    c                   8    e Zd Zdej        ddfdZdeddfdZdS )_CudaModulemoduler   Nc                 "    || _         i | _        d S r%   )_module_kernels)selfr   s     r   __init__z_CudaModule.__init__  s    02r    name_CudaKernelc           	         || j         v r| j         |         S ddlm}  |            }t          j                    }	 t          |                    t          j        |          | j        |	                    d                               t          || j                  }|| j         |<   |S # t          $ r}t          d| d          |d }~ww xY w)Nr   )r&   rt   zNo kernel named 'z' in this module)r   torch.cuda._utilsr&   r   r   r3   r   r,   r   rw   r   r/   AttributeError)r   r   r&   r1   funckernelerrs          r   __getattr__z_CudaModule.__getattr__  s    4=  =&& 	?>>>>>**,,  	V++L&&dkk'6J6J   
 !t|44F"(DM$M 	V 	V 	V !KT!K!K!KLLRUU	Vs   A.B* *
C4CC)__name__
__module____qualname__r   r   r   r   r   r"   r    r   r   r     sb        3v 34 3 3 3 3V V V V V V V Vr    r   c                       e Zd ZdZdej        dej        ddfdZ	 	 	 	 	 dd	eeeef         d
eeeef         de	e
         dede	e         ddfdZdeddfdZdS )r   zT
    Represents a compiled CUDA kernel that can be called with PyTorch tensors.
    r   r   r   Nc                 0    || _         || _        d| _        d S )Nr   )r   r   _max_shared_mem_bytes)r   r   r   s      r   r   z_CudaKernel.__init__3  s    	%&"""r       r   r   r   gridblockargs
shared_memstreamc                 t   ddl }|j        j                                        }|sg }g }g }	|D ]l}
t	          |
|j                  r|
j        s*|
j        r|
                                st          d          t          j        |
                                          }|                    |           |	                    t          j        |                     t	          |
t                    r<t          j        |
          }|	                    t          j        |                     t	          |
t"                    r=t          j        |
          }|	                    t          j        |                     Ot'          dt)          |
                     t          j        t+          |	          z              }t-          |	          D ]'\  }}
t          j        |
t          j                  ||<   (|ddl}|j                                        }|dk    rD| j        dk    s|| j        k    r.| j        dk    rdn
d| j         d}t7          d	| d
| d          t9          |                    | j        |d         |d         |d         |d         |d         |d         ||j        |d                     dS )a  
        Call the compiled CUDA kernel

        Args:
            grid (tuple): Grid dimensions (grid_x, grid_y, grid_z)
            block (tuple): Block dimensions (block_x, block_y, block_z)
            args (list): List of arguments to pass to the kernel.
                         PyTorch tensor arguments will be automatically converted to pointers.
            shared_mem (int): Shared memory size in bytes
            stream (torch.cuda.Stream): CUDA stream to use. If None, uses current stream.
        r   Nz?All tensor arguments must be CUDA tensors or pinned CPU tensorszUnsupported argument type:    znot configuredzonly z bytes configuredzKernel requires z' bytes of shared memory (>= 48KB), but ze. Call kernel.set_shared_memory_config(shared_mem) after compilation and before launching the kernel.r   r7   ) r   rT   _utilsr&   
isinstanceTensoris_cudais_cpu	is_pinned
ValueErrorr   r   data_ptrr~   r,   rS   c_intfloatc_double	TypeErrortyper   	enumeratecastrx   current_streamr   r/   r3   r   r   _as_parameter_)r   r   r   r   r   r   r   r1   processed_argsc_argsargptrr   r   c_args_arrayiconfigured_msgs                    r   __call__z_CudaKernel.__call__8  s   & 	*#<<>> 	D 13 	K 	KC#u|,, K{ CJ 3==?? $Y   ocllnn55%%c***fl3//0000C%% KS))fl5112222C'' K!?3//fl8445555 Id3ii I IJJJ #f++588'' 	@ 	@FAs$k#v??LOO >Z..00F ""&!++zD<V/V/V -22 ! JT7JJJ 
 3: 3 3%3 3 3   	""	QQQaaa% 	
 	
 	
 	
 	
r    shared_mem_bytesc                 |   |dk     r	|| _         d S t                      }t          j                                        }t          j        j        r|j        dk    rdnd}nt          |dd          }||k    rt          d| d| d          d	}t          |                    | j        ||                     || _         d S )
Nr   gfx950i   i  shared_memory_per_block_optinzRequested shared memory (z bytes) exceeds device limit (z= bytes). Consider reducing block size or shared memory usage.   )r   r&   r   rT   ry   r   r   r{   getattrr/   r3   r   r   )r   r   r1   device_propsmax_shared_mem+cudaFuncAttributeMaxDynamicSharedMemorySizes         r   set_shared_memory_configz$_CudaKernel.set_shared_memory_config  s   i'')9D&F*,, z7799= 		 &1X==: N %=u N n,,G,< G G!/G G G   783&&	;  	
 	
 	
 &6"""r    )r   r   Nr   N)r   r   r   __doc__r   r   r   tuplerS   r   listr   r   r   r"   r    r   r   r   .  s         'V_ 'fo '$ ' ' ' ' &/&/# $_
 _
CcM"_
 S#s]#_
 tn	_

 _
 _
 
_
 _
 _
 _
B(6 (6 (6 (6 (6 (6 (6 (6r    r   r   kernel_namesc           
         ddl }t                      }t          | t                    r|                     d          } t          j                    }|j                                        }|5  t          |
                    t          j        |          |                      ddd           n# 1 swxY w Y   |st          |          S i }|D ]q}t          j                    }t          |                    t          j        |          ||                    d                               t          ||          ||<   r|S )a,  
    Loads a CUDA module from PTX code and returns a module object that can access kernels.

    Args:
        ptx (bytes or str): The PTX code to load
        kernel_names (list, optional): List of kernel names to extract from the module.
                                      If None, will return a module object with __getattr__.

    Returns:
        object: If kernel_names is None, returns a module object with __getattr__ to access kernels.
               If kernel_names is provided, returns a dict mapping kernel names to _CudaKernel objects.
    r   Nrt   )rx   r&   r   r   rw   r   r   rT   r   r3   r   r,   r   r   r   )	r   r   r   r1   r   r   kernelsr   r   s	            r   _cuda_load_moduler     s      '((G #s "jj!! _FZ&&((F	 I IG,,V\&-A-A3GGHHHI I I I I I I I I I I I I I I  #6""" G 2 2  ''T""FDKK,@,@ 	
 	
 	

 $D&11Ns   +6B--B14B1deviceoptional	allow_cpuc                    t          | t                    r| S t          | t                    rt          j        |           } t          | t          j                  r;|r| j        dvrt          d|            n| j        dk    rt          d|            t          j                                        s&t          | t          j	        j                  r| j
        S t          | ||          S )a  Get the device index from :attr:`device`, which can be a torch.device object, a Python integer, or ``None``.

    If :attr:`device` is a torch.device object, returns the device index if it
    is a CUDA device. Note that for a CUDA device without a specified index,
    i.e., ``torch.device('cuda')``, this will return the current default CUDA
    device if :attr:`optional` is ``True``. If :attr:`allow_cpu` is ``True``,
    CPU devices will be accepted and ``-1`` will be returned in this case.

    If :attr:`device` is a Python integer, it is returned as is.

    If :attr:`device` is ``None``, this will return the current default CUDA
    device if :attr:`optional` is ``True``.
    )rT   cpuz(Expected a cuda or cpu device, but got: rT   z!Expected a cuda device, but got: )r   rS   r   r   r   r   r   jitis_scriptingrT   idx_torch_get_device_index)r   r   r   s      r   r   r     s      &# &# &f%%&%,'' K 	K{/11 !TF!T!TUUU 2[F""IIIJJJ9!!## fej/00 	:"68Y???r    )NNNFr%   )FF)r   r   typingr   r   r   r   torch._utilsr   r   r   r   r#   r&   rS   r3   rO   rZ   r\   r   r   rh   boolr   bytesr   r   r   dictr   r"   r    r   <module>r      s    



 ' ' ' ' ' ' ' ' ' '  F E E E E E
&+ 
 
 
 
+6; + + + +#&+ # # # #	7 	7 	7 	7 	7 	7V[    *6FK 6 6 6 6&$fk $ $ $ $tCy    6 )-(,#'P# P#P#P# !P#  ~	P#
 4.P# P# 5#:P# P# P# P#fV V V V V V V V<S6 S6 S6 S6 S6 S6 S6 S6n AE- -	sEz	-*249*=-
;S-/001- - - -b <A@ @@@48@@ @ @ @ @ @r    