Mercurial > hg > truffle
view src/gpu/ptx/vm/gpu_ptx.hpp @ 15490:5d0dd6a6f6b3
mx: improved heuristics for sorting remaining tasks in parallel Java build worklist
author | Doug Simon <doug.simon@oracle.com> |
---|---|
date | Sun, 04 May 2014 01:28:07 +0200 |
parents | 66e3af78ea96 |
children | e86659e1f813 |
line wrap: on
line source
/* * Copyright (c) 2013, Oracle and/or its affiliates. All rights reserved. * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. * * This code is free software; you can redistribute it and/or modify it * under the terms of the GNU General Public License version 2 only, as * published by the Free Software Foundation. * * This code is distributed in the hope that it will be useful, but WITHOUT * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License * version 2 for more details (a copy is included in the LICENSE file that * accompanied this code). * * You should have received a copy of the GNU General Public License version * 2 along with this work; if not, write to the Free Software Foundation, * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. * * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA * or visit www.oracle.com if you need additional information or have any * questions. * */ #ifndef GPU_PTX_HPP #define GPU_PTX_HPP /* * Some useful macro definitions from publicly available cuda.h. * These definitions are for convenience. */ #define GRAAL_CUDA_SUCCESS 0 /**< Device shares a unified address space with the host */ #define GRAAL_CU_DEVICE_ATTRIBUTE_MAX_THREADS_PER_BLOCK 1 #define GRAAL_CU_DEVICE_ATTRIBUTE_UNIFIED_ADDRESSING 41 #define GRAAL_CU_DEVICE_ATTRIBUTE_COMPUTE_CAPABILITY_MAJOR 75 #define GRAAL_CU_DEVICE_ATTRIBUTE_COMPUTE_CAPABILITY_MINOR 76 #define GRAAL_CU_DEVICE_ATTRIBUTE_MULTIPROCESSOR_COUNT 16 #define GRAAL_CU_DEVICE_ATTRIBUTE_WARP_SIZE 10 #define GRAAL_CU_DEVICE_ATTRIBUTE_CAN_MAP_HOST_MEMORY 19 #define GRAAL_CU_DEVICE_ATTRIBUTE_CONCURRENT_KERNELS 31 #define GRAAL_CU_DEVICE_ATTRIBUTE_ASYNC_ENGINE_COUNT 40 #define GRAAL_CU_JIT_MAX_REGISTERS 0 #define GRAAL_CU_JIT_THREADS_PER_BLOCK 1 #define GRAAL_CU_JIT_INFO_LOG_BUFFER 3 #define GRAAL_CU_JIT_INFO_LOG_BUFFER_SIZE_BYTES 4 #define GRAAL_CUDA_ERROR_NO_BINARY_FOR_GPU 209 /* * Flags for cuMemHostRegister */ #define GRAAL_CU_MEMHOSTREGISTER_PORTABLE 1 #define GRAAL_CU_MEMHOSTREGISTER_DEVICEMAP 2 /** * End of array terminator for the extra parameter to * ::cuLaunchKernel */ #define GRAAL_CU_LAUNCH_PARAM_END ((void*) 0x00) /** * Indicator that the next value in the extra parameter to * ::cuLaunchKernel will be a pointer to a buffer containing all kernel * parameters used for launching kernel f. This buffer needs to * honor all alignment/padding requirements of the individual parameters. * If ::GRAAL_CU_LAUNCH_PARAM_BUFFER_SIZE is not also specified in the * extra array, then ::GRAAL_CU_LAUNCH_PARAM_BUFFER_POINTER will have no * effect. */ #define GRAAL_CU_LAUNCH_PARAM_BUFFER_POINTER ((void*) 0x01) /** * Indicator that the next value in the extra parameter to * ::cuLaunchKernel will be a pointer to a size_t which contains the * size of the buffer specified with ::GRAAL_CU_LAUNCH_PARAM_BUFFER_POINTER. * It is required that ::GRAAL_CU_LAUNCH_PARAM_BUFFER_POINTER also be specified * in the extra array if the value associated with * ::GRAAL_CU_LAUNCH_PARAM_BUFFER_SIZE is not zero. */ #define GRAAL_CU_LAUNCH_PARAM_BUFFER_SIZE ((void*) 0x02) /* * Context creation flags */ #define GRAAL_CU_CTX_MAP_HOST 0x08 #define GRAAL_CU_CTX_SCHED_BLOCKING_SYNC 0x04 /** * Support compute capability 3.0 and later */ #define GRAAL_SUPPORTED_COMPUTE_CAPABILITY_VERSION 3.0 class Ptx : public Gpu { friend class PtxCall; private: static JNINativeMethod PTX_methods[]; // static native boolean initialize(); JNIEXPORT static jboolean initialize(JNIEnv* env, jclass); // static native long generateKernel(byte[] targetCode, String name); JNIEXPORT static jlong generate_kernel(JNIEnv *env, jclass, jbyteArray code_handle, jstring name_handle); // static native long getLaunchKernelAddress(); JNIEXPORT static jlong get_execute_kernel_from_vm_address(JNIEnv *env, jclass); // static native int getAvailableProcessors0(); JNIEXPORT static jint get_total_cores(JNIEnv *env, jobject); JNIEXPORT static void destroy_ptx_context(); // Links the CUDA driver library functions static bool link(); static int ncores(int major, int minor); public: virtual const char* name() { return "PTX"; } // Registers the implementations for the native methods in PTXHotSpotBackend static bool register_natives(JNIEnv* env); #if defined(__x86_64) || defined(AMD64) || defined(_M_AMD64) typedef unsigned long long CUdeviceptr; #else typedef unsigned int CUdeviceptr; #endif typedef int CUdevice; /* CUDA device */ static jlong execute_kernel_from_vm(JavaThread* thread, jlong kernel, jint dimX, jint dimY, jint dimZ, jlong buffer, jint bufferSize, jint objectParametersCount, jlong objectParametersOffsets, jlong pinnedObjects, int encodedReturnTypeSize); private: typedef int (*cuda_cu_init_func_t)(unsigned int); typedef int (*cuda_cu_ctx_create_func_t)(void*, unsigned int, CUdevice); typedef int (*cuda_cu_ctx_destroy_func_t)(void*); typedef int (*cuda_cu_ctx_synchronize_func_t)(void); typedef int (*cuda_cu_ctx_get_current_func_t)(void*); typedef int (*cuda_cu_ctx_set_current_func_t)(void*); typedef int (*cuda_cu_device_get_count_func_t)(int*); typedef int (*cuda_cu_device_get_name_func_t)(char*, int, int); typedef int (*cuda_cu_device_get_func_t)(int*, int); typedef int (*cuda_cu_device_compute_capability_func_t)(int*, int*, int); typedef int (*cuda_cu_device_get_attribute_func_t)(int*, int, int); typedef int (*cuda_cu_launch_kernel_func_t)(struct CUfunc_st*, unsigned int, unsigned int, unsigned int, unsigned int, unsigned int, unsigned int, unsigned int, void*, void**, void**); typedef int (*cuda_cu_module_get_function_func_t)(void*, void*, const char*); typedef int (*cuda_cu_module_load_data_ex_func_t)(void*, void*, unsigned int, void*, void**); typedef int (*cuda_cu_memalloc_func_t)(Ptx::CUdeviceptr*, size_t); typedef int (*cuda_cu_memfree_func_t)(Ptx::CUdeviceptr); typedef int (*cuda_cu_memcpy_htod_func_t)(Ptx::CUdeviceptr, const void*, unsigned int); typedef int (*cuda_cu_memcpy_dtoh_func_t)(const void*, Ptx::CUdeviceptr, unsigned int); typedef int (*cuda_cu_mem_host_register_func_t)(void*, size_t, unsigned int); typedef int (*cuda_cu_mem_host_get_device_pointer_func_t)(Ptx::CUdeviceptr*, void*, unsigned int); typedef int (*cuda_cu_mem_host_unregister_func_t)(void*); public: static cuda_cu_init_func_t _cuda_cu_init; static cuda_cu_ctx_create_func_t _cuda_cu_ctx_create; static cuda_cu_ctx_destroy_func_t _cuda_cu_ctx_destroy; static cuda_cu_ctx_synchronize_func_t _cuda_cu_ctx_synchronize; static cuda_cu_device_get_count_func_t _cuda_cu_device_get_count; static cuda_cu_device_get_name_func_t _cuda_cu_device_get_name; static cuda_cu_device_get_func_t _cuda_cu_device_get; static cuda_cu_device_compute_capability_func_t _cuda_cu_device_compute_capability; /* Deprecated as of CUDA 5.0 */ static cuda_cu_device_get_attribute_func_t _cuda_cu_device_get_attribute; static cuda_cu_launch_kernel_func_t _cuda_cu_launch_kernel; static cuda_cu_module_get_function_func_t _cuda_cu_module_get_function; static cuda_cu_module_load_data_ex_func_t _cuda_cu_module_load_data_ex; static cuda_cu_memalloc_func_t _cuda_cu_memalloc; static cuda_cu_memfree_func_t _cuda_cu_memfree; static cuda_cu_memcpy_htod_func_t _cuda_cu_memcpy_htod; static cuda_cu_memcpy_dtoh_func_t _cuda_cu_memcpy_dtoh; static cuda_cu_ctx_get_current_func_t _cuda_cu_ctx_get_current; static cuda_cu_ctx_set_current_func_t _cuda_cu_ctx_set_current; static cuda_cu_mem_host_register_func_t _cuda_cu_mem_host_register; static cuda_cu_mem_host_get_device_pointer_func_t _cuda_cu_mem_host_get_device_pointer; static cuda_cu_mem_host_unregister_func_t _cuda_cu_mem_host_unregister; protected: static void* _device_context; static int _cu_device; }; #endif // GPU_PTX_HPP