/* StarPU --- Runtime system for heterogeneous multicore architectures. * * Copyright (C) 2009-2019 Université de Bordeaux * Copyright (C) 2011,2012,2017 Inria * Copyright (C) 2010-2017,2019 CNRS * * StarPU is free software; you can redistribute it and/or modify * it under the terms of the GNU Lesser General Public License as published by * the Free Software Foundation; either version 2.1 of the License, or (at * your option) any later version. * * StarPU is distributed in the hope that it will be useful, but * WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. * * See the GNU Lesser General Public License in COPYING.LGPL for more details. */ #include #ifdef STARPU_USE_CUDA static int copy_ram_to_cuda(void *src_interface, unsigned src_node STARPU_ATTRIBUTE_UNUSED, void *dst_interface, unsigned dst_node STARPU_ATTRIBUTE_UNUSED); static int copy_cuda_to_ram(void *src_interface, unsigned src_node STARPU_ATTRIBUTE_UNUSED, void *dst_interface, unsigned dst_node STARPU_ATTRIBUTE_UNUSED); static int copy_ram_to_cuda_async(void *src_interface, unsigned src_node STARPU_ATTRIBUTE_UNUSED, void *dst_interface, unsigned dst_node STARPU_ATTRIBUTE_UNUSED, cudaStream_t stream); static int copy_cuda_to_ram_async(void *src_interface, unsigned src_node STARPU_ATTRIBUTE_UNUSED, void *dst_interface, unsigned dst_node STARPU_ATTRIBUTE_UNUSED, cudaStream_t stream); static int copy_cuda_to_cuda(void *src_interface, unsigned src_node STARPU_ATTRIBUTE_UNUSED, void *dst_interface, unsigned dst_node STARPU_ATTRIBUTE_UNUSED); #endif #ifdef STARPU_USE_OPENCL static int copy_ram_to_opencl(void *src_interface, unsigned src_node STARPU_ATTRIBUTE_UNUSED, void *dst_interface, unsigned dst_node STARPU_ATTRIBUTE_UNUSED); static int copy_opencl_to_ram(void *src_interface, unsigned src_node STARPU_ATTRIBUTE_UNUSED, void *dst_interface, unsigned dst_node STARPU_ATTRIBUTE_UNUSED); static int copy_opencl_to_opencl(void *src_interface, unsigned src_node STARPU_ATTRIBUTE_UNUSED, void *dst_interface, unsigned dst_node STARPU_ATTRIBUTE_UNUSED); static int copy_ram_to_opencl_async(void *src_interface, unsigned src_node STARPU_ATTRIBUTE_UNUSED, void *dst_interface, unsigned dst_node STARPU_ATTRIBUTE_UNUSED, cl_event *event); static int copy_opencl_to_ram_async(void *src_interface, unsigned src_node STARPU_ATTRIBUTE_UNUSED, void *dst_interface, unsigned dst_node STARPU_ATTRIBUTE_UNUSED, cl_event *event); static int copy_opencl_to_opencl_async(void *src_interface, unsigned src_node STARPU_ATTRIBUTE_UNUSED, void *dst_interface, unsigned dst_node STARPU_ATTRIBUTE_UNUSED, cl_event *event); #endif static int copy_any_to_any(void *src_interface, unsigned src_node, void *dst_interface, unsigned dst_node, void *async_data); static const struct starpu_data_copy_methods tensor_copy_data_methods_s = { #ifdef STARPU_USE_CUDA .ram_to_cuda = copy_ram_to_cuda, .cuda_to_ram = copy_cuda_to_ram, .ram_to_cuda_async = copy_ram_to_cuda_async, .cuda_to_ram_async = copy_cuda_to_ram_async, .cuda_to_cuda = copy_cuda_to_cuda, #endif #ifdef STARPU_USE_OPENCL .ram_to_opencl = copy_ram_to_opencl, .opencl_to_ram = copy_opencl_to_ram, .opencl_to_opencl = copy_opencl_to_opencl, .ram_to_opencl_async = copy_ram_to_opencl_async, .opencl_to_ram_async = copy_opencl_to_ram_async, .opencl_to_opencl_async = copy_opencl_to_opencl_async, #endif .any_to_any = copy_any_to_any, }; static void register_tensor_handle(starpu_data_handle_t handle, unsigned home_node, void *data_interface); static void *tensor_to_pointer(void *data_interface, unsigned node); static int tensor_pointer_is_inside(void *data_interface, unsigned node, void *ptr); static starpu_ssize_t allocate_tensor_buffer_on_node(void *data_interface_, unsigned dst_node); static void free_tensor_buffer_on_node(void *data_interface, unsigned node); static size_t tensor_interface_get_size(starpu_data_handle_t handle); static uint32_t footprint_tensor_interface_crc32(starpu_data_handle_t handle); static int tensor_compare(void *data_interface_a, void *data_interface_b); static void display_tensor_interface(starpu_data_handle_t handle, FILE *f); static int pack_tensor_handle(starpu_data_handle_t handle, unsigned node, void **ptr, starpu_ssize_t *count); static int unpack_tensor_handle(starpu_data_handle_t handle, unsigned node, void *ptr, size_t count); static starpu_ssize_t describe(void *data_interface, char *buf, size_t size); struct starpu_data_interface_ops starpu_interface_tensor_ops = { .register_data_handle = register_tensor_handle, .allocate_data_on_node = allocate_tensor_buffer_on_node, .to_pointer = tensor_to_pointer, .pointer_is_inside = tensor_pointer_is_inside, .free_data_on_node = free_tensor_buffer_on_node, .copy_methods = &tensor_copy_data_methods_s, .get_size = tensor_interface_get_size, .footprint = footprint_tensor_interface_crc32, .compare = tensor_compare, .interfaceid = STARPU_TENSOR_INTERFACE_ID, .interface_size = sizeof(struct starpu_tensor_interface), .display = display_tensor_interface, .pack_data = pack_tensor_handle, .unpack_data = unpack_tensor_handle, .describe = describe, .name = "STARPU_TENSOR_INTERFACE" }; static void *tensor_to_pointer(void *data_interface, unsigned node) { (void) node; struct starpu_tensor_interface *tensor_interface = data_interface; return (void*) tensor_interface->ptr; } static int tensor_pointer_is_inside(void *data_interface, unsigned node, void *ptr) { (void) node; struct starpu_tensor_interface *tensor_interface = data_interface; uint32_t ldy = tensor_interface->ldy; uint32_t ldz = tensor_interface->ldz; uint32_t ldt = tensor_interface->ldt; uint32_t nx = tensor_interface->nx; uint32_t ny = tensor_interface->ny; uint32_t nz = tensor_interface->nz; uint32_t nt = tensor_interface->nt; size_t elemsize = tensor_interface->elemsize; return (char*) ptr >= (char*) tensor_interface->ptr && (char*) ptr < (char*) tensor_interface->ptr + (nt-1)*ldt*elemsize + (nz-1)*ldz*elemsize + (ny-1)*ldy*elemsize + nx*elemsize; } static void register_tensor_handle(starpu_data_handle_t handle, unsigned home_node, void *data_interface) { struct starpu_tensor_interface *tensor_interface = (struct starpu_tensor_interface *) data_interface; unsigned node; for (node = 0; node < STARPU_MAXNODES; node++) { struct starpu_tensor_interface *local_interface = (struct starpu_tensor_interface *) starpu_data_get_interface_on_node(handle, node); if (node == home_node) { local_interface->ptr = tensor_interface->ptr; local_interface->dev_handle = tensor_interface->dev_handle; local_interface->offset = tensor_interface->offset; local_interface->ldy = tensor_interface->ldy; local_interface->ldz = tensor_interface->ldz; local_interface->ldt = tensor_interface->ldt; } else { local_interface->ptr = 0; local_interface->dev_handle = 0; local_interface->offset = 0; local_interface->ldy = 0; local_interface->ldz = 0; local_interface->ldt = 0; } local_interface->id = tensor_interface->id; local_interface->nx = tensor_interface->nx; local_interface->ny = tensor_interface->ny; local_interface->nz = tensor_interface->nz; local_interface->nt = tensor_interface->nt; local_interface->elemsize = tensor_interface->elemsize; } } /* declare a new data with the BLAS interface */ void starpu_tensor_data_register(starpu_data_handle_t *handleptr, int home_node, uintptr_t ptr, uint32_t ldy, uint32_t ldz, uint32_t ldt, uint32_t nx, uint32_t ny, uint32_t nz, uint32_t nt, size_t elemsize) { struct starpu_tensor_interface tensor_interface = { .id = STARPU_TENSOR_INTERFACE_ID, .ptr = ptr, .dev_handle = ptr, .offset = 0, .ldy = ldy, .ldz = ldz, .ldt = ldt, .nx = nx, .ny = ny, .nz = nz, .nt = nt, .elemsize = elemsize }; #ifndef STARPU_SIMGRID if (home_node >= 0 && starpu_node_get_kind(home_node) == STARPU_CPU_RAM) { STARPU_ASSERT_ACCESSIBLE(ptr); STARPU_ASSERT_ACCESSIBLE(ptr + (nt-1)*ldt*elemsize + (nz-1)*ldz*elemsize + (ny-1)*ldy*elemsize + nx*elemsize - 1); } #endif starpu_data_register(handleptr, home_node, &tensor_interface, &starpu_interface_tensor_ops); } void starpu_tensor_ptr_register(starpu_data_handle_t handle, unsigned node, uintptr_t ptr, uintptr_t dev_handle, size_t offset, uint32_t ldy, uint32_t ldz, uint32_t ldt) { struct starpu_tensor_interface *tensor_interface = starpu_data_get_interface_on_node(handle, node); starpu_data_ptr_register(handle, node); tensor_interface->ptr = ptr; tensor_interface->dev_handle = dev_handle; tensor_interface->offset = offset; tensor_interface->ldy = ldy; tensor_interface->ldz = ldz; tensor_interface->ldt = ldt; } static uint32_t footprint_tensor_interface_crc32(starpu_data_handle_t handle) { uint32_t hash; hash = starpu_hash_crc32c_be(starpu_tensor_get_nx(handle), 0); hash = starpu_hash_crc32c_be(starpu_tensor_get_ny(handle), hash); hash = starpu_hash_crc32c_be(starpu_tensor_get_nz(handle), hash); hash = starpu_hash_crc32c_be(starpu_tensor_get_nt(handle), hash); return hash; } static int tensor_compare(void *data_interface_a, void *data_interface_b) { struct starpu_tensor_interface *tensor_a = (struct starpu_tensor_interface *) data_interface_a; struct starpu_tensor_interface *tensor_b = (struct starpu_tensor_interface *) data_interface_b; /* Two matricess are considered compatible if they have the same size */ return (tensor_a->nx == tensor_b->nx) && (tensor_a->ny == tensor_b->ny) && (tensor_a->nz == tensor_b->nz) && (tensor_a->nt == tensor_b->nt) && (tensor_a->elemsize == tensor_b->elemsize); } static void display_tensor_interface(starpu_data_handle_t handle, FILE *f) { struct starpu_tensor_interface *tensor_interface; tensor_interface = (struct starpu_tensor_interface *) starpu_data_get_interface_on_node(handle, STARPU_MAIN_RAM); fprintf(f, "%u\t%u\t%u\t%u\t", tensor_interface->nx, tensor_interface->ny, tensor_interface->nz, tensor_interface->nt); } static int pack_tensor_handle(starpu_data_handle_t handle, unsigned node, void **ptr, starpu_ssize_t *count) { STARPU_ASSERT(starpu_data_test_if_allocated_on_node(handle, node)); struct starpu_tensor_interface *tensor_interface = (struct starpu_tensor_interface *) starpu_data_get_interface_on_node(handle, node); *count = tensor_interface->nx*tensor_interface->ny*tensor_interface->nz*tensor_interface->nt*tensor_interface->elemsize; if (ptr != NULL) { uint32_t t, z, y; char *block = (void *)tensor_interface->ptr; *ptr = (void *)starpu_malloc_on_node_flags(node, *count, 0); char *cur = *ptr; char *block_t = block; for(t=0 ; tnt ; t++) { char *block_z = block_t; for(z=0 ; znz ; z++) { char *block_y = block_z; for(y=0 ; yny ; y++) { memcpy(cur, block_y, tensor_interface->nx*tensor_interface->elemsize); cur += tensor_interface->nx*tensor_interface->elemsize; block_y += tensor_interface->ldy * tensor_interface->elemsize; } block_z += tensor_interface->ldz * tensor_interface->elemsize; } block_t += tensor_interface->ldt * tensor_interface->elemsize; } } return 0; } static int unpack_tensor_handle(starpu_data_handle_t handle, unsigned node, void *ptr, size_t count) { STARPU_ASSERT(starpu_data_test_if_allocated_on_node(handle, node)); struct starpu_tensor_interface *tensor_interface = (struct starpu_tensor_interface *) starpu_data_get_interface_on_node(handle, node); STARPU_ASSERT(count == tensor_interface->elemsize * tensor_interface->nx * tensor_interface->ny * tensor_interface->nz * tensor_interface->nt); uint32_t t, z, y; char *cur = ptr; char *block = (void *)tensor_interface->ptr; char *block_t = block; for(t=0 ; tnt ; t++) { char *block_z = block_t; for(z=0 ; znz ; z++) { char *block_y = block_z; for(y=0 ; yny ; y++) { memcpy(block_y, cur, tensor_interface->nx*tensor_interface->elemsize); cur += tensor_interface->nx*tensor_interface->elemsize; block_y += tensor_interface->ldy * tensor_interface->elemsize; } block_z += tensor_interface->ldz * tensor_interface->elemsize; } block_t += tensor_interface->ldt * tensor_interface->elemsize; } starpu_free_on_node_flags(node, (uintptr_t)ptr, count, 0); return 0; } static size_t tensor_interface_get_size(starpu_data_handle_t handle) { size_t size; struct starpu_tensor_interface *tensor_interface; tensor_interface = (struct starpu_tensor_interface *) starpu_data_get_interface_on_node(handle, STARPU_MAIN_RAM); #ifdef STARPU_DEBUG STARPU_ASSERT_MSG(tensor_interface->id == STARPU_TENSOR_INTERFACE_ID, "Error. The given data is not a block."); #endif size = tensor_interface->nx*tensor_interface->ny*tensor_interface->nz*tensor_interface->nt*tensor_interface->elemsize; return size; } /* offer an access to the data parameters */ uint32_t starpu_tensor_get_nx(starpu_data_handle_t handle) { struct starpu_tensor_interface *tensor_interface = (struct starpu_tensor_interface *) starpu_data_get_interface_on_node(handle, STARPU_MAIN_RAM); #ifdef STARPU_DEBUG STARPU_ASSERT_MSG(tensor_interface->id == STARPU_TENSOR_INTERFACE_ID, "Error. The given data is not a block."); #endif return tensor_interface->nx; } uint32_t starpu_tensor_get_ny(starpu_data_handle_t handle) { struct starpu_tensor_interface *tensor_interface = (struct starpu_tensor_interface *) starpu_data_get_interface_on_node(handle, STARPU_MAIN_RAM); #ifdef STARPU_DEBUG STARPU_ASSERT_MSG(tensor_interface->id == STARPU_TENSOR_INTERFACE_ID, "Error. The given data is not a block."); #endif return tensor_interface->ny; } uint32_t starpu_tensor_get_nz(starpu_data_handle_t handle) { struct starpu_tensor_interface *tensor_interface = (struct starpu_tensor_interface *) starpu_data_get_interface_on_node(handle, STARPU_MAIN_RAM); #ifdef STARPU_DEBUG STARPU_ASSERT_MSG(tensor_interface->id == STARPU_TENSOR_INTERFACE_ID, "Error. The given data is not a block."); #endif return tensor_interface->nz; } uint32_t starpu_tensor_get_nt(starpu_data_handle_t handle) { struct starpu_tensor_interface *tensor_interface = (struct starpu_tensor_interface *) starpu_data_get_interface_on_node(handle, STARPU_MAIN_RAM); #ifdef STARPU_DEBUG STARPU_ASSERT_MSG(tensor_interface->id == STARPU_TENSOR_INTERFACE_ID, "Error. The given data is not a block."); #endif return tensor_interface->nt; } uint32_t starpu_tensor_get_local_ldy(starpu_data_handle_t handle) { unsigned node; node = starpu_worker_get_local_memory_node(); STARPU_ASSERT(starpu_data_test_if_allocated_on_node(handle, node)); struct starpu_tensor_interface *tensor_interface = (struct starpu_tensor_interface *) starpu_data_get_interface_on_node(handle, node); #ifdef STARPU_DEBUG STARPU_ASSERT_MSG(tensor_interface->id == STARPU_TENSOR_INTERFACE_ID, "Error. The given data is not a block."); #endif return tensor_interface->ldy; } uint32_t starpu_tensor_get_local_ldz(starpu_data_handle_t handle) { unsigned node; node = starpu_worker_get_local_memory_node(); STARPU_ASSERT(starpu_data_test_if_allocated_on_node(handle, node)); struct starpu_tensor_interface *tensor_interface = (struct starpu_tensor_interface *) starpu_data_get_interface_on_node(handle, node); #ifdef STARPU_DEBUG STARPU_ASSERT_MSG(tensor_interface->id == STARPU_TENSOR_INTERFACE_ID, "Error. The given data is not a block."); #endif return tensor_interface->ldz; } uint32_t starpu_tensor_get_local_ldt(starpu_data_handle_t handle) { unsigned node; node = starpu_worker_get_local_memory_node(); STARPU_ASSERT(starpu_data_test_if_allocated_on_node(handle, node)); struct starpu_tensor_interface *tensor_interface = (struct starpu_tensor_interface *) starpu_data_get_interface_on_node(handle, node); #ifdef STARPU_DEBUG STARPU_ASSERT_MSG(tensor_interface->id == STARPU_TENSOR_INTERFACE_ID, "Error. The given data is not a block."); #endif return tensor_interface->ldt; } uintptr_t starpu_tensor_get_local_ptr(starpu_data_handle_t handle) { unsigned node; node = starpu_worker_get_local_memory_node(); STARPU_ASSERT(starpu_data_test_if_allocated_on_node(handle, node)); struct starpu_tensor_interface *tensor_interface = (struct starpu_tensor_interface *) starpu_data_get_interface_on_node(handle, node); #ifdef STARPU_DEBUG STARPU_ASSERT_MSG(tensor_interface->id == STARPU_TENSOR_INTERFACE_ID, "Error. The given data is not a block."); #endif return tensor_interface->ptr; } size_t starpu_tensor_get_elemsize(starpu_data_handle_t handle) { struct starpu_tensor_interface *tensor_interface = (struct starpu_tensor_interface *) starpu_data_get_interface_on_node(handle, STARPU_MAIN_RAM); #ifdef STARPU_DEBUG STARPU_ASSERT_MSG(tensor_interface->id == STARPU_TENSOR_INTERFACE_ID, "Error. The given data is not a block."); #endif return tensor_interface->elemsize; } /* memory allocation/deallocation primitives for the BLOCK interface */ /* returns the size of the allocated area */ static starpu_ssize_t allocate_tensor_buffer_on_node(void *data_interface_, unsigned dst_node) { uintptr_t addr = 0, handle; struct starpu_tensor_interface *dst_block = (struct starpu_tensor_interface *) data_interface_; uint32_t nx = dst_block->nx; uint32_t ny = dst_block->ny; uint32_t nz = dst_block->nz; uint32_t nt = dst_block->nt; size_t elemsize = dst_block->elemsize; starpu_ssize_t allocated_memory; handle = starpu_malloc_on_node(dst_node, nx*ny*nz*nt*elemsize); if (!handle) return -ENOMEM; if (starpu_node_get_kind(dst_node) != STARPU_OPENCL_RAM) addr = handle; allocated_memory = nx*ny*nz*nt*elemsize; /* update the data properly in consequence */ dst_block->ptr = addr; dst_block->dev_handle = handle; dst_block->offset = 0; dst_block->ldy = nx; dst_block->ldz = nx*ny; dst_block->ldt = nx*ny*nz; return allocated_memory; } static void free_tensor_buffer_on_node(void *data_interface, unsigned node) { struct starpu_tensor_interface *tensor_interface = (struct starpu_tensor_interface *) data_interface; uint32_t nx = tensor_interface->nx; uint32_t ny = tensor_interface->ny; uint32_t nz = tensor_interface->nz; uint32_t nt = tensor_interface->nt; size_t elemsize = tensor_interface->elemsize; starpu_free_on_node(node, tensor_interface->dev_handle, nx*ny*nz*nt*elemsize); } #ifdef STARPU_USE_CUDA static int copy_cuda_common(void *src_interface, unsigned src_node STARPU_ATTRIBUTE_UNUSED, void *dst_interface, unsigned dst_node STARPU_ATTRIBUTE_UNUSED, enum cudaMemcpyKind kind) { struct starpu_tensor_interface *src_block = src_interface; struct starpu_tensor_interface *dst_block = dst_interface; uint32_t nx = src_block->nx; uint32_t ny = src_block->ny; uint32_t nz = src_block->nz; uint32_t nt = src_block->nt; size_t elemsize = src_block->elemsize; cudaError_t cures; if (src_block->ldy == dst_block->ldy && src_block->ldz == dst_block->ldz && src_block->ldt == dst_block->ldt && nx*ny*nz == src_block->ldt) { /* Same lds on both sides, and contiguous, simple */ starpu_cuda_copy_async_sync((void *)src_block->ptr, src_node, (void *)dst_block->ptr, dst_node, nx*ny*nz*nt*elemsize, NULL, kind); } else { /* TODO: use cudaMemcpy2D for whole 3D blocks etc. when they are contiguous */ /* Default case: we transfer all blocks one by one: nz transfers */ /* TODO: use cudaMemcpy3D now that it works (except on cuda 4.2) */ unsigned t; for (t = 0; t < src_block->nt; t++) { unsigned z; for (z = 0; z < src_block->nz; z++) { uint8_t *src_ptr = ((uint8_t *)src_block->ptr) + t*src_block->ldt*src_block->elemsize + z*src_block->ldz*src_block->elemsize; uint8_t *dst_ptr = ((uint8_t *)dst_block->ptr) + t*dst_block->ldt*src_block->elemsize + z*dst_block->ldz*dst_block->elemsize; cures = cudaMemcpy2D((char *)dst_ptr, dst_block->ldy*elemsize, (char *)src_ptr, src_block->ldy*elemsize, nx*elemsize, ny, kind); if (!cures) cures = cudaThreadSynchronize(); if (STARPU_UNLIKELY(cures)) STARPU_CUDA_REPORT_ERROR(cures); } } } starpu_interface_data_copy(src_node, dst_node, src_block->nx*src_block->ny*src_block->nz*src_block->nt*src_block->elemsize); return 0; } static int copy_cuda_async_common(void *src_interface, unsigned src_node STARPU_ATTRIBUTE_UNUSED, void *dst_interface, unsigned dst_node STARPU_ATTRIBUTE_UNUSED, cudaStream_t stream, enum cudaMemcpyKind kind) { struct starpu_tensor_interface *src_block = src_interface; struct starpu_tensor_interface *dst_block = dst_interface; uint32_t nx = src_block->nx; uint32_t ny = src_block->ny; uint32_t nz = src_block->nz; uint32_t nt = src_block->nt; size_t elemsize = src_block->elemsize; cudaError_t cures; int ret; if (src_block->ldy == dst_block->ldy && src_block->ldz == dst_block->ldz && src_block->ldt == dst_block->ldt && nx*ny*nz == src_block->ldt) { /* Same lds on both sides, and contiguous, simple */ ret = starpu_cuda_copy_async_sync((void *)src_block->ptr, src_node, (void *)dst_block->ptr, dst_node, nx*ny*nz*nt*elemsize, stream, kind); } else { /* TODO: use cudaMemcpy2D for whole 3D blocks etc. when they are contiguous */ /* Default case: we transfer all blocks one by one: nz transfers */ /* TODO: use cudaMemcpy3D now that it works (except on cuda 4.2) */ unsigned t; for (t = 0; t < src_block->nt; t++) { unsigned z; for (z = 0; z < src_block->nz; z++) { uint8_t *src_ptr = ((uint8_t *)src_block->ptr) + t*src_block->ldt*src_block->elemsize + z*src_block->ldz*src_block->elemsize; uint8_t *dst_ptr = ((uint8_t *)dst_block->ptr) + t*dst_block->ldt*dst_block->elemsize + z*dst_block->ldz*dst_block->elemsize; double start; starpu_interface_start_driver_copy_async(src_node, dst_node, &start); cures = cudaMemcpy2DAsync((char *)dst_ptr, dst_block->ldy*elemsize, (char *)src_ptr, src_block->ldy*elemsize, nx*elemsize, ny, kind, stream); starpu_interface_end_driver_copy_async(src_node, dst_node, start); if (STARPU_UNLIKELY(cures)) { /* I don't know how to do that "better" */ goto no_async_default; } } } ret = -EAGAIN; } starpu_interface_data_copy(src_node, dst_node, src_block->nx*src_block->ny*src_block->nz*src_block->nt*src_block->elemsize); return ret; no_async_default: { unsigned t; for (t = 0; t < src_block->nt; t++) { unsigned z; for (z = 0; z < src_block->nz; z++) { uint8_t *src_ptr = ((uint8_t *)src_block->ptr) + t*src_block->ldt*src_block->elemsize + z*src_block->ldz*src_block->elemsize; uint8_t *dst_ptr = ((uint8_t *)dst_block->ptr) + t*dst_block->ldt*dst_block->elemsize + z*dst_block->ldz*dst_block->elemsize; cures = cudaMemcpy2D((char *)dst_ptr, dst_block->ldy*elemsize, (char *)src_ptr, src_block->ldy*elemsize, nx*elemsize, ny, kind); if (!cures) cures = cudaThreadSynchronize(); if (STARPU_UNLIKELY(cures)) STARPU_CUDA_REPORT_ERROR(cures); } } starpu_interface_data_copy(src_node, dst_node, src_block->nx*src_block->ny*src_block->nz*src_block->nt*src_block->elemsize); return 0; } } static int copy_cuda_to_ram(void *src_interface, unsigned src_node, void *dst_interface, unsigned dst_node) { return copy_cuda_common(src_interface, src_node, dst_interface, dst_node, cudaMemcpyDeviceToHost); } static int copy_ram_to_cuda(void *src_interface, unsigned src_node STARPU_ATTRIBUTE_UNUSED, void *dst_interface, unsigned dst_node STARPU_ATTRIBUTE_UNUSED) { return copy_cuda_common(src_interface, src_node, dst_interface, dst_node, cudaMemcpyHostToDevice); } static int copy_cuda_to_cuda(void *src_interface, unsigned src_node STARPU_ATTRIBUTE_UNUSED, void *dst_interface, unsigned dst_node STARPU_ATTRIBUTE_UNUSED) { return copy_cuda_common(src_interface, src_node, dst_interface, dst_node, cudaMemcpyDeviceToDevice); } static int copy_cuda_to_ram_async(void *src_interface, unsigned src_node STARPU_ATTRIBUTE_UNUSED, void *dst_interface, unsigned dst_node STARPU_ATTRIBUTE_UNUSED, cudaStream_t stream) { return copy_cuda_async_common(src_interface, src_node, dst_interface, dst_node, stream, cudaMemcpyDeviceToHost); } static int copy_ram_to_cuda_async(void *src_interface, unsigned src_node STARPU_ATTRIBUTE_UNUSED, void *dst_interface, unsigned dst_node STARPU_ATTRIBUTE_UNUSED, cudaStream_t stream) { return copy_cuda_async_common(src_interface, src_node, dst_interface, dst_node, stream, cudaMemcpyHostToDevice); } #endif // STARPU_USE_CUDA #ifdef STARPU_USE_OPENCL static int copy_opencl_common(void *src_interface, unsigned src_node, void *dst_interface, unsigned dst_node, cl_event *event) { struct starpu_tensor_interface *src_block = src_interface; struct starpu_tensor_interface *dst_block = dst_interface; int ret = 0; uint32_t nx = src_block->nx; uint32_t ny = src_block->ny; uint32_t nz = src_block->nz; /* We may have a contiguous buffer for the entire block, or contiguous * plans within the block, we can avoid many small transfers that way */ if (src_block->ldy == dst_block->ldy && src_block->ldz == dst_block->ldz && src_block->ldt == dst_block->ldt && nx*ny*nz == src_block->ldt) { ret = starpu_opencl_copy_async_sync(src_block->dev_handle, src_block->offset, src_node, dst_block->dev_handle, dst_block->offset, dst_node, src_block->nx*src_block->ny*src_block->nz*src_block->nt*src_block->elemsize, event); } else { /* Default case: we transfer all lines one by one: ny*nz transfers */ /* TODO: rect support */ unsigned t; for (t = 0; t < src_block->nt; t++) { unsigned z; for (z = 0; z < src_block->nz; z++) { unsigned j; for(j=0 ; jny ; j++) { ret = starpu_opencl_copy_async_sync(src_block->dev_handle, src_block->offset + t*src_block->ldt*src_block->elemsize + z*src_block->ldz*src_block->elemsize + j*src_block->ldy*src_block->elemsize, src_node, dst_block->dev_handle, dst_block->offset + t*dst_block->ldt*dst_block->elemsize + z*dst_block->ldz*dst_block->elemsize + j*dst_block->ldy*dst_block->elemsize, dst_node, src_block->nx*src_block->elemsize, event); } } } } starpu_interface_data_copy(src_node, dst_node, src_block->nx*src_block->ny*src_block->nz*src_block->nt*src_block->elemsize); return ret; } static int copy_ram_to_opencl_async(void *src_interface, unsigned src_node, void *dst_interface, unsigned dst_node, cl_event *event) { return copy_opencl_common(src_interface, src_node, dst_interface, dst_node, event); } static int copy_opencl_to_ram_async(void *src_interface, unsigned src_node, void *dst_interface, unsigned dst_node, cl_event *event) { return copy_opencl_common(src_interface, src_node, dst_interface, dst_node, event); } static int copy_opencl_to_opencl_async(void *src_interface, unsigned src_node, void *dst_interface, unsigned dst_node, cl_event *event) { return copy_opencl_common(src_interface, src_node, dst_interface, dst_node, event); } static int copy_ram_to_opencl(void *src_interface, unsigned src_node STARPU_ATTRIBUTE_UNUSED, void *dst_interface, unsigned dst_node STARPU_ATTRIBUTE_UNUSED) { return copy_ram_to_opencl_async(src_interface, src_node, dst_interface, dst_node, NULL); } static int copy_opencl_to_ram(void *src_interface, unsigned src_node STARPU_ATTRIBUTE_UNUSED, void *dst_interface, unsigned dst_node STARPU_ATTRIBUTE_UNUSED) { return copy_opencl_to_ram_async(src_interface, src_node, dst_interface, dst_node, NULL); } static int copy_opencl_to_opencl(void *src_interface, unsigned src_node STARPU_ATTRIBUTE_UNUSED, void *dst_interface, unsigned dst_node STARPU_ATTRIBUTE_UNUSED) { return copy_opencl_to_opencl_async(src_interface, src_node, dst_interface, dst_node, NULL); } #endif static int copy_any_to_any(void *src_interface, unsigned src_node, void *dst_interface, unsigned dst_node, void *async_data) { struct starpu_tensor_interface *src_block = (struct starpu_tensor_interface *) src_interface; struct starpu_tensor_interface *dst_block = (struct starpu_tensor_interface *) dst_interface; int ret = 0; uint32_t nx = dst_block->nx; uint32_t ny = dst_block->ny; uint32_t nz = dst_block->nz; uint32_t nt = dst_block->nt; size_t elemsize = dst_block->elemsize; uint32_t ldy_src = src_block->ldy; uint32_t ldz_src = src_block->ldz; uint32_t ldt_src = src_block->ldt; uint32_t ldy_dst = dst_block->ldy; uint32_t ldz_dst = dst_block->ldz; uint32_t ldt_dst = dst_block->ldt; if (ldy_src == nx && ldy_dst == nx && ldz_src == nx*ny && ldz_dst == nx*ny && ldt_src == nx*ny*nz && ldt_dst == nx*ny*nz) { /* Optimise non-partitioned and z-partitioned case */ if (starpu_interface_copy(src_block->dev_handle, src_block->offset, src_node, dst_block->dev_handle, dst_block->offset, dst_node, nx*ny*nz*nt*elemsize, async_data)) ret = -EAGAIN; } else { unsigned t; for (t = 0; t < nt; t++) { unsigned z; for (z = 0; z < nz; z++) { if (ldy_src == nx && ldy_dst == nx) { /* Optimise y-partitioned case */ uint32_t src_offset = t*ldt_src*elemsize + z*ldz_src*elemsize; uint32_t dst_offset = t*ldt_dst*elemsize + z*ldz_dst*elemsize; if (starpu_interface_copy(src_block->dev_handle, src_block->offset + src_offset, src_node, dst_block->dev_handle, dst_block->offset + dst_offset, dst_node, nx*ny*elemsize, async_data)) ret = -EAGAIN; } else { unsigned y; for (y = 0; y < ny; y++) { /* Eerf, x-partitioned case */ uint32_t src_offset = (y*ldy_src + z*ldz_src + t*ldt_src)*elemsize; uint32_t dst_offset = (y*ldy_dst + z*ldz_dst + t*ldt_dst)*elemsize; if (starpu_interface_copy(src_block->dev_handle, src_block->offset + src_offset, src_node, dst_block->dev_handle, dst_block->offset + dst_offset, dst_node, nx*elemsize, async_data)) ret = -EAGAIN; } } } } } starpu_interface_data_copy(src_node, dst_node, nx*ny*nz*nt*elemsize); return ret; } static starpu_ssize_t describe(void *data_interface, char *buf, size_t size) { struct starpu_tensor_interface *block = (struct starpu_tensor_interface *) data_interface; return snprintf(buf, size, "T%ux%ux%ux%ux%u", (unsigned) block->nx, (unsigned) block->ny, (unsigned) block->nz, (unsigned) block->nt, (unsigned) block->elemsize); }