implicit-stencil-blocks.c 12 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449
  1. /* StarPU --- Runtime system for heterogeneous multicore architectures.
  2. *
  3. * Copyright (C) 2016,2017 Inria
  4. * Copyright (C) 2016,2017 CNRS
  5. * Copyright (C) 2010,2013-2017 Université de Bordeaux
  6. *
  7. * StarPU is free software; you can redistribute it and/or modify
  8. * it under the terms of the GNU Lesser General Public License as published by
  9. * the Free Software Foundation; either version 2.1 of the License, or (at
  10. * your option) any later version.
  11. *
  12. * StarPU is distributed in the hope that it will be useful, but
  13. * WITHOUT ANY WARRANTY; without even the implied warranty of
  14. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
  15. *
  16. * See the GNU Lesser General Public License in COPYING.LGPL for more details.
  17. */
  18. #include "implicit-stencil.h"
  19. #include <math.h>
  20. /* Manage block and tags allocation */
  21. static struct block_description *blocks;
  22. static unsigned sizex, sizey, sizez;
  23. static unsigned nbz;
  24. static unsigned *block_sizes_z;
  25. /*
  26. * Tags for various codelet completion
  27. */
  28. /*
  29. * common tag format:
  30. */
  31. static starpu_tag_t tag_common(int z, int dir, int type)
  32. {
  33. return (((((starpu_tag_t)type) << 4) | ((dir+1)/2)) << 32)|(starpu_tag_t)z;
  34. }
  35. /* Completion of last update tasks */
  36. starpu_tag_t TAG_FINISH(int z)
  37. {
  38. z = (z + nbz)%nbz;
  39. starpu_tag_t tag = tag_common(z, 0, 1);
  40. return tag;
  41. }
  42. /* Completion of the save codelet for MPI send/recv */
  43. starpu_tag_t TAG_START(int z, int dir)
  44. {
  45. z = (z + nbz)%nbz;
  46. starpu_tag_t tag = tag_common(z, dir, 2);
  47. return tag;
  48. }
  49. /*
  50. * common MPI tag format:
  51. */
  52. static int mpi_tag_common(int z, int dir, int layer_or_boundary, int buffer)
  53. {
  54. return (z<<12) | (layer_or_boundary << 8) | ((((1+dir)/2))<<4) | buffer;
  55. }
  56. int MPI_TAG_LAYERS(int z, int buffer)
  57. {
  58. z = (z + nbz)%nbz;
  59. /* No direction for layers ; layer is 0 */
  60. int tag = mpi_tag_common(z, 0, 0, buffer);
  61. return tag;
  62. }
  63. int MPI_TAG_BOUNDARIES(int z, int dir, int buffer)
  64. {
  65. z = (z + nbz)%nbz;
  66. int tag = mpi_tag_common(z, dir, 1, buffer);
  67. return tag;
  68. }
  69. /*
  70. * Block descriptors
  71. */
  72. /* Compute the size of the different blocks */
  73. static void compute_block_sizes(void)
  74. {
  75. block_sizes_z = (unsigned *) malloc(nbz*sizeof(unsigned));
  76. STARPU_ASSERT(block_sizes_z);
  77. /* Perhaps the last chunk is smaller */
  78. unsigned default_block_size = (sizez+nbz-1)/nbz;
  79. unsigned remaining = sizez;
  80. unsigned b;
  81. for (b = 0; b < nbz; b++)
  82. {
  83. block_sizes_z[b] = MIN(default_block_size, remaining);
  84. remaining -= block_sizes_z[b];
  85. }
  86. STARPU_ASSERT(remaining == 0);
  87. }
  88. unsigned get_block_size(int bz)
  89. {
  90. return block_sizes_z[bz];
  91. }
  92. struct block_description *get_block_description(int z)
  93. {
  94. z = (z + nbz)%nbz;
  95. STARPU_ASSERT(&blocks[z]);
  96. return &blocks[z];
  97. }
  98. int get_block_mpi_node(int z)
  99. {
  100. z = (z + nbz)%nbz;
  101. return blocks[z].mpi_node;
  102. }
  103. void create_blocks_array(unsigned _sizex, unsigned _sizey, unsigned _sizez, unsigned _nbz)
  104. {
  105. /* Store the parameters */
  106. nbz = _nbz;
  107. sizex = _sizex;
  108. sizey = _sizey;
  109. sizez = _sizez;
  110. /* Create a grid of block descriptors */
  111. blocks = (struct block_description *) calloc(nbz, sizeof(struct block_description));
  112. STARPU_ASSERT(blocks);
  113. /* What is the size of the different blocks ? */
  114. compute_block_sizes();
  115. unsigned bz;
  116. for (bz = 0; bz < nbz; bz++)
  117. {
  118. struct block_description * block =
  119. get_block_description(bz);
  120. /* Which block is it ? */
  121. block->bz = bz;
  122. /* For simplicity, we store which are the neighbours blocks */
  123. block->boundary_blocks[B] = get_block_description((bz-1+nbz)%nbz);
  124. block->boundary_blocks[T] = get_block_description((bz+1)%nbz);
  125. }
  126. }
  127. void free_blocks_array()
  128. {
  129. free(blocks);
  130. free(block_sizes_z);
  131. }
  132. /*
  133. * Initialization of the blocks
  134. */
  135. void assign_blocks_to_workers(int rank)
  136. {
  137. unsigned bz;
  138. /* NB: perhaps we could count a GPU as multiple workers */
  139. /* how many workers are there ? */
  140. /*unsigned nworkers = starpu_worker_get_count();*/
  141. /* how many blocks are on that MPI node ? */
  142. unsigned nblocks = 0;
  143. for (bz = 0; bz < nbz; bz++)
  144. {
  145. struct block_description *block =
  146. get_block_description(bz);
  147. if (block->mpi_node == rank)
  148. nblocks++;
  149. }
  150. /* how many blocks per worker ? */
  151. /*unsigned nblocks_per_worker = (nblocks + nworkers - 1)/nworkers;*/
  152. /* we now attribute up to nblocks_per_worker blocks per workers */
  153. unsigned attributed = 0;
  154. for (bz = 0; bz < nbz; bz++)
  155. {
  156. struct block_description *block =
  157. get_block_description(bz);
  158. if (block->mpi_node == rank)
  159. {
  160. unsigned workerid;
  161. /* Manage initial block distribution between CPU and GPU */
  162. #if 0
  163. #if 1
  164. /* GPUs then CPUs */
  165. if (attributed < 3*18)
  166. workerid = attributed / 18;
  167. else
  168. workerid = 3+ (attributed - 3*18) / 2;
  169. #else
  170. /* GPUs interleaved with CPUs */
  171. if ((attributed % 20) <= 1)
  172. workerid = 3 + attributed / 20;
  173. else if (attributed < 60)
  174. workerid = attributed / 20;
  175. else
  176. workerid = (attributed - 60)/2 + 6;
  177. #endif
  178. #else
  179. /* Only GPUS */
  180. workerid = (attributed / 21) % 3;
  181. #endif
  182. /*= attributed/nblocks_per_worker;*/
  183. block->preferred_worker = workerid;
  184. attributed++;
  185. }
  186. }
  187. }
  188. void assign_blocks_to_mpi_nodes(int world_size)
  189. {
  190. unsigned nzblocks_per_process = (nbz + world_size - 1) / world_size;
  191. unsigned bz;
  192. for (bz = 0; bz < nbz; bz++)
  193. {
  194. struct block_description *block =
  195. get_block_description(bz);
  196. block->mpi_node = bz / nzblocks_per_process;
  197. }
  198. }
  199. static size_t allocated = 0;
  200. static void allocate_block_on_node(starpu_data_handle_t *handleptr, unsigned bz, TYPE **ptr, unsigned nx, unsigned ny, unsigned nz)
  201. {
  202. int ret;
  203. size_t block_size = nx*ny*nz*sizeof(TYPE);
  204. /* Allocate memory */
  205. #if 1
  206. ret = starpu_malloc_flags((void **)ptr, block_size, STARPU_MALLOC_PINNED|STARPU_MALLOC_SIMULATION_FOLDED);
  207. STARPU_ASSERT(ret == 0);
  208. #else
  209. *ptr = malloc(block_size);
  210. STARPU_ASSERT(*ptr);
  211. #endif
  212. allocated += block_size;
  213. //#ifndef STARPU_SIMGRID
  214. // /* Fill the blocks with 0 */
  215. // memset(*ptr, 0, block_size);
  216. //#endif
  217. /* Register it to StarPU */
  218. starpu_block_data_register(handleptr, STARPU_MAIN_RAM, (uintptr_t)*ptr, nx, nx*ny, nx, ny, nz, sizeof(TYPE));
  219. starpu_data_set_coordinates(*handleptr, 1, bz);
  220. }
  221. static void free_block_on_node(starpu_data_handle_t handleptr, unsigned nx, unsigned ny, unsigned nz)
  222. {
  223. void *ptr = (void *) starpu_block_get_local_ptr(handleptr);
  224. size_t block_size = nx*ny*nz*sizeof(TYPE);
  225. starpu_data_unregister(handleptr);
  226. starpu_free_flags(ptr, block_size, STARPU_MALLOC_PINNED|STARPU_MALLOC_SIMULATION_FOLDED);
  227. }
  228. void display_memory_consumption(int rank, double time)
  229. {
  230. FPRINTF(stderr, "%lu B of memory were allocated on node %d in %f ms\n", (unsigned long)allocated, rank, time/1000);
  231. }
  232. void allocate_memory_on_node(int rank)
  233. {
  234. unsigned bz;
  235. /* Correctly allocate and declare all data handles to StarPU. */
  236. for (bz = 0; bz < nbz; bz++)
  237. {
  238. struct block_description *block = get_block_description(bz);
  239. int node = block->mpi_node;
  240. unsigned size_bz = block_sizes_z[bz];
  241. if (node == rank)
  242. {
  243. /* Main blocks */
  244. allocate_block_on_node(&block->layers_handle[0], bz, &block->layers[0],
  245. (sizex + 2*K), (sizey + 2*K), (size_bz + 2*K));
  246. allocate_block_on_node(&block->layers_handle[1], bz, &block->layers[1],
  247. (sizex + 2*K), (sizey + 2*K), (size_bz + 2*K));
  248. /* Boundary blocks : Top */
  249. allocate_block_on_node(&block->boundaries_handle[T][0], bz, &block->boundaries[T][0],
  250. (sizex + 2*K), (sizey + 2*K), K);
  251. allocate_block_on_node(&block->boundaries_handle[T][1], bz, &block->boundaries[T][1],
  252. (sizex + 2*K), (sizey + 2*K), K);
  253. /* Boundary blocks : Bottom */
  254. allocate_block_on_node(&block->boundaries_handle[B][0], bz, &block->boundaries[B][0],
  255. (sizex + 2*K), (sizey + 2*K), K);
  256. allocate_block_on_node(&block->boundaries_handle[B][1], bz, &block->boundaries[B][1],
  257. (sizex + 2*K), (sizey + 2*K), K);
  258. }
  259. /* Register void blocks to StarPU, that StarPU-MPI will request to
  260. * neighbour nodes if needed for the local computation */
  261. else
  262. {
  263. /* Main blocks */
  264. starpu_block_data_register(&block->layers_handle[0], -1, (uintptr_t) NULL, (sizex + 2*K), (sizex + 2*K)*(sizey + 2*K), (sizex + 2*K), (sizey + 2*K), (size_bz + 2*K), sizeof(TYPE));
  265. starpu_block_data_register(&block->layers_handle[1], -1, (uintptr_t) NULL, (sizex + 2*K), (sizex + 2*K)*(sizey + 2*K), (sizex + 2*K), (sizey + 2*K), (size_bz + 2*K), sizeof(TYPE));
  266. /* Boundary blocks : Top */
  267. starpu_block_data_register(&block->boundaries_handle[T][0], -1, (uintptr_t) NULL, (sizex + 2*K), (sizex + 2*K)*(sizey + 2*K), (sizex + 2*K), (sizey + 2*K), K, sizeof(TYPE));
  268. starpu_block_data_register(&block->boundaries_handle[T][1], -1, (uintptr_t) NULL, (sizex + 2*K), (sizex + 2*K)*(sizey + 2*K), (sizex + 2*K), (sizey + 2*K), K, sizeof(TYPE));
  269. /* Boundary blocks : Bottom */
  270. starpu_block_data_register(&block->boundaries_handle[B][0], -1, (uintptr_t) NULL, (sizex + 2*K), (sizex + 2*K)*(sizey + 2*K), (sizex + 2*K), (sizey + 2*K), K, sizeof(TYPE));
  271. starpu_block_data_register(&block->boundaries_handle[B][1], -1, (uintptr_t) NULL, (sizex + 2*K), (sizex + 2*K)*(sizey + 2*K), (sizex + 2*K), (sizey + 2*K), K, sizeof(TYPE));
  272. }
  273. #if defined(STARPU_USE_MPI) && !defined(STARPU_USE_MPI_MASTER_SLAVE)
  274. /* Register all data to StarPU-MPI, even the ones that are not
  275. * allocated on the local node. */
  276. /* Main blocks */
  277. starpu_mpi_data_register(block->layers_handle[0], MPI_TAG_LAYERS(bz, 0), node);
  278. starpu_mpi_data_register(block->layers_handle[1], MPI_TAG_LAYERS(bz, 1), node);
  279. /* Boundary blocks : Top */
  280. starpu_mpi_data_register(block->boundaries_handle[T][0], MPI_TAG_BOUNDARIES(bz, T, 0), node);
  281. starpu_mpi_data_register(block->boundaries_handle[T][1], MPI_TAG_BOUNDARIES(bz, T, 1), node);
  282. /* Boundary blocks : Bottom */
  283. starpu_mpi_data_register(block->boundaries_handle[B][0], MPI_TAG_BOUNDARIES(bz, B, 0), node);
  284. starpu_mpi_data_register(block->boundaries_handle[B][1], MPI_TAG_BOUNDARIES(bz, B, 1), node);
  285. #endif
  286. }
  287. /* Initialize all the data in parallel */
  288. for (bz = 0; bz < nbz; bz++)
  289. {
  290. struct block_description *block = get_block_description(bz);
  291. int node = block->mpi_node;
  292. if (node == rank)
  293. {
  294. /* Set all the data to 0 */
  295. create_task_memset(sizex, sizey, bz);
  296. /* Initialize the first layer with some random data */
  297. create_task_initlayer(sizex, sizey, bz);
  298. }
  299. }
  300. starpu_task_wait_for_all();
  301. }
  302. void free_memory_on_node(int rank)
  303. {
  304. unsigned bz;
  305. for (bz = 0; bz < nbz; bz++)
  306. {
  307. struct block_description *block = get_block_description(bz);
  308. int node = block->mpi_node;
  309. /* Main blocks */
  310. if (node == rank)
  311. {
  312. free_block_on_node(block->layers_handle[0], (sizex + 2*K), (sizey + 2*K), K);
  313. free_block_on_node(block->layers_handle[1], (sizex + 2*K), (sizey + 2*K), K);
  314. }
  315. else
  316. {
  317. starpu_data_unregister(block->layers_handle[0]);
  318. starpu_data_unregister(block->layers_handle[1]);
  319. }
  320. /* Boundary blocks : Top */
  321. if (node == rank)
  322. {
  323. free_block_on_node(block->boundaries_handle[T][0], (sizex + 2*K), (sizey + 2*K), K);
  324. free_block_on_node(block->boundaries_handle[T][1], (sizex + 2*K), (sizey + 2*K), K);
  325. }
  326. else
  327. {
  328. starpu_data_unregister(block->boundaries_handle[T][0]);
  329. starpu_data_unregister(block->boundaries_handle[T][1]);
  330. }
  331. /* Boundary blocks : Bottom */
  332. if (node == rank)
  333. {
  334. free_block_on_node(block->boundaries_handle[B][0], (sizex + 2*K), (sizey + 2*K), K);
  335. free_block_on_node(block->boundaries_handle[B][1], (sizex + 2*K), (sizey + 2*K), K);
  336. }
  337. else
  338. {
  339. starpu_data_unregister(block->boundaries_handle[B][0]);
  340. starpu_data_unregister(block->boundaries_handle[B][1]);
  341. }
  342. }
  343. }
  344. /* check how many cells are alive */
  345. void check(int rank)
  346. {
  347. unsigned bz;
  348. for (bz = 0; bz < nbz; bz++)
  349. {
  350. struct block_description *block = get_block_description(bz);
  351. int node = block->mpi_node;
  352. /* Main blocks */
  353. if (node == rank)
  354. {
  355. unsigned size_bz = block_sizes_z[bz];
  356. #ifdef LIFE
  357. unsigned x, y, z;
  358. unsigned sum = 0;
  359. for (x = 0; x < sizex; x++)
  360. for (y = 0; y < sizey; y++)
  361. for (z = 0; z < size_bz; z++)
  362. sum += block->layers[0][(K+x)+(K+y)*(sizex + 2*K)+(K+z)*(sizex+2*K)*(sizey+2*K)];
  363. printf("block %u got %u/%u alive\n", bz, sum, sizex*sizey*size_bz);
  364. #endif
  365. }
  366. }
  367. }