memory_nodes.c 6.3 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239
  1. /* StarPU --- Runtime system for heterogeneous multicore architectures.
  2. *
  3. * Copyright (C) 2009-2015 Université de Bordeaux
  4. * Copyright (C) 2010, 2011, 2012, 2013, 2014, 2015 CNRS
  5. *
  6. * StarPU is free software; you can redistribute it and/or modify
  7. * it under the terms of the GNU Lesser General Public License as published by
  8. * the Free Software Foundation; either version 2.1 of the License, or (at
  9. * your option) any later version.
  10. *
  11. * StarPU is distributed in the hope that it will be useful, but
  12. * WITHOUT ANY WARRANTY; without even the implied warranty of
  13. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
  14. *
  15. * See the GNU Lesser General Public License in COPYING.LGPL for more details.
  16. */
  17. #include <common/config.h>
  18. #include <core/sched_policy.h>
  19. #include <datawizard/datastats.h>
  20. #include <datawizard/memory_manager.h>
  21. #include <datawizard/memory_nodes.h>
  22. #include <common/fxt.h>
  23. #include "copy_driver.h"
  24. #include "memalloc.h"
  25. static struct _starpu_memory_node_descr descr;
  26. static starpu_pthread_key_t memory_node_key;
  27. void _starpu_memory_nodes_init(void)
  28. {
  29. /* there is no node yet, subsequent nodes will be
  30. * added using _starpu_memory_node_register */
  31. descr.nnodes = 0;
  32. STARPU_PTHREAD_KEY_CREATE(&memory_node_key, NULL);
  33. unsigned i;
  34. for (i = 0; i < STARPU_MAXNODES; i++)
  35. {
  36. descr.nodes[i] = STARPU_UNUSED;
  37. descr.nworkers[i] = 0;
  38. }
  39. _starpu_init_mem_chunk_lists();
  40. _starpu_init_data_request_lists();
  41. _starpu_memory_manager_init();
  42. STARPU_PTHREAD_RWLOCK_INIT(&descr.conditions_rwlock, NULL);
  43. descr.total_condition_count = 0;
  44. }
  45. void _starpu_memory_nodes_deinit(void)
  46. {
  47. _starpu_deinit_data_request_lists();
  48. _starpu_deinit_mem_chunk_lists();
  49. STARPU_PTHREAD_RWLOCK_DESTROY(&descr.conditions_rwlock);
  50. STARPU_PTHREAD_KEY_DELETE(memory_node_key);
  51. }
  52. void _starpu_memory_node_set_local_key(unsigned *node)
  53. {
  54. STARPU_PTHREAD_SETSPECIFIC(memory_node_key, node);
  55. }
  56. unsigned _starpu_memory_node_get_local_key(void)
  57. {
  58. unsigned *memory_node;
  59. memory_node = (unsigned *) STARPU_PTHREAD_GETSPECIFIC(memory_node_key);
  60. /* in case this is called by the programmer, we assume the RAM node
  61. is the appropriate memory node ... XXX */
  62. if (STARPU_UNLIKELY(!memory_node))
  63. return STARPU_MAIN_RAM;
  64. return *memory_node;
  65. }
  66. void _starpu_memory_node_add_nworkers(unsigned node)
  67. {
  68. descr.nworkers[node]++;
  69. }
  70. unsigned _starpu_memory_node_get_nworkers(unsigned node)
  71. {
  72. return descr.nworkers[node];
  73. }
  74. struct _starpu_memory_node_descr *_starpu_memory_node_get_description(void)
  75. {
  76. return &descr;
  77. }
  78. enum starpu_node_kind starpu_node_get_kind(unsigned node)
  79. {
  80. return descr.nodes[node];
  81. }
  82. int _starpu_memory_node_get_devid(unsigned node)
  83. {
  84. return descr.devid[node];
  85. }
  86. unsigned starpu_memory_nodes_get_count(void)
  87. {
  88. return descr.nnodes;
  89. }
  90. void _starpu_memory_node_get_name(unsigned node, char *name, int size)
  91. {
  92. const char *prefix;
  93. switch (descr.nodes[node])
  94. {
  95. case STARPU_CPU_RAM:
  96. prefix = "RAM";
  97. break;
  98. case STARPU_CUDA_RAM:
  99. prefix = "CUDA";
  100. break;
  101. case STARPU_OPENCL_RAM:
  102. prefix = "OpenCL";
  103. break;
  104. case STARPU_DISK_RAM:
  105. prefix = "Disk";
  106. break;
  107. case STARPU_MIC_RAM:
  108. prefix = "MIC";
  109. break;
  110. case STARPU_SCC_RAM:
  111. prefix = "SCC_RAM";
  112. break;
  113. case STARPU_SCC_SHM:
  114. prefix = "SCC_shared";
  115. break;
  116. case STARPU_UNUSED:
  117. default:
  118. prefix = "unknown";
  119. STARPU_ASSERT(0);
  120. }
  121. snprintf(name, size, "%s %u", prefix, descr.devid[node]);
  122. }
  123. unsigned _starpu_memory_node_register(enum starpu_node_kind kind, int devid)
  124. {
  125. unsigned node;
  126. /* ATOMIC_ADD returns the new value ... */
  127. node = STARPU_ATOMIC_ADD(&descr.nnodes, 1) - 1;
  128. STARPU_ASSERT_MSG(node < STARPU_MAXNODES,"Too many nodes (%u) for maximum %u. Use configure option --enable-maxnodes=xxx to update the maximum number of nodes.", node, STARPU_MAXNODES);
  129. descr.nodes[node] = kind;
  130. _STARPU_TRACE_NEW_MEM_NODE(node);
  131. descr.devid[node] = devid;
  132. /* for now, there is no condition associated to that newly created node */
  133. descr.condition_count[node] = 0;
  134. return node;
  135. }
  136. #ifdef STARPU_SIMGRID
  137. void _starpu_simgrid_memory_node_set_host(unsigned node, msg_host_t host)
  138. {
  139. descr.host[node] = host;
  140. }
  141. msg_host_t _starpu_simgrid_memory_node_get_host(unsigned node)
  142. {
  143. return descr.host[node];
  144. }
  145. #endif
  146. /* TODO move in a more appropriate file !! */
  147. /* Register a condition variable associated to worker which is associated to a
  148. * memory node itself. */
  149. void _starpu_memory_node_register_condition(starpu_pthread_cond_t *cond, starpu_pthread_mutex_t *mutex, unsigned nodeid)
  150. {
  151. unsigned cond_id;
  152. unsigned nconds_total, nconds;
  153. STARPU_PTHREAD_RWLOCK_WRLOCK(&descr.conditions_rwlock);
  154. /* we only insert the queue if it's not already in the list */
  155. nconds = descr.condition_count[nodeid];
  156. for (cond_id = 0; cond_id < nconds; cond_id++)
  157. {
  158. if (descr.conditions_attached_to_node[nodeid][cond_id].cond == cond)
  159. {
  160. STARPU_ASSERT(descr.conditions_attached_to_node[nodeid][cond_id].mutex == mutex);
  161. /* the condition is already in the list */
  162. STARPU_PTHREAD_RWLOCK_UNLOCK(&descr.conditions_rwlock);
  163. return;
  164. }
  165. }
  166. /* it was not found locally */
  167. descr.conditions_attached_to_node[nodeid][cond_id].cond = cond;
  168. descr.conditions_attached_to_node[nodeid][cond_id].mutex = mutex;
  169. descr.condition_count[nodeid]++;
  170. /* do we have to add it in the global list as well ? */
  171. nconds_total = descr.total_condition_count;
  172. for (cond_id = 0; cond_id < nconds_total; cond_id++)
  173. {
  174. if (descr.conditions_all[cond_id].cond == cond)
  175. {
  176. /* the queue is already in the global list */
  177. STARPU_PTHREAD_RWLOCK_UNLOCK(&descr.conditions_rwlock);
  178. return;
  179. }
  180. }
  181. /* it was not in the global list either */
  182. descr.conditions_all[nconds_total].cond = cond;
  183. descr.conditions_all[nconds_total].mutex = mutex;
  184. descr.total_condition_count++;
  185. STARPU_PTHREAD_RWLOCK_UNLOCK(&descr.conditions_rwlock);
  186. }
  187. unsigned starpu_worker_get_memory_node(unsigned workerid)
  188. {
  189. struct _starpu_machine_config *config = _starpu_get_machine_config();
  190. /* This workerid may either be a basic worker or a combined worker */
  191. unsigned nworkers = config->topology.nworkers;
  192. if (workerid < config->topology.nworkers)
  193. return config->workers[workerid].memory_node;
  194. /* We have a combined worker */
  195. unsigned ncombinedworkers = config->topology.ncombinedworkers;
  196. STARPU_ASSERT_MSG(workerid < ncombinedworkers + nworkers, "Bad workerid %u, maximum %u", workerid, ncombinedworkers + nworkers);
  197. return config->combined_workers[workerid - nworkers].memory_node;
  198. }