mpi_reduction.c 3.9 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157
  1. /* StarPU --- Runtime system for heterogeneous multicore architectures.
  2. *
  3. * Copyright (C) 2012 Centre National de la Recherche Scientifique
  4. *
  5. * StarPU is free software; you can redistribute it and/or modify
  6. * it under the terms of the GNU Lesser General Public License as published by
  7. * the Free Software Foundation; either version 2.1 of the License, or (at
  8. * your option) any later version.
  9. *
  10. * StarPU is distributed in the hope that it will be useful, but
  11. * WITHOUT ANY WARRANTY; without even the implied warranty of
  12. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
  13. *
  14. * See the GNU Lesser General Public License in COPYING.LGPL for more details.
  15. */
  16. #include <starpu_mpi.h>
  17. #include <math.h>
  18. extern void init_cpu_func(void *descr[], void *cl_arg);
  19. extern void redux_cpu_func(void *descr[], void *cl_arg);
  20. extern void dot_cpu_func(void *descr[], void *cl_arg);
  21. static struct starpu_codelet init_codelet =
  22. {
  23. .where = STARPU_CPU,
  24. .cpu_funcs = {init_cpu_func, NULL},
  25. .nbuffers = 1,
  26. .name = "init_codelet"
  27. };
  28. static struct starpu_codelet redux_codelet =
  29. {
  30. .where = STARPU_CPU,
  31. .cpu_funcs = {redux_cpu_func, NULL},
  32. .nbuffers = 2,
  33. .name = "redux_codelet"
  34. };
  35. static struct starpu_codelet dot_codelet =
  36. {
  37. .where = STARPU_CPU,
  38. .cpu_funcs = {dot_cpu_func, NULL},
  39. .nbuffers = 2,
  40. .modes = {STARPU_R, STARPU_REDUX},
  41. .name = "dot_codelet"
  42. };
  43. /* Returns the MPI node number where data indexes index is */
  44. int my_distrib(int x, int nb_nodes)
  45. {
  46. return x % nb_nodes;
  47. }
  48. int main(int argc, char **argv)
  49. {
  50. int my_rank, size, x, y;
  51. long int *vector;
  52. long int dot, sum=0;
  53. starpu_data_handle_t *handles;
  54. starpu_data_handle_t dot_handle;
  55. int nb_elements, step;
  56. int ret = starpu_init(NULL);
  57. STARPU_CHECK_RETURN_VALUE(ret, "starpu_init");
  58. starpu_mpi_initialize_extended(&my_rank, &size);
  59. nb_elements = size*8000;
  60. step = 4;
  61. vector = (long int *) malloc(nb_elements*sizeof(vector[0]));
  62. for(x = 0; x < nb_elements; x+=step)
  63. {
  64. int mpi_rank = my_distrib(x/step, size);
  65. if (mpi_rank == my_rank)
  66. {
  67. for(y=0 ; y<step ; y++)
  68. {
  69. vector[x+y] = x+y+1;
  70. }
  71. }
  72. }
  73. if (my_rank == 0) {
  74. dot = 14;
  75. sum = (nb_elements * (nb_elements + 1)) / 2;
  76. sum+= dot;
  77. starpu_variable_data_register(&dot_handle, 0, (uintptr_t)&dot, sizeof(dot));
  78. }
  79. else
  80. {
  81. starpu_variable_data_register(&dot_handle, -1, (uintptr_t)NULL, sizeof(dot));
  82. }
  83. handles = (starpu_data_handle_t *) malloc(nb_elements*sizeof(handles[0]));
  84. for(x = 0; x < nb_elements; x+=step)
  85. {
  86. int mpi_rank = my_distrib(x/step, size);
  87. if (mpi_rank == my_rank)
  88. {
  89. /* Owning data */
  90. starpu_vector_data_register(&handles[x], 0, (uintptr_t)&(vector[x]), step, sizeof(vector[0]));
  91. }
  92. else
  93. {
  94. starpu_vector_data_register(&handles[x], -1, (uintptr_t)NULL, step, sizeof(vector[0]));
  95. }
  96. if (handles[x])
  97. {
  98. starpu_data_set_rank(handles[x], mpi_rank);
  99. starpu_data_set_tag(handles[x], x);
  100. }
  101. }
  102. starpu_data_set_rank(dot_handle, 0);
  103. starpu_data_set_tag(dot_handle, nb_elements+1);
  104. starpu_data_set_reduction_methods(dot_handle, &redux_codelet, &init_codelet);
  105. for (x = 0; x < nb_elements; x+=step)
  106. {
  107. starpu_mpi_insert_task(MPI_COMM_WORLD,
  108. &dot_codelet,
  109. STARPU_R, handles[x],
  110. STARPU_REDUX, dot_handle,
  111. 0);
  112. }
  113. starpu_mpi_redux_data(MPI_COMM_WORLD, dot_handle);
  114. fprintf(stderr, "Waiting ...\n");
  115. starpu_task_wait_for_all();
  116. for(x = 0; x < nb_elements; x+=step)
  117. {
  118. if (handles[x]) starpu_data_unregister(handles[x]);
  119. }
  120. if (dot_handle)
  121. {
  122. starpu_data_unregister(dot_handle);
  123. }
  124. free(vector);
  125. free(handles);
  126. starpu_mpi_shutdown();
  127. starpu_shutdown();
  128. if (my_rank == 0)
  129. {
  130. fprintf(stderr, "[%d] sum=%ld\n", my_rank, sum);
  131. fprintf(stderr, "[%d] dot=%ld\n", my_rank, dot);
  132. fprintf(stderr, "%s when computing reduction\n", (sum == dot) ? "Success" : "Error");
  133. }
  134. return 0;
  135. }