Samuel Thibault 4 år sedan
förälder
incheckning
07f8c7c6c2
45 ändrade filer med 50 tillägg och 81 borttagningar
  1. 1 1
      doc/doxygen/chapters/code/disk_compute.c
  2. 0 1
      examples/basic_examples/task_insert_color.c
  3. 1 1
      examples/cholesky/cholesky_kernels.c
  4. 1 1
      examples/filters/fmultiple_submit_readonly.c
  5. 1 1
      examples/filters/fplan_notautomatic.c
  6. 0 2
      examples/heat/dw_factolu_kernels.c
  7. 0 2
      examples/lu/xlu_kernels.c
  8. 1 1
      examples/pi/SobolQRNG/sobol_gold.c
  9. 2 0
      examples/sched_ctx/gpu_partition.c
  10. 1 2
      examples/sched_ctx/sched_ctx.c
  11. 3 2
      examples/sched_ctx/sched_ctx_without_sched_policy.c
  12. 1 2
      examples/spmv/matrix_market/mm_to_bcsr.c
  13. 0 1
      julia/examples/task_insert_color/task_insert_color.c
  14. 3 4
      mpi/examples/benchs/abstract_sendrecv_bench.c
  15. 4 7
      mpi/examples/benchs/sendrecv_parallel_tasks_bench.c
  16. 2 2
      mpi/examples/cg/cg.c
  17. 1 1
      mpi/examples/matrix_decomposition/mpi_cholesky_codelets.c
  18. 0 3
      mpi/examples/matrix_decomposition/mpi_cholesky_distributed.c
  19. 1 1
      mpi/examples/matrix_decomposition/mpi_cholesky_kernels.c
  20. 0 3
      mpi/examples/mpi_lu/pxlu.c
  21. 0 2
      mpi/examples/mpi_lu/pxlu_kernels.c
  22. 0 1
      mpi/examples/mpi_redux/mpi_redux.c
  23. 0 1
      mpi/src/mpi/starpu_mpi_early_data.c
  24. 1 1
      mpi/tests/early_stuff.c
  25. 0 1
      mpi/tests/mpi_barrier.c
  26. 1 1
      mpi/tests/pingpong.c
  27. 0 1
      mpi/tests/temporary.c
  28. 0 1
      src/core/perfmodel/energy_model.c
  29. 6 6
      src/datawizard/copy_driver.c
  30. 1 1
      src/drivers/mp_common/sink_common.c
  31. 1 0
      src/drivers/mpi/driver_mpi_common.c
  32. 5 1
      src/sched_policies/component_eager.c
  33. 0 1
      src/sched_policies/work_stealing_policy.c
  34. 1 1
      starpurm/dev/chameleon_test/dgemm.c
  35. 0 2
      tests/datawizard/acquire_try.c
  36. 1 1
      tests/datawizard/commute2.c
  37. 0 1
      tests/datawizard/interfaces/variable/variable_opencl.c
  38. 2 2
      tests/datawizard/specific_node.c
  39. 0 1
      tests/energy/energy_efficiency.c
  40. 1 2
      tests/main/callback.c
  41. 1 1
      tests/microbenchs/tasks_size_overhead.c
  42. 2 2
      tests/perfmodels/memory.c
  43. 1 4
      tests/perfmodels/regression_based_energy.c
  44. 2 4
      tests/perfmodels/regression_based_gpu.c
  45. 1 4
      tests/perfmodels/regression_based_multiimpl.c

+ 1 - 1
doc/doxygen/chapters/code/disk_compute.c

@@ -128,7 +128,7 @@ int main(int argc, char **argv)
 	if (f == NULL)
 	if (f == NULL)
 		goto enoent;
 		goto enoent;
 	/* take datas */
 	/* take datas */
-	int size = fread(C, sizeof(int), NX, f);
+	fread(C, sizeof(int), NX, f);
 
 
 	/* close the file */
 	/* close the file */
 	fclose(f);
 	fclose(f);

+ 0 - 1
examples/basic_examples/task_insert_color.c

@@ -45,7 +45,6 @@ struct starpu_codelet mycodelet_color =
 
 
 int main(void)
 int main(void)
 {
 {
-	unsigned i;
 	int value=42;
 	int value=42;
 	starpu_data_handle_t handle;
 	starpu_data_handle_t handle;
 	int ret;
 	int ret;

+ 1 - 1
examples/cholesky/cholesky_kernels.c

@@ -218,8 +218,8 @@ static inline void chol_common_codelet_update_u11(void *descr[], int s, void *_a
 			{
 			{
 			int ret;
 			int ret;
 			int info;
 			int info;
-			cudaStream_t stream = starpu_cuda_get_local_stream();
 #if (MAGMA_VERSION_MAJOR > 1) || (MAGMA_VERSION_MAJOR == 1 && MAGMA_VERSION_MINOR >= 4)
 #if (MAGMA_VERSION_MAJOR > 1) || (MAGMA_VERSION_MAJOR == 1 && MAGMA_VERSION_MINOR >= 4)
+			cudaStream_t stream = starpu_cuda_get_local_stream();
 			cublasSetKernelStream(stream);
 			cublasSetKernelStream(stream);
 			magmablasSetKernelStream(stream);
 			magmablasSetKernelStream(stream);
 #else
 #else

+ 1 - 1
examples/filters/fmultiple_submit_readonly.c

@@ -146,7 +146,7 @@ struct starpu_codelet cl_check =
 int main(void)
 int main(void)
 {
 {
 	int start, factor;
 	int start, factor;
-	unsigned j, n=1;
+	unsigned n=1;
 	int matrix[NX][NY];
 	int matrix[NX][NY];
 	int ret, i;
 	int ret, i;
 
 

+ 1 - 1
examples/filters/fplan_notautomatic.c

@@ -58,7 +58,7 @@ void split_callback(void *arg)
 
 
 void supertask_callback(void *arg)
 void supertask_callback(void *arg)
 {
 {
-	(arg);
+	(void)arg;
 	starpu_data_handle_t sub_handles[PARTS];
 	starpu_data_handle_t sub_handles[PARTS];
 	int add;
 	int add;
 	struct starpu_task *task = starpu_task_get_current();
 	struct starpu_task *task = starpu_task_get_current();

+ 0 - 2
examples/heat/dw_factolu_kernels.c

@@ -338,7 +338,6 @@ static inline void dw_common_codelet_update_u11(void *descr[], int s, void *_arg
 
 
 #ifdef STARPU_USE_CUDA
 #ifdef STARPU_USE_CUDA
 	cudaStream_t stream;
 	cudaStream_t stream;
-	cublasHandle_t handle;
 	cublasStatus_t status;
 	cublasStatus_t status;
 #endif
 #endif
 
 
@@ -372,7 +371,6 @@ static inline void dw_common_codelet_update_u11(void *descr[], int s, void *_arg
 #ifdef STARPU_USE_CUDA
 #ifdef STARPU_USE_CUDA
 		case 1:
 		case 1:
 			stream = starpu_cuda_get_local_stream();
 			stream = starpu_cuda_get_local_stream();
-			handle = starpu_cublas_get_local_handle();
 			for (z = 0; z < nx; z++)
 			for (z = 0; z < nx; z++)
 			{
 			{
 				float pivot;
 				float pivot;

+ 0 - 2
examples/lu/xlu_kernels.c

@@ -55,7 +55,6 @@ static inline void STARPU_LU(common_u22)(void *descr[], int s, void *_args)
 
 
 #ifdef STARPU_USE_CUDA
 #ifdef STARPU_USE_CUDA
 	cublasStatus status;
 	cublasStatus status;
-	cudaError_t cures;
 #endif
 #endif
 
 
 	switch (s)
 	switch (s)
@@ -181,7 +180,6 @@ static inline void STARPU_LU(common_u12)(void *descr[], int s, void *_args)
 
 
 #ifdef STARPU_USE_CUDA
 #ifdef STARPU_USE_CUDA
 	cublasStatus status;
 	cublasStatus status;
-	cudaError_t cures;
 #endif
 #endif
 
 
 	/* solve L11 U12 = A12 (find U12) */
 	/* solve L11 U12 = A12 (find U12) */

+ 1 - 1
examples/pi/SobolQRNG/sobol_gold.c

@@ -70,7 +70,7 @@ void initSobolDirectionVectors(int n_dimensions, unsigned int *directions)
             for (i = 0 ; i < n_directions ; i++)
             for (i = 0 ; i < n_directions ; i++)
             {
             {
                 /* All m's are 1 */
                 /* All m's are 1 */
-                v[i] = 1 << (31 - i);
+                v[i] = 1U << (31 - i);
             }
             }
         }
         }
         else
         else

+ 2 - 0
examples/sched_ctx/gpu_partition.c

@@ -103,8 +103,10 @@ int main(void)
 {
 {
 	int ret, exit_value = 0;
 	int ret, exit_value = 0;
 	int iter;
 	int iter;
+#ifdef STARPU_USE_CUDA
 	int ncuda = 0;
 	int ncuda = 0;
 	int gpu_devid = -1;
 	int gpu_devid = -1;
+#endif
 
 
 #ifdef STARPU_DEVEL
 #ifdef STARPU_DEVEL
 #warning temporary fix: skip test as cuda computation fails
 #warning temporary fix: skip test as cuda computation fails

+ 1 - 2
examples/sched_ctx/sched_ctx.c

@@ -82,7 +82,6 @@ int main(void)
 {
 {
 	int ntasks = NTASKS;
 	int ntasks = NTASKS;
 	int ret;
 	int ret;
-	unsigned ncuda = 0;
 	int nprocs1 = 0;
 	int nprocs1 = 0;
 	int nprocs2 = 0;
 	int nprocs2 = 0;
 	int procs1[STARPU_NMAXWORKERS], procs2[STARPU_NMAXWORKERS];
 	int procs1[STARPU_NMAXWORKERS], procs2[STARPU_NMAXWORKERS];
@@ -105,7 +104,7 @@ int main(void)
 	if (nprocs1 == 0) goto enodev;
 	if (nprocs1 == 0) goto enodev;
 
 
 #ifdef STARPU_USE_CUDA
 #ifdef STARPU_USE_CUDA
-	ncuda = nprocs2 = starpu_cuda_worker_get_count();
+	nprocs2 = starpu_cuda_worker_get_count();
 	starpu_worker_get_ids_by_type(STARPU_CUDA_WORKER, procs2, nprocs2);
 	starpu_worker_get_ids_by_type(STARPU_CUDA_WORKER, procs2, nprocs2);
 #endif
 #endif
 	if (nprocs2 == 0)
 	if (nprocs2 == 0)

+ 3 - 2
examples/sched_ctx/sched_ctx_without_sched_policy.c

@@ -79,10 +79,11 @@ int main(void)
 
 
 	int nprocs1 = 1;
 	int nprocs1 = 1;
 	int nprocs2 = 1;
 	int nprocs2 = 1;
-	int ncuda = 0;
-	int *procs1, *procs2, *procscuda;
+	int *procs1, *procs2;
 
 
 #ifdef STARPU_USE_CUDA
 #ifdef STARPU_USE_CUDA
+	int ncuda = 0;
+	int *procscuda;
 	ncuda = starpu_cuda_worker_get_count();
 	ncuda = starpu_cuda_worker_get_count();
 	procscuda = (int*)malloc(ncuda*sizeof(int));
 	procscuda = (int*)malloc(ncuda*sizeof(int));
 	starpu_worker_get_ids_by_type(STARPU_CUDA_WORKER, procscuda, ncuda);
 	starpu_worker_get_ids_by_type(STARPU_CUDA_WORKER, procscuda, ncuda);

+ 1 - 2
examples/spmv/matrix_market/mm_to_bcsr.c

@@ -315,7 +315,6 @@ bcsr_t *mm_file_to_bcsr(char *filename, unsigned c, unsigned r)
 {
 {
 	FILE *f;
 	FILE *f;
 	MM_typecode matcode;
 	MM_typecode matcode;
-	int ret_code;
 	int M, N;
 	int M, N;
 	int nz;
 	int nz;
 	int i;
 	int i;
@@ -348,7 +347,7 @@ bcsr_t *mm_file_to_bcsr(char *filename, unsigned c, unsigned r)
 
 
 	/* find out size of sparse matrix .... */
 	/* find out size of sparse matrix .... */
 
 
-	if ((ret_code = mm_read_mtx_crd_size(f, &M, &N, &nz)) !=0)
+	if ((mm_read_mtx_crd_size(f, &M, &N, &nz)) !=0)
 		exit(1);
 		exit(1);
 
 
 
 

+ 0 - 1
julia/examples/task_insert_color/task_insert_color.c

@@ -45,7 +45,6 @@ struct starpu_codelet mycodelet_color =
 
 
 int main(void)
 int main(void)
 {
 {
-	unsigned i;
 	int value=42;
 	int value=42;
 	starpu_data_handle_t handle;
 	starpu_data_handle_t handle;
 	int ret;
 	int ret;

+ 3 - 4
mpi/examples/benchs/abstract_sendrecv_bench.c

@@ -20,9 +20,9 @@
 void sendrecv_bench(int mpi_rank, starpu_pthread_barrier_t* thread_barrier, int bidir)
 void sendrecv_bench(int mpi_rank, starpu_pthread_barrier_t* thread_barrier, int bidir)
 {
 {
 	uint64_t iterations = LOOPS_DEFAULT;
 	uint64_t iterations = LOOPS_DEFAULT;
-	uint64_t s = 0;
-	uint64_t j = 0;
-	uint64_t k = 0;
+	uint64_t s;
+	uint64_t j;
+	uint64_t k;
 
 
 	if (mpi_rank >= 2)
 	if (mpi_rank >= 2)
 	{
 	{
@@ -54,7 +54,6 @@ void sendrecv_bench(int mpi_rank, starpu_pthread_barrier_t* thread_barrier, int
 		printf("# size  (Bytes)\t|  latency \t| 10^6 B/s \t| MB/s   \t| d1    \t|median  \t| avg    \t| d9    \t| max\n");
 		printf("# size  (Bytes)\t|  latency \t| 10^6 B/s \t| MB/s   \t| d1    \t|median  \t| avg    \t| d9    \t| max\n");
 	}
 	}
 
 
-	int array_size = 0;
 	starpu_data_handle_t handle_send, handle_recv;
 	starpu_data_handle_t handle_send, handle_recv;
 	float* vector_send = NULL;
 	float* vector_send = NULL;
 	float* vector_recv = NULL;
 	float* vector_recv = NULL;

+ 4 - 7
mpi/examples/benchs/sendrecv_parallel_tasks_bench.c

@@ -60,8 +60,8 @@ void cpu_task(void* descr[], void* args)
 	double t1, t2;
 	double t1, t2;
 	int asked_worker;
 	int asked_worker;
 	int current_worker = starpu_worker_get_id();
 	int current_worker = starpu_worker_get_id();
-	uint64_t j = 0;
-	uint64_t k = 0;
+	uint64_t j;
+	uint64_t k;
 
 
 	starpu_codelet_unpack_args(args, &mpi_rank, &asked_worker, &s, &handle_send, &handle_recv);
 	starpu_codelet_unpack_args(args, &mpi_rank, &asked_worker, &s, &handle_send, &handle_recv);
 
 
@@ -171,10 +171,8 @@ int main(int argc, char **argv)
 
 
 
 
 	unsigned cpu_count = starpu_cpu_worker_get_count();
 	unsigned cpu_count = starpu_cpu_worker_get_count();
-	unsigned* mpi_tags = malloc(cpu_count * sizeof(unsigned));
-	unsigned tag = 0;
-	uint64_t s = 0;
-	unsigned i = 0;
+	uint64_t s;
+	unsigned i;
 
 
 	int* workers = malloc(cpu_count * sizeof(int));
 	int* workers = malloc(cpu_count * sizeof(int));
 	float** vectors_send = malloc(cpu_count * sizeof(float*));
 	float** vectors_send = malloc(cpu_count * sizeof(float*));
@@ -223,7 +221,6 @@ int main(int argc, char **argv)
 	free(vectors_recv);
 	free(vectors_recv);
 	free(handles_send);
 	free(handles_send);
 	free(handles_recv);
 	free(handles_recv);
-	free(mpi_tags);
 
 
 	starpu_mpi_shutdown();
 	starpu_mpi_shutdown();
 
 

+ 2 - 2
mpi/examples/cg/cg.c

@@ -143,7 +143,7 @@ static void generate_random_problem(void)
 
 
 static void free_data(void)
 static void free_data(void)
 {
 {
-	unsigned ii, jj, j, i;
+	unsigned j, i;
 	int mpi_rank;
 	int mpi_rank;
 
 
 	for (j = 0; j < nblocks; j++)
 	for (j = 0; j < nblocks; j++)
@@ -377,7 +377,7 @@ int main(int argc, char **argv)
 
 
 	if (n % nblocks != 0)
 	if (n % nblocks != 0)
 	{
 	{
-		FPRINTF_SERVER(stderr, "The number of blocks (%d) must divide the matrix size (%lld).\n", nblocks, n);
+		FPRINTF_SERVER(stderr, "The number of blocks (%u) must divide the matrix size (%lld).\n", nblocks, n);
 		starpu_mpi_shutdown();
 		starpu_mpi_shutdown();
 		return 1;
 		return 1;
 	}
 	}

+ 1 - 1
mpi/examples/matrix_decomposition/mpi_cholesky_codelets.c

@@ -469,7 +469,7 @@ void dw_cholesky(float ***matA, unsigned ld, int rank, int nodes, double *timing
 	double start;
 	double start;
 	double end;
 	double end;
 	starpu_data_handle_t **data_handles;
 	starpu_data_handle_t **data_handles;
-	unsigned k, m, n;
+	unsigned m, n;
 
 
 	/* create all the DAG nodes */
 	/* create all the DAG nodes */
 
 

+ 0 - 3
mpi/examples/matrix_decomposition/mpi_cholesky_distributed.c

@@ -31,9 +31,6 @@ int main(int argc, char **argv)
 	float ***bmat;
 	float ***bmat;
 	int rank, nodes, ret;
 	int rank, nodes, ret;
 	double timing, flops;
 	double timing, flops;
-#ifndef STARPU_SIMGRID
-	int correctness=1;
-#endif
 
 
 	ret = starpu_mpi_init_conf(&argc, &argv, 1, MPI_COMM_WORLD, NULL);
 	ret = starpu_mpi_init_conf(&argc, &argv, 1, MPI_COMM_WORLD, NULL);
 	STARPU_CHECK_RETURN_VALUE(ret, "starpu_mpi_init_conf");
 	STARPU_CHECK_RETURN_VALUE(ret, "starpu_mpi_init_conf");

+ 1 - 1
mpi/examples/matrix_decomposition/mpi_cholesky_kernels.c

@@ -198,8 +198,8 @@ static inline void chol_common_codelet_update_u11(void *descr[], int s, void *_a
 			{
 			{
 				int ret;
 				int ret;
 				int info;
 				int info;
-			cudaStream_t stream = starpu_cuda_get_local_stream();
 #if (MAGMA_VERSION_MAJOR > 1) || (MAGMA_VERSION_MAJOR == 1 && MAGMA_VERSION_MINOR >= 4)
 #if (MAGMA_VERSION_MAJOR > 1) || (MAGMA_VERSION_MAJOR == 1 && MAGMA_VERSION_MINOR >= 4)
+			cudaStream_t stream = starpu_cuda_get_local_stream();
 			cublasSetKernelStream(stream);
 			cublasSetKernelStream(stream);
 			magmablasSetKernelStream(stream);
 			magmablasSetKernelStream(stream);
 #else
 #else

+ 0 - 3
mpi/examples/mpi_lu/pxlu.c

@@ -199,10 +199,7 @@ static void create_task_11_recv(unsigned k)
 			if (rank == get_block_rank(i, k-1))
 			if (rank == get_block_rank(i, k-1))
 				tag_array[ndeps++] = TAG21(k-1, i);
 				tag_array[ndeps++] = TAG21(k-1, i);
 		}
 		}
-	}
 
 
-	if (k > 0)
-	{
 		unsigned j;
 		unsigned j;
 		for (j = (k-1)+1; j < nblocks; j++)
 		for (j = (k-1)+1; j < nblocks; j++)
 		{
 		{

+ 0 - 2
mpi/examples/mpi_lu/pxlu_kernels.c

@@ -55,7 +55,6 @@ static inline void STARPU_PLU(common_u22)(void *descr[], int s, void *_args)
 
 
 #ifdef STARPU_USE_CUDA
 #ifdef STARPU_USE_CUDA
 	cublasStatus status;
 	cublasStatus status;
-	cudaError_t cures;
 #endif
 #endif
 
 
 	switch (s)
 	switch (s)
@@ -171,7 +170,6 @@ static inline void STARPU_PLU(common_u12)(void *descr[], int s, void *_args)
 
 
 #ifdef STARPU_USE_CUDA
 #ifdef STARPU_USE_CUDA
 	cublasStatus status;
 	cublasStatus status;
-	cudaError_t cures;
 #endif
 #endif
 
 
 	/* solve L11 U12 = A12 (find U12) */
 	/* solve L11 U12 = A12 (find U12) */

+ 0 - 1
mpi/examples/mpi_redux/mpi_redux.c

@@ -123,7 +123,6 @@ int main(int argc, char *argv[])
 	double a, b[comm_size];
 	double a, b[comm_size];
 	starpu_data_handle_t a_h, b_h[comm_size];
 	starpu_data_handle_t a_h, b_h[comm_size];
 	double work_coef = 2;
 	double work_coef = 2;
-	enum starpu_data_access_mode codelet_mode;
 	enum starpu_data_access_mode task_mode;
 	enum starpu_data_access_mode task_mode;
 	int i,j,work_node;
 	int i,j,work_node;
     	starpu_mpi_tag_t tag = 0;
     	starpu_mpi_tag_t tag = 0;

+ 0 - 1
mpi/src/mpi/starpu_mpi_early_data.c

@@ -95,7 +95,6 @@ struct _starpu_mpi_early_data_handle *_starpu_mpi_early_data_create(struct _star
 void _starpu_mpi_early_data_delete(struct _starpu_mpi_early_data_handle *early_data_handle)
 void _starpu_mpi_early_data_delete(struct _starpu_mpi_early_data_handle *early_data_handle)
 {
 {
 	free(early_data_handle);
 	free(early_data_handle);
-	early_data_handle = NULL;
 }
 }
 
 
 struct _starpu_mpi_early_data_handle *_starpu_mpi_early_data_find(struct _starpu_mpi_node_tag *node_tag)
 struct _starpu_mpi_early_data_handle *_starpu_mpi_early_data_find(struct _starpu_mpi_node_tag *node_tag)

+ 1 - 1
mpi/tests/early_stuff.c

@@ -108,7 +108,7 @@ void early_request()
 
 
 int main(int argc, char **argv)
 int main(int argc, char **argv)
 {
 {
-	int ret, rank, size, i;
+	int ret;
 	starpu_data_handle_t tab_handle[4];
 	starpu_data_handle_t tab_handle[4];
 	int mpi_init;
 	int mpi_init;
 
 

+ 0 - 1
mpi/tests/mpi_barrier.c

@@ -21,7 +21,6 @@
 int main(int argc, char **argv)
 int main(int argc, char **argv)
 {
 {
 	int ret, mpi_init;
 	int ret, mpi_init;
-	MPI_Status status;
 
 
 	MPI_INIT_THREAD(&argc, &argv, MPI_THREAD_SERIALIZED, &mpi_init);
 	MPI_INIT_THREAD(&argc, &argv, MPI_THREAD_SERIALIZED, &mpi_init);
 
 

+ 1 - 1
mpi/tests/pingpong.c

@@ -135,7 +135,7 @@ int main(int argc, char **argv)
 	int loop;
 	int loop;
 	int other_rank = rank%2 == 0 ? rank+1 : rank-1;
 	int other_rank = rank%2 == 0 ? rank+1 : rank-1;
 	int sender;
 	int sender;
-	int r = 0;
+	int r;
 
 
 	if (method == 0) // ping pongs
 	if (method == 0) // ping pongs
 	{
 	{

+ 0 - 1
mpi/tests/temporary.c

@@ -43,7 +43,6 @@ int main(int argc, char **argv)
 {
 {
 	int rank, size, n;
 	int rank, size, n;
 	int ret;
 	int ret;
-	int a;
 	int val0 = 0, val1 = 0;
 	int val0 = 0, val1 = 0;
 	starpu_data_handle_t data0, data1, tmp0, tmp, tmp2;
 	starpu_data_handle_t data0, data1, tmp0, tmp, tmp2;
 
 

+ 0 - 1
src/core/perfmodel/energy_model.c

@@ -234,7 +234,6 @@ int starpu_energy_stop(struct starpu_perfmodel *model, struct starpu_task *task,
 	{
 	{
 		printf("Error: worker type %d is not supported! \n", archi);
 		printf("Error: worker type %d is not supported! \n", archi);
 		return -1;
 		return -1;
-		break;
 	}
 	}
 	}
 	}
 
 

+ 6 - 6
src/datawizard/copy_driver.c

@@ -373,8 +373,8 @@ int starpu_interface_copy3d(uintptr_t src, size_t src_offset, unsigned src_node,
 	STARPU_ASSERT_MSG(ld1_src >= blocksize, "block size %lu is bigger than ld %lu in source", (unsigned long) blocksize, (unsigned long) ld1_src);
 	STARPU_ASSERT_MSG(ld1_src >= blocksize, "block size %lu is bigger than ld %lu in source", (unsigned long) blocksize, (unsigned long) ld1_src);
 	STARPU_ASSERT_MSG(ld1_dst >= blocksize, "block size %lu is bigger than ld %lu in destination", (unsigned long) blocksize, (unsigned long) ld1_dst);
 	STARPU_ASSERT_MSG(ld1_dst >= blocksize, "block size %lu is bigger than ld %lu in destination", (unsigned long) blocksize, (unsigned long) ld1_dst);
 
 
-	STARPU_ASSERT_MSG(ld2_src >= numblocks_1 * ld1_src, "block group size %lu is bigger than group ld %lu in source", (unsigned long) numblocks_1 * ld1_src, (unsigned long) ld2_src);
-	STARPU_ASSERT_MSG(ld2_dst >= numblocks_1 * ld1_dst, "block group size %lu is bigger than group ld %lu in destination", (unsigned long) numblocks_1 * ld1_dst, (unsigned long) ld2_dst);
+	STARPU_ASSERT_MSG(ld2_src >= numblocks_1 * ld1_src, "block group size %lu is bigger than group ld %lu in source", (unsigned long) (numblocks_1 * ld1_src), (unsigned long) ld2_src);
+	STARPU_ASSERT_MSG(ld2_dst >= numblocks_1 * ld1_dst, "block group size %lu is bigger than group ld %lu in destination", (unsigned long) (numblocks_1 * ld1_dst), (unsigned long) ld2_dst);
 
 
 	if (ld2_src == blocksize * numblocks_1 &&
 	if (ld2_src == blocksize * numblocks_1 &&
 	    ld2_dst == blocksize * numblocks_1)
 	    ld2_dst == blocksize * numblocks_1)
@@ -420,11 +420,11 @@ int starpu_interface_copy4d(uintptr_t src, size_t src_offset, unsigned src_node,
 	STARPU_ASSERT_MSG(ld1_src >= blocksize, "block size %lu is bigger than ld %lu in source", (unsigned long) blocksize, (unsigned long) ld1_src);
 	STARPU_ASSERT_MSG(ld1_src >= blocksize, "block size %lu is bigger than ld %lu in source", (unsigned long) blocksize, (unsigned long) ld1_src);
 	STARPU_ASSERT_MSG(ld1_dst >= blocksize, "block size %lu is bigger than ld %lu in destination", (unsigned long) blocksize, (unsigned long) ld1_dst);
 	STARPU_ASSERT_MSG(ld1_dst >= blocksize, "block size %lu is bigger than ld %lu in destination", (unsigned long) blocksize, (unsigned long) ld1_dst);
 
 
-	STARPU_ASSERT_MSG(ld2_src >= numblocks_1 * ld1_src, "block group size %lu is bigger than group ld %lu in source", (unsigned long) numblocks_1 * ld1_src, (unsigned long) ld2_src);
-	STARPU_ASSERT_MSG(ld2_dst >= numblocks_1 * ld1_dst, "block group size %lu is bigger than group ld %lu in destination", (unsigned long) numblocks_1 * ld1_dst, (unsigned long) ld2_dst);
+	STARPU_ASSERT_MSG(ld2_src >= numblocks_1 * ld1_src, "block group size %lu is bigger than group ld %lu in source", (unsigned long) (numblocks_1 * ld1_src), (unsigned long) ld2_src);
+	STARPU_ASSERT_MSG(ld2_dst >= numblocks_1 * ld1_dst, "block group size %lu is bigger than group ld %lu in destination", (unsigned long) (numblocks_1 * ld1_dst), (unsigned long) ld2_dst);
 
 
-	STARPU_ASSERT_MSG(ld3_src >= numblocks_2 * ld2_src, "block group group size %lu is bigger than group group ld %lu in source", (unsigned long) numblocks_2 * ld2_src, (unsigned long) ld3_src);
-	STARPU_ASSERT_MSG(ld3_dst >= numblocks_2 * ld2_dst, "block group group size %lu is bigger than group group ld %lu in destination", (unsigned long) numblocks_2 * ld2_dst, (unsigned long) ld3_dst);
+	STARPU_ASSERT_MSG(ld3_src >= numblocks_2 * ld2_src, "block group group size %lu is bigger than group group ld %lu in source", (unsigned long) (numblocks_2 * ld2_src), (unsigned long) ld3_src);
+	STARPU_ASSERT_MSG(ld3_dst >= numblocks_2 * ld2_dst, "block group group size %lu is bigger than group group ld %lu in destination", (unsigned long) (numblocks_2 * ld2_dst), (unsigned long) ld3_dst);
 
 
 	if (ld3_src == blocksize * numblocks_1 * numblocks_2 &&
 	if (ld3_src == blocksize * numblocks_1 * numblocks_2 &&
 	    ld3_dst == blocksize * numblocks_1 * numblocks_2)
 	    ld3_dst == blocksize * numblocks_1 * numblocks_2)

+ 1 - 1
src/drivers/mp_common/sink_common.c

@@ -308,7 +308,7 @@ static void _starpu_sink_common_recv_workers(struct _starpu_mp_node * node, void
 void _starpu_sink_common_worker(void)
 void _starpu_sink_common_worker(void)
 {
 {
 	struct _starpu_mp_node *node = NULL;
 	struct _starpu_mp_node *node = NULL;
-	enum _starpu_mp_command command = STARPU_MP_COMMAND_EXIT;
+	enum _starpu_mp_command command;
 	int arg_size = 0;
 	int arg_size = 0;
 	void *arg = NULL;
 	void *arg = NULL;
 	int exit_starpu = 0;
 	int exit_starpu = 0;

+ 1 - 0
src/drivers/mpi/driver_mpi_common.c

@@ -320,6 +320,7 @@ void _starpu_mpi_common_recv_from_device(const struct _starpu_mp_node *node STAR
                 struct _starpu_mpi_ms_event_request * req = _starpu_mpi_ms_event_request_new();
                 struct _starpu_mpi_ms_event_request * req = _starpu_mpi_ms_event_request_new();
 
 
                 res = MPI_Irecv(msg, len, MPI_BYTE, src_devid, ASYNC_TAG, MPI_COMM_WORLD, &req->request);
                 res = MPI_Irecv(msg, len, MPI_BYTE, src_devid, ASYNC_TAG, MPI_COMM_WORLD, &req->request);
+                STARPU_ASSERT_MSG(res == MPI_SUCCESS, "MPI Master/Slave cannot Ireceive a msg with a size of %d Bytes !", len);
 
 
                 channel->starpu_mp_common_finished_receiver++;
                 channel->starpu_mp_common_finished_receiver++;
                 channel->starpu_mp_common_finished_sender++;
                 channel->starpu_mp_common_finished_sender++;

+ 5 - 1
src/sched_policies/component_eager.c

@@ -35,11 +35,15 @@ static int eager_push_task(struct starpu_sched_component * component, struct sta
 	struct _starpu_eager_data *d = component->data;
 	struct _starpu_eager_data *d = component->data;
 	struct starpu_sched_component *target;
 	struct starpu_sched_component *target;
 
 
-	if (d->ntasks == 0)
+	//STARPU_COMPONENT_MUTEX_LOCK(&d->scheduling_mutex);
+	if (d->ntasks == 0) {
 		/* We have already pushed a task down */
 		/* We have already pushed a task down */
+		//STARPU_COMPONENT_MUTEX_UNLOCK(&d->scheduling_mutex);
 		return 1;
 		return 1;
+	}
 	if (d->ntasks > 0)
 	if (d->ntasks > 0)
 		d->ntasks--;
 		d->ntasks--;
+	//STARPU_COMPONENT_MUTEX_UNLOCK(&d->scheduling_mutex);
 
 
 	if ((target = d->target))
 	if ((target = d->target))
 	{
 	{

+ 0 - 1
src/sched_policies/work_stealing_policy.c

@@ -500,7 +500,6 @@ static int select_victim_overload(struct _starpu_work_stealing_data *ws, unsigne
  */
  */
 static unsigned select_worker_overload(struct _starpu_work_stealing_data *ws, struct starpu_task *task, unsigned sched_ctx_id)
 static unsigned select_worker_overload(struct _starpu_work_stealing_data *ws, struct starpu_task *task, unsigned sched_ctx_id)
 {
 {
-	struct _starpu_work_stealing_data *ws = (struct _starpu_work_stealing_data*)starpu_sched_ctx_get_policy_data(sched_ctx_id);
 	unsigned best_worker = 0;
 	unsigned best_worker = 0;
 	float best_ratio = FLT_MAX;
 	float best_ratio = FLT_MAX;
 
 

+ 1 - 1
starpurm/dev/chameleon_test/dgemm.c

@@ -295,7 +295,7 @@ int main( int argc, char const *argv[])
 			cpu_nb = 1;
 			cpu_nb = 1;
 		}
 		}
 		int cuda_offset = rm_nb_cuda_units/2;
 		int cuda_offset = rm_nb_cuda_units/2;
-		int cuda_nb = rm_nb_cuda_units/2;
+		int cuda_nb = rm_nb_cuda_units - cuda_offset;
 		spawn_tests(cpu_offset, cpu_nb, cuda_offset, cuda_nb, &test_args);
 		spawn_tests(cpu_offset, cpu_nb, cuda_offset, cuda_nb, &test_args);
 	}
 	}
 	_wait_pending_spawns();
 	_wait_pending_spawns();

+ 0 - 2
tests/datawizard/acquire_try.c

@@ -50,9 +50,7 @@ void callback(void *arg)
 
 
 int main(int argc, char **argv)
 int main(int argc, char **argv)
 {
 {
-	unsigned i;
 	int ret;
 	int ret;
-	int nacquired;
 
 
         ret = starpu_initialize(NULL, &argc, &argv);
         ret = starpu_initialize(NULL, &argc, &argv);
 	if (ret == -ENODEV) return STARPU_TEST_SKIPPED;
 	if (ret == -ENODEV) return STARPU_TEST_SKIPPED;

+ 1 - 1
tests/datawizard/commute2.c

@@ -61,7 +61,7 @@ int main(void)
 	double *res, *a;
 	double *res, *a;
 	unsigned n=100000, i;
 	unsigned n=100000, i;
 	starpu_data_handle_t res_handle, a_handle;
 	starpu_data_handle_t res_handle, a_handle;
-	unsigned nb_tasks = 10, worker;
+	unsigned nb_tasks = 10;
 	int ret;
 	int ret;
 
 
 	ret = starpu_init(NULL);
 	ret = starpu_init(NULL);

+ 0 - 1
tests/datawizard/interfaces/variable/variable_opencl.c

@@ -74,7 +74,6 @@ void test_variable_opencl_func(void *buffers[], void *args)
 	{
 	{
 		size_t global = 1;
 		size_t global = 1;
                 size_t local = 1;
                 size_t local = 1;
-                size_t s;
                 cl_device_id device;
                 cl_device_id device;
 
 
                 starpu_opencl_get_device(devid, &device);
                 starpu_opencl_get_device(devid, &device);

+ 2 - 2
tests/datawizard/specific_node.c

@@ -136,7 +136,7 @@ int main(void)
 	unsigned ntasks = 1000;
 	unsigned ntasks = 1000;
 #endif
 #endif
 
 
-	int ret, ret2;
+	int ret;
 
 
 	ret = starpu_init(NULL);
 	ret = starpu_init(NULL);
 	if (ret == -ENODEV) return STARPU_TEST_SKIPPED;
 	if (ret == -ENODEV) return STARPU_TEST_SKIPPED;
@@ -180,7 +180,7 @@ int main(void)
 	ret = (data == ntasks) ? EXIT_SUCCESS : EXIT_FAILURE;
 	ret = (data == ntasks) ? EXIT_SUCCESS : EXIT_FAILURE;
 
 
 #ifdef STARPU_USE_OPENCL
 #ifdef STARPU_USE_OPENCL
-        ret2 = starpu_opencl_unload_opencl(&opencl_program);
+        int ret2 = starpu_opencl_unload_opencl(&opencl_program);
         STARPU_CHECK_RETURN_VALUE(ret2, "starpu_opencl_unload_opencl");
         STARPU_CHECK_RETURN_VALUE(ret2, "starpu_opencl_unload_opencl");
 #endif
 #endif
 
 

+ 0 - 1
tests/energy/energy_efficiency.c

@@ -497,7 +497,6 @@ int main(int argc, char *argv[])
 		for (n = 0; n < N; n++)
 		for (n = 0; n < N; n++)
 			starpu_data_unregister(A[m][n]);
 			starpu_data_unregister(A[m][n]);
 
 
-out:
 	starpu_shutdown();
 	starpu_shutdown();
 	return 0;
 	return 0;
 }
 }

+ 1 - 2
tests/main/callback.c

@@ -1,6 +1,6 @@
 /* StarPU --- Runtime system for heterogeneous multicore architectures.
 /* StarPU --- Runtime system for heterogeneous multicore architectures.
  *
  *
- * Copyright (C) 2020       Université de Bordeaux, CNRS (LaBRI UMR 5800), Inria
+ * Copyright (C) 2020-2021       Université de Bordeaux, CNRS (LaBRI UMR 5800), Inria
  *
  *
  * StarPU is free software; you can redistribute it and/or modify
  * StarPU is free software; you can redistribute it and/or modify
  * it under the terms of the GNU Lesser General Public License as published by
  * it under the terms of the GNU Lesser General Public License as published by
@@ -89,7 +89,6 @@ int main(void)
 	starpu_shutdown();
 	starpu_shutdown();
 	return 0;
 	return 0;
 
 
-enodev:
 	starpu_shutdown();
 	starpu_shutdown();
 	fprintf(stderr, "WARNING: No one can execute this task\n");
 	fprintf(stderr, "WARNING: No one can execute this task\n");
 	/* yes, we do not perform the computation but we did detect that no one
 	/* yes, we do not perform the computation but we did detect that no one

+ 1 - 1
tests/microbenchs/tasks_size_overhead.c

@@ -228,7 +228,7 @@ int main(int argc, char **argv)
 		goto error;
 		goto error;
 	}
 	}
 
 
-	if (mincpus <= 0)
+	if (mincpus == 0)
 		mincpus = 1;
 		mincpus = 1;
 	/* For each number of cpus, benchmark */
 	/* For each number of cpus, benchmark */
 	for (ncpus= mincpus; ncpus <= maxcpus; ncpus += cpustep)
 	for (ncpus= mincpus; ncpus <= maxcpus; ncpus += cpustep)

+ 2 - 2
tests/perfmodels/memory.c

@@ -43,8 +43,8 @@ static struct starpu_codelet my_codelet =
 
 
 double cuda_cost_function(struct starpu_task *t, struct starpu_perfmodel_arch *a, unsigned i)
 double cuda_cost_function(struct starpu_task *t, struct starpu_perfmodel_arch *a, unsigned i)
 {
 {
-	t;
-	a;
+	(void) t;
+	(void) a;
 	return (double)i;
 	return (double)i;
 }
 }
 
 

+ 1 - 4
tests/perfmodels/regression_based_energy.c

@@ -284,14 +284,11 @@ int main(int argc, char **argv)
 	starpu_vector_data_register(&handle, -1, (uintptr_t)NULL, size, sizeof(int));
 	starpu_vector_data_register(&handle, -1, (uintptr_t)NULL, size, sizeof(int));
 
 
 	struct starpu_task *task = starpu_task_create();
 	struct starpu_task *task = starpu_task_create();
-	task->cl = &memset_cl;
 	task->handles[0] = handle;
 	task->handles[0] = handle;
 	task->destroy = 0;
 	task->destroy = 0;
 
 
-	task->cl = &nl_memset_cl;
-
 	FPRINTF(stdout, "\n ////non linear regression results////\n");
 	FPRINTF(stdout, "\n ////non linear regression results////\n");
-
+	task->cl = &nl_memset_cl;
 	compare_performance(size, &nl_memset_cl, task);
 	compare_performance(size, &nl_memset_cl, task);
 
 
 	starpu_task_destroy(task);
 	starpu_task_destroy(task);

+ 2 - 4
tests/perfmodels/regression_based_gpu.c

@@ -366,17 +366,15 @@ int main(int argc, char **argv)
 	starpu_vector_data_register(&handle, -1, (uintptr_t)NULL, size, sizeof(int));
 	starpu_vector_data_register(&handle, -1, (uintptr_t)NULL, size, sizeof(int));
 
 
 	struct starpu_task *task = starpu_task_create();
 	struct starpu_task *task = starpu_task_create();
-	task->cl = &memset_cl;
 	task->handles[0] = handle;
 	task->handles[0] = handle;
 	task->destroy = 0;
 	task->destroy = 0;
 
 
 	//FPRINTF(stdout, "\n ////linear regression results////\n");
 	//FPRINTF(stdout, "\n ////linear regression results////\n");
+	//task->cl = &memset_cl;
 	//compare_performance(size, &memset_cl, task);
 	//compare_performance(size, &memset_cl, task);
 
 
-	task->cl = &nl_memset_cl;
-
 	FPRINTF(stdout, "\n ////non linear regression results////\n");
 	FPRINTF(stdout, "\n ////non linear regression results////\n");
-
+	task->cl = &nl_memset_cl;
 	compare_performance(size, &nl_memset_cl, task);
 	compare_performance(size, &nl_memset_cl, task);
 
 
 	starpu_task_destroy(task);
 	starpu_task_destroy(task);

+ 1 - 4
tests/perfmodels/regression_based_multiimpl.c

@@ -285,14 +285,11 @@ int main(int argc, char **argv)
 	starpu_vector_data_register(&handle, -1, (uintptr_t)NULL, size, sizeof(int));
 	starpu_vector_data_register(&handle, -1, (uintptr_t)NULL, size, sizeof(int));
 
 
 	struct starpu_task *task = starpu_task_create();
 	struct starpu_task *task = starpu_task_create();
-	task->cl = &memset_cl;
 	task->handles[0] = handle;
 	task->handles[0] = handle;
 	task->destroy = 0;
 	task->destroy = 0;
 
 
-	task->cl = &nl_memset_cl;
-
 	FPRINTF(stdout, "\n ////non linear regression results////\n");
 	FPRINTF(stdout, "\n ////non linear regression results////\n");
-
+	task->cl = &nl_memset_cl;
 	compare_performance(size, &nl_memset_cl, task);
 	compare_performance(size, &nl_memset_cl, task);
 
 
 	starpu_task_destroy(task);
 	starpu_task_destroy(task);