/* StarPU --- Runtime system for heterogeneous multicore architectures. * * Copyright (C) 2009-2012, 2014-2016 Université de Bordeaux * Copyright (C) 2010, 2011, 2013 CNRS * * StarPU is free software; you can redistribute it and/or modify * it under the terms of the GNU Lesser General Public License as published by * the Free Software Foundation; either version 2.1 of the License, or (at * your option) any later version. * * StarPU is distributed in the hope that it will be useful, but * WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. * * See the GNU Lesser General Public License in COPYING.LGPL for more details. */ #ifndef __MEMORY_NODES_H__ #define __MEMORY_NODES_H__ #include #include #include #include #include #include #ifdef STARPU_SIMGRID #include #endif #define _STARPU_MEMORY_NODE_TUPLE(node1,node2) (node1 | (node2 << 4)) #define _STARPU_MEMORY_NODE_TUPLE_FIRST(tuple) (tuple & 0x0F) #define _STARPU_MEMORY_NODE_TUPLE_SECOND(tuple) (tuple & 0xF0) extern char _starpu_worker_drives_memory[STARPU_NMAXWORKERS][STARPU_MAXNODES]; struct _starpu_cond_and_mutex { starpu_pthread_cond_t *cond; starpu_pthread_mutex_t *mutex; }; struct _starpu_memory_node_descr { unsigned nnodes; enum starpu_node_kind nodes[STARPU_MAXNODES]; /* Get the device id associated to this node, or -1 if not applicable */ int devid[STARPU_MAXNODES]; unsigned nworkers[STARPU_MAXNODES]; #ifdef STARPU_SIMGRID msg_host_t host[STARPU_MAXNODES]; #endif // TODO move this 2 lists outside struct _starpu_memory_node_descr /* Every worker is associated to a condition variable on which the * worker waits when there is task available. It is possible that * multiple worker share the same condition variable, so we maintain a * list of all these condition variables so that we can wake up all * worker attached to a memory node that are waiting on a task. */ starpu_pthread_rwlock_t conditions_rwlock; struct _starpu_cond_and_mutex conditions_attached_to_node[STARPU_MAXNODES][STARPU_NMAXWORKERS]; struct _starpu_cond_and_mutex conditions_all[STARPU_MAXNODES*STARPU_NMAXWORKERS]; /* the number of queues attached to each node */ unsigned total_condition_count; unsigned condition_count[STARPU_MAXNODES]; }; extern struct _starpu_memory_node_descr _starpu_descr; void _starpu_memory_nodes_init(void); void _starpu_memory_nodes_deinit(void); extern starpu_pthread_key_t _starpu_memory_node_key STARPU_ATTRIBUTE_INTERNAL; static inline void _starpu_memory_node_set_local_key(unsigned *node) { STARPU_PTHREAD_SETSPECIFIC(_starpu_memory_node_key, node); } static inline unsigned _starpu_memory_node_get_local_key(void) { unsigned *memory_node; memory_node = (unsigned *) STARPU_PTHREAD_GETSPECIFIC(_starpu_memory_node_key); /* in case this is called by the programmer, we assume the RAM node is the appropriate memory node ... XXX */ if (STARPU_UNLIKELY(!memory_node)) return STARPU_MAIN_RAM; return *memory_node; } static inline void _starpu_memory_node_add_nworkers(unsigned node) { _starpu_descr.nworkers[node]++; } /* same utility as _starpu_memory_node_add_nworkers */ static inline void _starpu_worker_drives_memory_node(unsigned worker_id, unsigned memnode) { _starpu_worker_drives_memory[worker_id][memnode] = 1; } static inline unsigned _starpu_memory_node_get_nworkers(unsigned node) { return _starpu_descr.nworkers[node]; } #ifdef STARPU_SIMGRID static inline void _starpu_simgrid_memory_node_set_host(unsigned node, msg_host_t host) { _starpu_descr.host[node] = host; } static inline msg_host_t _starpu_simgrid_memory_node_get_host(unsigned node) { return _starpu_descr.host[node]; } #endif unsigned _starpu_memory_node_register(enum starpu_node_kind kind, int devid); //void _starpu_memory_node_attach_queue(struct starpu_jobq_s *q, unsigned nodeid); void _starpu_memory_node_register_condition(starpu_pthread_cond_t *cond, starpu_pthread_mutex_t *mutex, unsigned memory_node); static inline int _starpu_memory_node_get_devid(unsigned node) { return _starpu_descr.devid[node]; } void _starpu_memory_node_get_name(unsigned node, char *name, int size); static inline struct _starpu_memory_node_descr *_starpu_memory_node_get_description(void) { return &_starpu_descr; } static inline enum starpu_node_kind _starpu_node_get_kind(unsigned node) { return _starpu_descr.nodes[node]; } #define starpu_node_get_kind _starpu_node_get_kind static inline unsigned _starpu_memory_nodes_get_count(void) { return _starpu_descr.nnodes; } #define starpu_memory_nodes_get_count _starpu_memory_nodes_get_count static inline unsigned _starpu_worker_get_memory_node(unsigned workerid) { struct _starpu_machine_config *config = _starpu_get_machine_config(); /* This workerid may either be a basic worker or a combined worker */ unsigned nworkers = config->topology.nworkers; if (workerid < config->topology.nworkers) return config->workers[workerid].memory_node; /* We have a combined worker */ unsigned ncombinedworkers STARPU_ATTRIBUTE_UNUSED = config->topology.ncombinedworkers; STARPU_ASSERT_MSG(workerid < ncombinedworkers + nworkers, "Bad workerid %u, maximum %u", workerid, ncombinedworkers + nworkers); return config->combined_workers[workerid - nworkers].memory_node; } #define starpu_worker_get_memory_node _starpu_worker_get_memory_node #endif // __MEMORY_NODES_H__