Warning: That file was not part of the compilation database. It may have many parsing errors.

1/* Copyright (C) 2003,2004 Andi Kleen, SuSE Labs.
2
3 libnuma is free software; you can redistribute it and/or
4 modify it under the terms of the GNU Lesser General Public
5 License as published by the Free Software Foundation; version
6 2.1.
7
8 libnuma is distributed in the hope that it will be useful,
9 but WITHOUT ANY WARRANTY; without even the implied warranty of
10 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
11 Lesser General Public License for more details.
12
13 You should find a copy of v2.1 of the GNU Lesser General Public License
14 somewhere on your Linux system; if not, write to the Free Software
15 Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA */
16
17#ifndef _NUMA_H
18#define _NUMA_H 1
19
20/* allow an application to test for the current programming interface: */
21#define LIBNUMA_API_VERSION 2
22
23/* Simple NUMA policy library */
24
25#include <stddef.h>
26#include <string.h>
27#include <sys/types.h>
28#include <stdlib.h>
29
30#if defined(__x86_64__) || defined(__i386__)
31#define NUMA_NUM_NODES 128
32#else
33#define NUMA_NUM_NODES 2048
34#endif
35
36#ifdef __cplusplus
37extern "C" {
38#endif
39
40typedef struct {
41 unsigned long n[NUMA_NUM_NODES/(sizeof(unsigned long)*8)];
42} nodemask_t;
43
44struct bitmask {
45 unsigned long size; /* number of bits in the map */
46 unsigned long *maskp;
47};
48
49/* operations on struct bitmask */
50int numa_bitmask_isbitset(const struct bitmask *, unsigned int);
51struct bitmask *numa_bitmask_setall(struct bitmask *);
52struct bitmask *numa_bitmask_clearall(struct bitmask *);
53struct bitmask *numa_bitmask_setbit(struct bitmask *, unsigned int);
54struct bitmask *numa_bitmask_clearbit(struct bitmask *, unsigned int);
55unsigned int numa_bitmask_nbytes(struct bitmask *);
56unsigned int numa_bitmask_weight(const struct bitmask *);
57struct bitmask *numa_bitmask_alloc(unsigned int);
58void numa_bitmask_free(struct bitmask *);
59int numa_bitmask_equal(const struct bitmask *, const struct bitmask *);
60void copy_nodemask_to_bitmask(nodemask_t *, struct bitmask *);
61void copy_bitmask_to_nodemask(struct bitmask *, nodemask_t *);
62void copy_bitmask_to_bitmask(struct bitmask *, struct bitmask *);
63
64/* compatibility for codes that used them: */
65
66static inline void nodemask_zero(nodemask_t *mask)
67{
68 struct bitmask tmp;
69
70 tmp.maskp = (unsigned long *)mask;
71 tmp.size = sizeof(nodemask_t) * 8;
72 numa_bitmask_clearall(&tmp);
73}
74
75static inline void nodemask_zero_compat(nodemask_t *mask)
76{
77 struct bitmask tmp;
78
79 tmp.maskp = (unsigned long *)mask;
80 tmp.size = sizeof(nodemask_t) * 8;
81 numa_bitmask_clearall(&tmp);
82}
83
84static inline void nodemask_set_compat(nodemask_t *mask, int node)
85{
86 mask->n[node / (8*sizeof(unsigned long))] |=
87 (1UL<<(node%(8*sizeof(unsigned long))));
88}
89
90static inline void nodemask_clr_compat(nodemask_t *mask, int node)
91{
92 mask->n[node / (8*sizeof(unsigned long))] &=
93 ~(1UL<<(node%(8*sizeof(unsigned long))));
94}
95
96static inline int nodemask_isset_compat(const nodemask_t *mask, int node)
97{
98 if ((unsigned)node >= NUMA_NUM_NODES)
99 return 0;
100 if (mask->n[node / (8*sizeof(unsigned long))] &
101 (1UL<<(node%(8*sizeof(unsigned long)))))
102 return 1;
103 return 0;
104}
105
106static inline int nodemask_equal(const nodemask_t *a, const nodemask_t *b)
107{
108 struct bitmask tmp_a, tmp_b;
109
110 tmp_a.maskp = (unsigned long *)a;
111 tmp_a.size = sizeof(nodemask_t) * 8;
112
113 tmp_b.maskp = (unsigned long *)b;
114 tmp_b.size = sizeof(nodemask_t) * 8;
115
116 return numa_bitmask_equal(&tmp_a, &tmp_b);
117}
118
119static inline int nodemask_equal_compat(const nodemask_t *a, const nodemask_t *b)
120{
121 struct bitmask tmp_a, tmp_b;
122
123 tmp_a.maskp = (unsigned long *)a;
124 tmp_a.size = sizeof(nodemask_t) * 8;
125
126 tmp_b.maskp = (unsigned long *)b;
127 tmp_b.size = sizeof(nodemask_t) * 8;
128
129 return numa_bitmask_equal(&tmp_a, &tmp_b);
130}
131
132/* NUMA support available. If this returns a negative value all other function
133 in this library are undefined. */
134int numa_available(void);
135
136/* Basic NUMA state */
137
138/* Get max available node */
139int numa_max_node(void);
140int numa_max_possible_node(void);
141/* Return preferred node */
142int numa_preferred(void);
143
144/* Return node size and free memory */
145long long numa_node_size64(int node, long long *freep);
146long numa_node_size(int node, long *freep);
147
148int numa_pagesize(void);
149
150/* Set with all nodes from which the calling process may allocate memory.
151 Only valid after numa_available. */
152extern struct bitmask *numa_all_nodes_ptr;
153
154/* Set with all nodes the kernel has exposed to userspace */
155extern struct bitmask *numa_nodes_ptr;
156
157/* For source compatibility */
158extern nodemask_t numa_all_nodes;
159
160/* Set with all cpus. */
161extern struct bitmask *numa_all_cpus_ptr;
162
163/* Set with no nodes */
164extern struct bitmask *numa_no_nodes_ptr;
165
166/* Source compatibility */
167extern nodemask_t numa_no_nodes;
168
169/* Only run and allocate memory from a specific set of nodes. */
170void numa_bind(struct bitmask *nodes);
171
172/* Set the NUMA node interleaving mask. 0 to turn off interleaving */
173void numa_set_interleave_mask(struct bitmask *nodemask);
174
175/* Return the current interleaving mask */
176struct bitmask *numa_get_interleave_mask(void);
177
178/* allocate a bitmask big enough for all nodes */
179struct bitmask *numa_allocate_nodemask(void);
180
181static inline void numa_free_nodemask(struct bitmask *b)
182{
183 numa_bitmask_free(b);
184}
185
186/* Some node to preferably allocate memory from for task. */
187void numa_set_preferred(int node);
188
189/* Set local memory allocation policy for task */
190void numa_set_localalloc(void);
191
192/* Only allocate memory from the nodes set in mask. 0 to turn off */
193void numa_set_membind(struct bitmask *nodemask);
194
195/* Return current membind */
196struct bitmask *numa_get_membind(void);
197
198/* Return allowed memories [nodes] */
199struct bitmask *numa_get_mems_allowed(void);
200
201int numa_get_interleave_node(void);
202
203/* NUMA memory allocation. These functions always round to page size
204 and are relatively slow. */
205
206/* Alloc memory page interleaved on nodes in mask */
207void *numa_alloc_interleaved_subset(size_t size, struct bitmask *nodemask);
208/* Alloc memory page interleaved on all nodes. */
209void *numa_alloc_interleaved(size_t size);
210/* Alloc memory located on node */
211void *numa_alloc_onnode(size_t size, int node);
212/* Alloc memory on local node */
213void *numa_alloc_local(size_t size);
214/* Allocation with current policy */
215void *numa_alloc(size_t size);
216/* Change the size of a memory area preserving the memory policy */
217void *numa_realloc(void *old_addr, size_t old_size, size_t new_size);
218/* Free memory allocated by the functions above */
219void numa_free(void *mem, size_t size);
220
221/* Low level functions, primarily for shared memory. All memory
222 processed by these must not be touched yet */
223
224/* Interleave an memory area. */
225void numa_interleave_memory(void *mem, size_t size, struct bitmask *mask);
226
227/* Allocate a memory area on a specific node. */
228void numa_tonode_memory(void *start, size_t size, int node);
229
230/* Allocate memory on a mask of nodes. */
231void numa_tonodemask_memory(void *mem, size_t size, struct bitmask *mask);
232
233/* Allocate a memory area on the current node. */
234void numa_setlocal_memory(void *start, size_t size);
235
236/* Allocate memory area with current memory policy */
237void numa_police_memory(void *start, size_t size);
238
239/* Run current task only on nodes in mask */
240int numa_run_on_node_mask(struct bitmask *mask);
241/* Run current task on nodes in mask without any cpuset awareness */
242int numa_run_on_node_mask_all(struct bitmask *mask);
243/* Run current task only on node */
244int numa_run_on_node(int node);
245/* Return current mask of nodes the task can run on */
246struct bitmask * numa_get_run_node_mask(void);
247
248/* When strict fail allocation when memory cannot be allocated in target node(s). */
249void numa_set_bind_policy(int strict);
250
251/* Fail when existing memory has incompatible policy */
252void numa_set_strict(int flag);
253
254/* maximum nodes (size of kernel nodemask_t) */
255int numa_num_possible_nodes();
256
257/* maximum cpus (size of kernel cpumask_t) */
258int numa_num_possible_cpus();
259
260/* nodes in the system */
261int numa_num_configured_nodes();
262
263/* maximum cpus */
264int numa_num_configured_cpus();
265
266/* maximum cpus allowed to current task */
267int numa_num_task_cpus();
268int numa_num_thread_cpus(); /* backward compatibility */
269
270/* maximum nodes allowed to current task */
271int numa_num_task_nodes();
272int numa_num_thread_nodes(); /* backward compatibility */
273
274/* allocate a bitmask the size of the kernel cpumask_t */
275struct bitmask *numa_allocate_cpumask();
276
277static inline void numa_free_cpumask(struct bitmask *b)
278{
279 numa_bitmask_free(b);
280}
281
282/* Convert node to CPU mask. -1/errno on failure, otherwise 0. */
283int numa_node_to_cpus(int, struct bitmask *);
284
285/* report the node of the specified cpu. -1/errno on invalid cpu. */
286int numa_node_of_cpu(int cpu);
287
288/* Report distance of node1 from node2. 0 on error.*/
289int numa_distance(int node1, int node2);
290
291/* Error handling. */
292/* This is an internal function in libnuma that can be overwritten by an user
293 program. Default is to print an error to stderr and exit if numa_exit_on_error
294 is true. */
295void numa_error(char *where);
296
297/* When true exit the program when a NUMA system call (except numa_available)
298 fails */
299extern int numa_exit_on_error;
300/* Warning function. Can also be overwritten. Default is to print on stderr
301 once. */
302void numa_warn(int num, char *fmt, ...);
303
304/* When true exit the program on a numa_warn() call */
305extern int numa_exit_on_warn;
306
307int numa_migrate_pages(int pid, struct bitmask *from, struct bitmask *to);
308
309int numa_move_pages(int pid, unsigned long count, void **pages,
310 const int *nodes, int *status, int flags);
311
312int numa_sched_getaffinity(pid_t, struct bitmask *);
313int numa_sched_setaffinity(pid_t, struct bitmask *);
314
315/* Convert an ascii list of nodes to a bitmask */
316struct bitmask *numa_parse_nodestring(const char *);
317
318/* Convert an ascii list of nodes to a bitmask without current nodeset
319 * dependency */
320struct bitmask *numa_parse_nodestring_all(const char *);
321
322/* Convert an ascii list of cpu to a bitmask */
323struct bitmask *numa_parse_cpustring(const char *);
324
325/* Convert an ascii list of cpu to a bitmask without current taskset
326 * dependency */
327struct bitmask *numa_parse_cpustring_all(const char *);
328
329/*
330 * The following functions are for source code compatibility
331 * with releases prior to version 2.
332 * Such codes should be compiled with NUMA_VERSION1_COMPATIBILITY defined.
333 */
334
335static inline void numa_set_interleave_mask_compat(nodemask_t *nodemask)
336{
337 struct bitmask tmp;
338
339 tmp.maskp = (unsigned long *)nodemask;
340 tmp.size = sizeof(nodemask_t) * 8;
341 numa_set_interleave_mask(&tmp);
342}
343
344static inline nodemask_t numa_get_interleave_mask_compat()
345{
346 struct bitmask *tp;
347 nodemask_t mask;
348
349 tp = numa_get_interleave_mask();
350 copy_bitmask_to_nodemask(tp, &mask);
351 numa_bitmask_free(tp);
352 return mask;
353}
354
355static inline void numa_bind_compat(nodemask_t *mask)
356{
357 struct bitmask *tp;
358
359 tp = numa_allocate_nodemask();
360 copy_nodemask_to_bitmask(mask, tp);
361 numa_bind(tp);
362 numa_bitmask_free(tp);
363}
364
365static inline void numa_set_membind_compat(nodemask_t *mask)
366{
367 struct bitmask tmp;
368
369 tmp.maskp = (unsigned long *)mask;
370 tmp.size = sizeof(nodemask_t) * 8;
371 numa_set_membind(&tmp);
372}
373
374static inline nodemask_t numa_get_membind_compat()
375{
376 struct bitmask *tp;
377 nodemask_t mask;
378
379 tp = numa_get_membind();
380 copy_bitmask_to_nodemask(tp, &mask);
381 numa_bitmask_free(tp);
382 return mask;
383}
384
385static inline void *numa_alloc_interleaved_subset_compat(size_t size,
386 const nodemask_t *mask)
387{
388 struct bitmask tmp;
389
390 tmp.maskp = (unsigned long *)mask;
391 tmp.size = sizeof(nodemask_t) * 8;
392 return numa_alloc_interleaved_subset(size, &tmp);
393}
394
395static inline int numa_run_on_node_mask_compat(const nodemask_t *mask)
396{
397 struct bitmask tmp;
398
399 tmp.maskp = (unsigned long *)mask;
400 tmp.size = sizeof(nodemask_t) * 8;
401 return numa_run_on_node_mask(&tmp);
402}
403
404static inline nodemask_t numa_get_run_node_mask_compat()
405{
406 struct bitmask *tp;
407 nodemask_t mask;
408
409 tp = numa_get_run_node_mask();
410 copy_bitmask_to_nodemask(tp, &mask);
411 numa_bitmask_free(tp);
412 return mask;
413}
414
415static inline void numa_interleave_memory_compat(void *mem, size_t size,
416 const nodemask_t *mask)
417{
418 struct bitmask tmp;
419
420 tmp.maskp = (unsigned long *)mask;
421 tmp.size = sizeof(nodemask_t) * 8;
422 numa_interleave_memory(mem, size, &tmp);
423}
424
425static inline void numa_tonodemask_memory_compat(void *mem, size_t size,
426 const nodemask_t *mask)
427{
428 struct bitmask tmp;
429
430 tmp.maskp = (unsigned long *)mask;
431 tmp.size = sizeof(nodemask_t) * 8;
432 numa_tonodemask_memory(mem, size, &tmp);
433}
434
435static inline int numa_sched_getaffinity_compat(pid_t pid, unsigned len,
436 unsigned long *mask)
437{
438 struct bitmask tmp;
439
440 tmp.maskp = (unsigned long *)mask;
441 tmp.size = len * 8;
442 return numa_sched_getaffinity(pid, &tmp);
443}
444
445static inline int numa_sched_setaffinity_compat(pid_t pid, unsigned len,
446 unsigned long *mask)
447{
448 struct bitmask tmp;
449
450 tmp.maskp = (unsigned long *)mask;
451 tmp.size = len * 8;
452 return numa_sched_setaffinity(pid, &tmp);
453}
454
455static inline int numa_node_to_cpus_compat(int node, unsigned long *buffer,
456 int buffer_len)
457{
458 struct bitmask tmp;
459
460 tmp.maskp = (unsigned long *)buffer;
461 tmp.size = buffer_len * 8;
462 return numa_node_to_cpus(node, &tmp);
463}
464
465/* end of version 1 compatibility functions */
466
467/*
468 * To compile an application that uses libnuma version 1:
469 * add -DNUMA_VERSION1_COMPATIBILITY to your Makefile's CFLAGS
470 */
471#ifdef NUMA_VERSION1_COMPATIBILITY
472#include <numacompat1.h>
473#endif
474
475#ifdef __cplusplus
476}
477#endif
478
479#endif
480

Warning: That file was not part of the compilation database. It may have many parsing errors.