1/* System specific fork hooks. Linux version.
2 Copyright (C) 2021-2022 Free Software Foundation, Inc.
3 This file is part of the GNU C Library.
4
5 The GNU C Library is free software; you can redistribute it and/or
6 modify it under the terms of the GNU Lesser General Public
7 License as published by the Free Software Foundation; either
8 version 2.1 of the License, or (at your option) any later version.
9
10 The GNU C Library is distributed in the hope that it will be useful,
11 but WITHOUT ANY WARRANTY; without even the implied warranty of
12 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
13 Lesser General Public License for more details.
14
15 You should have received a copy of the GNU Lesser General Public
16 License along with the GNU C Library; if not, see
17 <https://www.gnu.org/licenses/>. */
18
19#ifndef _FORK_H
20#define _FORK_H
21
22#include <assert.h>
23#include <kernel-posix-timers.h>
24#include <ldsodefs.h>
25#include <list.h>
26#include <mqueue.h>
27#include <pthreadP.h>
28#include <sysdep.h>
29
30static inline void
31fork_system_setup (void)
32{
33 /* See __pthread_once. */
34 __fork_generation += __PTHREAD_ONCE_FORK_GEN_INCR;
35}
36
37static void
38fork_system_setup_after_fork (void)
39{
40 /* There is one thread running. */
41 __nptl_nthreads = 1;
42
43 /* Initialize thread library locks. */
44 GL (dl_stack_cache_lock) = LLL_LOCK_INITIALIZER;
45 __default_pthread_attr_lock = LLL_LOCK_INITIALIZER;
46
47 call_function_static_weak (__mq_notify_fork_subprocess);
48 call_function_static_weak (__timer_fork_subprocess);
49}
50
51/* In case of a fork() call the memory allocation in the child will be
52 the same but only one thread is running. All stacks except that of
53 the one running thread are not used anymore. We have to recycle
54 them. */
55static void
56reclaim_stacks (void)
57{
58 struct pthread *self = (struct pthread *) THREAD_SELF;
59
60 /* No locking necessary. The caller is the only stack in use. But
61 we have to be aware that we might have interrupted a list
62 operation. */
63
64 if (GL (dl_in_flight_stack) != 0)
65 {
66 bool add_p = GL (dl_in_flight_stack) & 1;
67 list_t *elem = (list_t *) (GL (dl_in_flight_stack) & ~(uintptr_t) 1);
68
69 if (add_p)
70 {
71 /* We always add at the beginning of the list. So in this case we
72 only need to check the beginning of these lists to see if the
73 pointers at the head of the list are inconsistent. */
74 list_t *l = NULL;
75
76 if (GL (dl_stack_used).next->prev != &GL (dl_stack_used))
77 l = &GL (dl_stack_used);
78 else if (GL (dl_stack_cache).next->prev != &GL (dl_stack_cache))
79 l = &GL (dl_stack_cache);
80
81 if (l != NULL)
82 {
83 assert (l->next->prev == elem);
84 elem->next = l->next;
85 elem->prev = l;
86 l->next = elem;
87 }
88 }
89 else
90 {
91 /* We can simply always replay the delete operation. */
92 elem->next->prev = elem->prev;
93 elem->prev->next = elem->next;
94 }
95
96 GL (dl_in_flight_stack) = 0;
97 }
98
99 /* Mark all stacks except the still running one as free. */
100 list_t *runp;
101 list_for_each (runp, &GL (dl_stack_used))
102 {
103 struct pthread *curp = list_entry (runp, struct pthread, list);
104 if (curp != self)
105 {
106 /* This marks the stack as free. */
107 curp->tid = 0;
108
109 /* Account for the size of the stack. */
110 GL (dl_stack_cache_actsize) += curp->stackblock_size;
111
112 if (curp->specific_used)
113 {
114 /* Clear the thread-specific data. */
115 memset (curp->specific_1stblock, '\0',
116 sizeof (curp->specific_1stblock));
117
118 curp->specific_used = false;
119
120 for (size_t cnt = 1; cnt < PTHREAD_KEY_1STLEVEL_SIZE; ++cnt)
121 if (curp->specific[cnt] != NULL)
122 {
123 memset (curp->specific[cnt], '\0',
124 sizeof (curp->specific_1stblock));
125
126 /* We have allocated the block which we do not
127 free here so re-set the bit. */
128 curp->specific_used = true;
129 }
130 }
131 }
132 }
133
134 /* Add the stack of all running threads to the cache. */
135 list_splice (add: &GL (dl_stack_used), head: &GL (dl_stack_cache));
136
137 /* Remove the entry for the current thread to from the cache list
138 and add it to the list of running threads. Which of the two
139 lists is decided by the user_stack flag. */
140 list_del (elem: &self->list);
141
142 /* Re-initialize the lists for all the threads. */
143 INIT_LIST_HEAD (&GL (dl_stack_used));
144 INIT_LIST_HEAD (&GL (dl_stack_user));
145
146 if (__glibc_unlikely (THREAD_GETMEM (self, user_stack)))
147 list_add (newp: &self->list, head: &GL (dl_stack_user));
148 else
149 list_add (newp: &self->list, head: &GL (dl_stack_used));
150}
151
152
153#endif
154

source code of glibc/sysdeps/nptl/fork.h