1 1.23 christos /* $NetBSD: tls.c,v 1.23 2024/11/30 01:04:05 christos Exp $ */ 2 1.1 joerg /*- 3 1.1 joerg * Copyright (c) 2011 The NetBSD Foundation, Inc. 4 1.1 joerg * All rights reserved. 5 1.1 joerg * 6 1.1 joerg * This code is derived from software contributed to The NetBSD Foundation 7 1.1 joerg * by Joerg Sonnenberger. 8 1.1 joerg * 9 1.1 joerg * Redistribution and use in source and binary forms, with or without 10 1.1 joerg * modification, are permitted provided that the following conditions 11 1.1 joerg * are met: 12 1.1 joerg * 1. Redistributions of source code must retain the above copyright 13 1.1 joerg * notice, this list of conditions and the following disclaimer. 14 1.1 joerg * 2. Redistributions in binary form must reproduce the above copyright 15 1.1 joerg * notice, this list of conditions and the following disclaimer in the 16 1.1 joerg * documentation and/or other materials provided with the distribution. 17 1.1 joerg * 18 1.1 joerg * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS 19 1.1 joerg * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED 20 1.1 joerg * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR 21 1.1 joerg * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS 22 1.1 joerg * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 23 1.1 joerg * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 24 1.1 joerg * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 25 1.1 joerg * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 26 1.1 joerg * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 27 1.1 joerg * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 28 1.1 joerg * POSSIBILITY OF SUCH DAMAGE. 29 1.1 joerg */ 30 1.1 joerg 31 1.1 joerg #include <sys/cdefs.h> 32 1.23 christos __RCSID("$NetBSD: tls.c,v 1.23 2024/11/30 01:04:05 christos Exp $"); 33 1.20 riastrad 34 1.20 riastrad /* 35 1.20 riastrad * Thread-local storage 36 1.20 riastrad * 37 1.20 riastrad * Reference: 38 1.20 riastrad * 39 1.20 riastrad * [ELFTLS] Ulrich Drepper, `ELF Handling For Thread-Local 40 1.20 riastrad * Storage', Version 0.21, 2023-08-22. 41 1.20 riastrad * https://akkadia.org/drepper/tls.pdf 42 1.20 riastrad * https://web.archive.org/web/20240718081934/https://akkadia.org/drepper/tls.pdf 43 1.20 riastrad */ 44 1.1 joerg 45 1.1 joerg #include <sys/param.h> 46 1.2 joerg #include <sys/ucontext.h> 47 1.1 joerg #include <lwp.h> 48 1.14 joerg #include <stdalign.h> 49 1.13 joerg #include <stddef.h> 50 1.1 joerg #include <string.h> 51 1.8 skrll #include "debug.h" 52 1.1 joerg #include "rtld.h" 53 1.1 joerg 54 1.23 christos #include <machine/lwp_private.h> 55 1.23 christos 56 1.1 joerg #if defined(__HAVE_TLS_VARIANT_I) || defined(__HAVE_TLS_VARIANT_II) 57 1.1 joerg 58 1.4 joerg static struct tls_tcb *_rtld_tls_allocate_locked(void); 59 1.18 joerg static void *_rtld_tls_module_allocate(struct tls_tcb *, size_t); 60 1.4 joerg 61 1.20 riastrad /* 62 1.20 riastrad * DTV offset 63 1.20 riastrad * 64 1.20 riastrad * On some architectures (m68k, mips, or1k, powerpc, and riscv), 65 1.20 riastrad * the DTV offsets passed to __tls_get_addr have a bias relative 66 1.20 riastrad * to the start of the DTV, in order to maximize the range of TLS 67 1.20 riastrad * offsets that can be used by instruction encodings with signed 68 1.20 riastrad * displacements. 69 1.20 riastrad */ 70 1.3 matt #ifndef TLS_DTV_OFFSET 71 1.3 matt #define TLS_DTV_OFFSET 0 72 1.3 matt #endif 73 1.3 matt 74 1.1 joerg static size_t _rtld_tls_static_space; /* Static TLS space allocated */ 75 1.1 joerg static size_t _rtld_tls_static_offset; /* Next offset for static TLS to use */ 76 1.20 riastrad size_t _rtld_tls_dtv_generation = 1; /* Bumped on each load of obj w/ TLS */ 77 1.20 riastrad size_t _rtld_tls_max_index = 1; /* Max index into up-to-date DTV */ 78 1.1 joerg 79 1.20 riastrad /* 80 1.20 riastrad * DTV -- Dynamic Thread Vector 81 1.20 riastrad * 82 1.20 riastrad * The DTV is a per-thread array that maps each module with 83 1.20 riastrad * thread-local storage to a pointer into part of the thread's TCB 84 1.20 riastrad * (thread control block), or dynamically loaded TLS blocks, 85 1.20 riastrad * reserved for that module's storage. 86 1.20 riastrad * 87 1.20 riastrad * The TCB itself, struct tls_tcb, has a pointer to the DTV at 88 1.20 riastrad * tcb->tcb_dtv. 89 1.20 riastrad * 90 1.20 riastrad * The layout is: 91 1.20 riastrad * 92 1.20 riastrad * +---------------+ 93 1.20 riastrad * | max index | -1 max index i for which dtv[i] is alloced 94 1.20 riastrad * +---------------+ 95 1.20 riastrad * | generation | 0 void **dtv points here 96 1.20 riastrad * +---------------+ 97 1.20 riastrad * | obj 1 tls ptr | 1 TLS pointer for obj w/ obj->tlsindex 1 98 1.20 riastrad * +---------------+ 99 1.20 riastrad * | obj 2 tls ptr | 2 TLS pointer for obj w/ obj->tlsindex 2 100 1.20 riastrad * +---------------+ 101 1.20 riastrad * . 102 1.20 riastrad * . 103 1.20 riastrad * . 104 1.20 riastrad * 105 1.20 riastrad * The values of obj->tlsindex start at 1; this way, 106 1.20 riastrad * dtv[obj->tlsindex] works, when dtv[0] is the generation. The 107 1.20 riastrad * TLS pointers go either into the static thread-local storage, 108 1.20 riastrad * for the initial objects (i.e., those loaded at startup), or 109 1.20 riastrad * into TLS blocks dynamically allocated for objects that 110 1.20 riastrad * dynamically loaded by dlopen. 111 1.20 riastrad * 112 1.20 riastrad * The generation field is a cache of the global generation number 113 1.20 riastrad * _rtld_tls_dtv_generation, which is bumped every time an object 114 1.20 riastrad * with TLS is loaded in _rtld_map_object, and cached by 115 1.20 riastrad * __tls_get_addr (via _rtld_tls_get_addr) when a newly loaded 116 1.20 riastrad * module lies outside the bounds of the current DTV. 117 1.20 riastrad * 118 1.20 riastrad * XXX Why do we keep max index and generation separately? They 119 1.20 riastrad * appear to be initialized the same, always incremented together, 120 1.20 riastrad * and always stored together. 121 1.20 riastrad * 122 1.20 riastrad * XXX Why is this not a struct? 123 1.20 riastrad * 124 1.20 riastrad * struct dtv { 125 1.20 riastrad * size_t dtv_gen; 126 1.20 riastrad * void *dtv_module[]; 127 1.20 riastrad * }; 128 1.20 riastrad */ 129 1.15 skrll #define DTV_GENERATION(dtv) ((size_t)((dtv)[0])) 130 1.15 skrll #define DTV_MAX_INDEX(dtv) ((size_t)((dtv)[-1])) 131 1.1 joerg #define SET_DTV_GENERATION(dtv, val) (dtv)[0] = (void *)(size_t)(val) 132 1.1 joerg #define SET_DTV_MAX_INDEX(dtv, val) (dtv)[-1] = (void *)(size_t)(val) 133 1.1 joerg 134 1.20 riastrad /* 135 1.20 riastrad * _rtld_tls_get_addr(tcb, idx, offset) 136 1.20 riastrad * 137 1.20 riastrad * Slow path for __tls_get_addr (see below), called to allocate 138 1.20 riastrad * TLS space if needed for the object obj with obj->tlsindex idx, 139 1.20 riastrad * at offset, which must be below obj->tlssize. 140 1.20 riastrad * 141 1.20 riastrad * This may allocate a DTV if the current one is too old, and it 142 1.20 riastrad * may allocate a dynamically loaded TLS block if there isn't one 143 1.20 riastrad * already allocated for it. 144 1.20 riastrad * 145 1.20 riastrad * XXX Why is the first argument passed as `void *tls' instead of 146 1.20 riastrad * just `struct tls_tcb *tcb'? 147 1.20 riastrad */ 148 1.1 joerg void * 149 1.1 joerg _rtld_tls_get_addr(void *tls, size_t idx, size_t offset) 150 1.1 joerg { 151 1.1 joerg struct tls_tcb *tcb = tls; 152 1.1 joerg void **dtv, **new_dtv; 153 1.5 joerg sigset_t mask; 154 1.1 joerg 155 1.5 joerg _rtld_exclusive_enter(&mask); 156 1.4 joerg 157 1.1 joerg dtv = tcb->tcb_dtv; 158 1.1 joerg 159 1.20 riastrad /* 160 1.20 riastrad * If the generation number has changed, we have to allocate a 161 1.20 riastrad * new DTV. 162 1.20 riastrad * 163 1.20 riastrad * XXX Do we really? Isn't it enough to check whether idx <= 164 1.20 riastrad * DTV_MAX_INDEX(dtv)? 165 1.20 riastrad */ 166 1.1 joerg if (__predict_false(DTV_GENERATION(dtv) != _rtld_tls_dtv_generation)) { 167 1.1 joerg size_t to_copy = DTV_MAX_INDEX(dtv); 168 1.1 joerg 169 1.21 riastrad /* 170 1.21 riastrad * "2 +" because the first element is the generation and 171 1.21 riastrad * the second one is the maximum index. 172 1.21 riastrad */ 173 1.1 joerg new_dtv = xcalloc((2 + _rtld_tls_max_index) * sizeof(*dtv)); 174 1.20 riastrad ++new_dtv; /* advance past DTV_MAX_INDEX */ 175 1.20 riastrad if (to_copy > _rtld_tls_max_index) /* XXX How? */ 176 1.1 joerg to_copy = _rtld_tls_max_index; 177 1.1 joerg memcpy(new_dtv + 1, dtv + 1, to_copy * sizeof(*dtv)); 178 1.20 riastrad xfree(dtv - 1); /* retreat back to DTV_MAX_INDEX */ 179 1.1 joerg dtv = tcb->tcb_dtv = new_dtv; 180 1.1 joerg SET_DTV_MAX_INDEX(dtv, _rtld_tls_max_index); 181 1.1 joerg SET_DTV_GENERATION(dtv, _rtld_tls_dtv_generation); 182 1.1 joerg } 183 1.1 joerg 184 1.1 joerg if (__predict_false(dtv[idx] == NULL)) 185 1.18 joerg dtv[idx] = _rtld_tls_module_allocate(tcb, idx); 186 1.1 joerg 187 1.5 joerg _rtld_exclusive_exit(&mask); 188 1.4 joerg 189 1.1 joerg return (uint8_t *)dtv[idx] + offset; 190 1.1 joerg } 191 1.1 joerg 192 1.20 riastrad /* 193 1.20 riastrad * _rtld_tls_initial_allocation() 194 1.20 riastrad * 195 1.20 riastrad * Allocate the TCB (thread control block) for the initial thread, 196 1.20 riastrad * once the static TLS space usage has been determined (plus some 197 1.20 riastrad * slop to allow certain special cases like Mesa to be dlopened). 198 1.20 riastrad * 199 1.20 riastrad * This must be done _after_ all initial objects (i.e., those 200 1.20 riastrad * loaded at startup, as opposed to objects dynamically loaded by 201 1.20 riastrad * dlopen) have had TLS offsets allocated if need be by 202 1.20 riastrad * _rtld_tls_offset_allocate, and have had relocations processed. 203 1.20 riastrad */ 204 1.1 joerg void 205 1.1 joerg _rtld_tls_initial_allocation(void) 206 1.1 joerg { 207 1.1 joerg struct tls_tcb *tcb; 208 1.1 joerg 209 1.1 joerg _rtld_tls_static_space = _rtld_tls_static_offset + 210 1.1 joerg RTLD_STATIC_TLS_RESERVATION; 211 1.1 joerg 212 1.1 joerg #ifndef __HAVE_TLS_VARIANT_I 213 1.1 joerg _rtld_tls_static_space = roundup2(_rtld_tls_static_space, 214 1.14 joerg alignof(max_align_t)); 215 1.1 joerg #endif 216 1.8 skrll dbg(("_rtld_tls_static_space %zu", _rtld_tls_static_space)); 217 1.1 joerg 218 1.4 joerg tcb = _rtld_tls_allocate_locked(); 219 1.3 matt #ifdef __HAVE___LWP_SETTCB 220 1.3 matt __lwp_settcb(tcb); 221 1.3 matt #else 222 1.1 joerg _lwp_setprivate(tcb); 223 1.3 matt #endif 224 1.1 joerg } 225 1.1 joerg 226 1.20 riastrad /* 227 1.20 riastrad * _rtld_tls_allocate_locked() 228 1.20 riastrad * 229 1.20 riastrad * Internal subroutine to allocate a TCB (thread control block) 230 1.20 riastrad * for the current thread. 231 1.20 riastrad * 232 1.20 riastrad * This allocates a DTV and a TCB that points to it, including 233 1.20 riastrad * static space in the TCB for the TLS of the initial objects. 234 1.20 riastrad * TLS blocks for dynamically loaded objects are allocated lazily. 235 1.20 riastrad * 236 1.20 riastrad * Caller must either be single-threaded (at startup via 237 1.20 riastrad * _rtld_tls_initial_allocation) or hold the rtld exclusive lock 238 1.20 riastrad * (via _rtld_tls_allocate). 239 1.20 riastrad */ 240 1.4 joerg static struct tls_tcb * 241 1.4 joerg _rtld_tls_allocate_locked(void) 242 1.1 joerg { 243 1.1 joerg Obj_Entry *obj; 244 1.1 joerg struct tls_tcb *tcb; 245 1.1 joerg uint8_t *p, *q; 246 1.1 joerg 247 1.1 joerg p = xcalloc(_rtld_tls_static_space + sizeof(struct tls_tcb)); 248 1.1 joerg #ifdef __HAVE_TLS_VARIANT_I 249 1.1 joerg tcb = (struct tls_tcb *)p; 250 1.1 joerg p += sizeof(struct tls_tcb); 251 1.1 joerg #else 252 1.1 joerg p += _rtld_tls_static_space; 253 1.1 joerg tcb = (struct tls_tcb *)p; 254 1.1 joerg tcb->tcb_self = tcb; 255 1.1 joerg #endif 256 1.17 riastrad dbg(("lwp %d tls tcb %p", _lwp_self(), tcb)); 257 1.21 riastrad /* 258 1.21 riastrad * "2 +" because the first element is the generation and the second 259 1.21 riastrad * one is the maximum index. 260 1.21 riastrad */ 261 1.1 joerg tcb->tcb_dtv = xcalloc(sizeof(*tcb->tcb_dtv) * (2 + _rtld_tls_max_index)); 262 1.20 riastrad ++tcb->tcb_dtv; /* advance past DTV_MAX_INDEX */ 263 1.1 joerg SET_DTV_MAX_INDEX(tcb->tcb_dtv, _rtld_tls_max_index); 264 1.1 joerg SET_DTV_GENERATION(tcb->tcb_dtv, _rtld_tls_dtv_generation); 265 1.1 joerg 266 1.1 joerg for (obj = _rtld_objlist; obj != NULL; obj = obj->next) { 267 1.18 joerg if (obj->tls_static) { 268 1.1 joerg #ifdef __HAVE_TLS_VARIANT_I 269 1.1 joerg q = p + obj->tlsoffset; 270 1.1 joerg #else 271 1.1 joerg q = p - obj->tlsoffset; 272 1.1 joerg #endif 273 1.17 riastrad dbg(("%s: [lwp %d] tls dtv %p index %zu offset %zu", 274 1.17 riastrad obj->path, _lwp_self(), 275 1.17 riastrad q, obj->tlsindex, obj->tlsoffset)); 276 1.11 joerg if (obj->tlsinitsize) 277 1.11 joerg memcpy(q, obj->tlsinit, obj->tlsinitsize); 278 1.1 joerg tcb->tcb_dtv[obj->tlsindex] = q; 279 1.1 joerg } 280 1.1 joerg } 281 1.1 joerg 282 1.1 joerg return tcb; 283 1.1 joerg } 284 1.1 joerg 285 1.20 riastrad /* 286 1.20 riastrad * _rtld_tls_allocate() 287 1.20 riastrad * 288 1.20 riastrad * Allocate a TCB (thread control block) for the current thread. 289 1.20 riastrad * 290 1.20 riastrad * Called by pthread_create for non-initial threads. (The initial 291 1.20 riastrad * thread's TCB is allocated by _rtld_tls_initial_allocation.) 292 1.20 riastrad */ 293 1.4 joerg struct tls_tcb * 294 1.4 joerg _rtld_tls_allocate(void) 295 1.4 joerg { 296 1.4 joerg struct tls_tcb *tcb; 297 1.5 joerg sigset_t mask; 298 1.4 joerg 299 1.5 joerg _rtld_exclusive_enter(&mask); 300 1.4 joerg tcb = _rtld_tls_allocate_locked(); 301 1.5 joerg _rtld_exclusive_exit(&mask); 302 1.4 joerg 303 1.4 joerg return tcb; 304 1.4 joerg } 305 1.4 joerg 306 1.20 riastrad /* 307 1.20 riastrad * _rtld_tls_free(tcb) 308 1.20 riastrad * 309 1.20 riastrad * Free a TCB allocated with _rtld_tls_allocate. 310 1.20 riastrad * 311 1.20 riastrad * Frees any TLS blocks for dynamically loaded objects that tcb's 312 1.20 riastrad * DTV points to, and frees tcb's DTV, and frees tcb. 313 1.20 riastrad */ 314 1.1 joerg void 315 1.1 joerg _rtld_tls_free(struct tls_tcb *tcb) 316 1.1 joerg { 317 1.1 joerg size_t i, max_index; 318 1.11 joerg uint8_t *p, *p_end; 319 1.5 joerg sigset_t mask; 320 1.1 joerg 321 1.5 joerg _rtld_exclusive_enter(&mask); 322 1.4 joerg 323 1.1 joerg #ifdef __HAVE_TLS_VARIANT_I 324 1.1 joerg p = (uint8_t *)tcb; 325 1.1 joerg #else 326 1.1 joerg p = (uint8_t *)tcb - _rtld_tls_static_space; 327 1.1 joerg #endif 328 1.11 joerg p_end = p + _rtld_tls_static_space; 329 1.11 joerg 330 1.11 joerg max_index = DTV_MAX_INDEX(tcb->tcb_dtv); 331 1.11 joerg for (i = 1; i <= max_index; ++i) { 332 1.11 joerg if ((uint8_t *)tcb->tcb_dtv[i] < p || 333 1.11 joerg (uint8_t *)tcb->tcb_dtv[i] >= p_end) 334 1.11 joerg xfree(tcb->tcb_dtv[i]); 335 1.11 joerg } 336 1.20 riastrad xfree(tcb->tcb_dtv - 1); /* retreat back to DTV_MAX_INDEX */ 337 1.1 joerg xfree(p); 338 1.4 joerg 339 1.5 joerg _rtld_exclusive_exit(&mask); 340 1.1 joerg } 341 1.1 joerg 342 1.20 riastrad /* 343 1.20 riastrad * _rtld_tls_module_allocate(tcb, idx) 344 1.20 riastrad * 345 1.20 riastrad * Allocate thread-local storage in the thread with the given TCB 346 1.20 riastrad * (thread control block) for the object obj whose obj->tlsindex 347 1.20 riastrad * is idx. 348 1.20 riastrad * 349 1.20 riastrad * If obj has had space in static TLS reserved (obj->tls_static), 350 1.20 riastrad * return a pointer into that. Otherwise, allocate a TLS block, 351 1.20 riastrad * mark obj as having a TLS block allocated (obj->tls_dynamic), 352 1.20 riastrad * and return it. 353 1.20 riastrad * 354 1.20 riastrad * Called by _rtld_tls_get_addr to get the thread-local storage 355 1.20 riastrad * for an object the first time around. 356 1.20 riastrad */ 357 1.18 joerg static void * 358 1.18 joerg _rtld_tls_module_allocate(struct tls_tcb *tcb, size_t idx) 359 1.1 joerg { 360 1.1 joerg Obj_Entry *obj; 361 1.1 joerg uint8_t *p; 362 1.1 joerg 363 1.1 joerg for (obj = _rtld_objlist; obj != NULL; obj = obj->next) { 364 1.1 joerg if (obj->tlsindex == idx) 365 1.1 joerg break; 366 1.1 joerg } 367 1.1 joerg if (obj == NULL) { 368 1.1 joerg _rtld_error("Module for TLS index %zu missing", idx); 369 1.1 joerg _rtld_die(); 370 1.1 joerg } 371 1.18 joerg if (obj->tls_static) { 372 1.18 joerg #ifdef __HAVE_TLS_VARIANT_I 373 1.19 joerg p = (uint8_t *)tcb + obj->tlsoffset + sizeof(struct tls_tcb); 374 1.18 joerg #else 375 1.18 joerg p = (uint8_t *)tcb - obj->tlsoffset; 376 1.18 joerg #endif 377 1.18 joerg return p; 378 1.18 joerg } 379 1.1 joerg 380 1.1 joerg p = xmalloc(obj->tlssize); 381 1.1 joerg memcpy(p, obj->tlsinit, obj->tlsinitsize); 382 1.1 joerg memset(p + obj->tlsinitsize, 0, obj->tlssize - obj->tlsinitsize); 383 1.1 joerg 384 1.18 joerg obj->tls_dynamic = 1; 385 1.18 joerg 386 1.1 joerg return p; 387 1.1 joerg } 388 1.1 joerg 389 1.20 riastrad /* 390 1.20 riastrad * _rtld_tls_offset_allocate(obj) 391 1.20 riastrad * 392 1.20 riastrad * Allocate a static thread-local storage offset for obj. 393 1.20 riastrad * 394 1.20 riastrad * Called by _rtld at startup for all initial objects. Called 395 1.20 riastrad * also by MD relocation logic, which is allowed (for Mesa) to 396 1.20 riastrad * allocate an additional 64 bytes (RTLD_STATIC_TLS_RESERVATION) 397 1.20 riastrad * of static thread-local storage in dlopened objects. 398 1.20 riastrad */ 399 1.1 joerg int 400 1.1 joerg _rtld_tls_offset_allocate(Obj_Entry *obj) 401 1.1 joerg { 402 1.1 joerg size_t offset, next_offset; 403 1.1 joerg 404 1.18 joerg if (obj->tls_dynamic) 405 1.18 joerg return -1; 406 1.18 joerg 407 1.18 joerg if (obj->tls_static) 408 1.1 joerg return 0; 409 1.1 joerg if (obj->tlssize == 0) { 410 1.1 joerg obj->tlsoffset = 0; 411 1.18 joerg obj->tls_static = 1; 412 1.1 joerg return 0; 413 1.1 joerg } 414 1.1 joerg 415 1.1 joerg #ifdef __HAVE_TLS_VARIANT_I 416 1.1 joerg offset = roundup2(_rtld_tls_static_offset, obj->tlsalign); 417 1.1 joerg next_offset = offset + obj->tlssize; 418 1.1 joerg #else 419 1.1 joerg offset = roundup2(_rtld_tls_static_offset + obj->tlssize, 420 1.1 joerg obj->tlsalign); 421 1.1 joerg next_offset = offset; 422 1.1 joerg #endif 423 1.1 joerg 424 1.1 joerg /* 425 1.1 joerg * Check if the static allocation was already done. 426 1.1 joerg * This happens if dynamically loaded modules want to use 427 1.1 joerg * static TLS space. 428 1.1 joerg * 429 1.1 joerg * XXX Keep an actual free list and callbacks for initialisation. 430 1.1 joerg */ 431 1.1 joerg if (_rtld_tls_static_space) { 432 1.1 joerg if (obj->tlsinitsize) { 433 1.1 joerg _rtld_error("%s: Use of initialized " 434 1.7 joerg "Thread Local Storage with model initial-exec " 435 1.1 joerg "and dlopen is not supported", 436 1.1 joerg obj->path); 437 1.1 joerg return -1; 438 1.1 joerg } 439 1.1 joerg if (next_offset > _rtld_tls_static_space) { 440 1.1 joerg _rtld_error("%s: No space available " 441 1.1 joerg "for static Thread Local Storage", 442 1.1 joerg obj->path); 443 1.1 joerg return -1; 444 1.1 joerg } 445 1.1 joerg } 446 1.1 joerg obj->tlsoffset = offset; 447 1.16 riastrad dbg(("%s: static tls offset 0x%zx size %zu\n", 448 1.16 riastrad obj->path, obj->tlsoffset, obj->tlssize)); 449 1.1 joerg _rtld_tls_static_offset = next_offset; 450 1.18 joerg obj->tls_static = 1; 451 1.1 joerg 452 1.1 joerg return 0; 453 1.1 joerg } 454 1.1 joerg 455 1.20 riastrad /* 456 1.20 riastrad * _rtld_tls_offset_free(obj) 457 1.20 riastrad * 458 1.20 riastrad * Free a static thread-local storage offset for obj. 459 1.20 riastrad * 460 1.20 riastrad * Called by dlclose (via _rtld_unload_object -> _rtld_obj_free). 461 1.20 riastrad * 462 1.20 riastrad * Since static thread-local storage is normally not used by 463 1.20 riastrad * dlopened objects (with the exception of Mesa), this doesn't do 464 1.20 riastrad * anything to recycle the space right now. 465 1.20 riastrad */ 466 1.1 joerg void 467 1.1 joerg _rtld_tls_offset_free(Obj_Entry *obj) 468 1.1 joerg { 469 1.1 joerg 470 1.1 joerg /* 471 1.1 joerg * XXX See above. 472 1.1 joerg */ 473 1.18 joerg obj->tls_static = 0; 474 1.1 joerg return; 475 1.1 joerg } 476 1.1 joerg 477 1.12 rin #if defined(__HAVE_COMMON___TLS_GET_ADDR) && defined(RTLD_LOADER) 478 1.2 joerg /* 479 1.20 riastrad * __tls_get_addr(tlsindex) 480 1.20 riastrad * 481 1.20 riastrad * Symbol directly called by code generated by the compiler for 482 1.20 riastrad * references thread-local storage in the general-dynamic or 483 1.20 riastrad * local-dynamic TLS models (but not initial-exec or local-exec). 484 1.20 riastrad * 485 1.20 riastrad * The argument is a pointer to 486 1.20 riastrad * 487 1.20 riastrad * struct { 488 1.20 riastrad * unsigned long int ti_module; 489 1.20 riastrad * unsigned long int ti_offset; 490 1.20 riastrad * }; 491 1.20 riastrad * 492 1.20 riastrad * as in, e.g., [ELFTLS] Sec. 3.4.3. This coincides with the 493 1.20 riastrad * type size_t[2] on all architectures that use this common 494 1.20 riastrad * __tls_get_addr definition (XXX but why do we write it as 495 1.20 riastrad * size_t[2]?). 496 1.20 riastrad * 497 1.20 riastrad * ti_module, i.e., arg[0], is the obj->tlsindex assigned at 498 1.20 riastrad * load-time by _rtld_map_object, and ti_offset, i.e., arg[1], is 499 1.20 riastrad * assigned at link-time by ld(1), possibly adjusted by 500 1.20 riastrad * TLS_DTV_OFFSET. 501 1.20 riastrad * 502 1.20 riastrad * Some architectures -- specifically IA-64 -- use a different 503 1.20 riastrad * calling convention. Some architectures -- specifically i386 504 1.20 riastrad * -- also use another entry point ___tls_get_addr (that's three 505 1.20 riastrad * leading underscores) with a different calling convention. 506 1.2 joerg */ 507 1.2 joerg void * 508 1.2 joerg __tls_get_addr(void *arg_) 509 1.2 joerg { 510 1.2 joerg size_t *arg = (size_t *)arg_; 511 1.2 joerg void **dtv; 512 1.3 matt #ifdef __HAVE___LWP_GETTCB_FAST 513 1.3 matt struct tls_tcb * const tcb = __lwp_gettcb_fast(); 514 1.3 matt #else 515 1.3 matt struct tls_tcb * const tcb = __lwp_getprivate_fast(); 516 1.3 matt #endif 517 1.3 matt size_t idx = arg[0], offset = arg[1] + TLS_DTV_OFFSET; 518 1.2 joerg 519 1.2 joerg dtv = tcb->tcb_dtv; 520 1.2 joerg 521 1.20 riastrad /* 522 1.20 riastrad * Fast path: access to an already allocated DTV entry. This 523 1.20 riastrad * checks the current limit and the entry without needing any 524 1.20 riastrad * locking. Entries are only freed on dlclose() and it is an 525 1.20 riastrad * application bug if code of the module is still running at 526 1.20 riastrad * that point. 527 1.20 riastrad */ 528 1.22 riastrad if (__predict_true(idx <= DTV_MAX_INDEX(dtv) && dtv[idx] != NULL)) 529 1.2 joerg return (uint8_t *)dtv[idx] + offset; 530 1.2 joerg 531 1.2 joerg return _rtld_tls_get_addr(tcb, idx, offset); 532 1.2 joerg } 533 1.2 joerg #endif 534 1.2 joerg 535 1.1 joerg #endif /* __HAVE_TLS_VARIANT_I || __HAVE_TLS_VARIANT_II */ 536