| 1 | /* |
| 2 | * Copyright (c) 2000-2019 Apple Inc. All rights reserved. |
| 3 | * |
| 4 | * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ |
| 5 | * |
| 6 | * This file contains Original Code and/or Modifications of Original Code |
| 7 | * as defined in and that are subject to the Apple Public Source License |
| 8 | * Version 2.0 (the 'License'). You may not use this file except in |
| 9 | * compliance with the License. The rights granted to you under the License |
| 10 | * may not be used to create, or enable the creation or redistribution of, |
| 11 | * unlawful or unlicensed copies of an Apple operating system, or to |
| 12 | * circumvent, violate, or enable the circumvention or violation of, any |
| 13 | * terms of an Apple operating system software license agreement. |
| 14 | * |
| 15 | * Please obtain a copy of the License at |
| 16 | * http://www.opensource.apple.com/apsl/ and read it before using this file. |
| 17 | * |
| 18 | * The Original Code and all software distributed under the License are |
| 19 | * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER |
| 20 | * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES, |
| 21 | * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY, |
| 22 | * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT. |
| 23 | * Please see the License for the specific language governing rights and |
| 24 | * limitations under the License. |
| 25 | * |
| 26 | * @APPLE_OSREFERENCE_LICENSE_HEADER_END@ |
| 27 | */ |
| 28 | /* |
| 29 | * @OSF_FREE_COPYRIGHT@ |
| 30 | */ |
| 31 | /* |
| 32 | * Mach Operating System |
| 33 | * Copyright (c) 1991,1990,1989,1988 Carnegie Mellon University |
| 34 | * All Rights Reserved. |
| 35 | * |
| 36 | * Permission to use, copy, modify and distribute this software and its |
| 37 | * documentation is hereby granted, provided that both the copyright |
| 38 | * notice and this permission notice appear in all copies of the |
| 39 | * software, derivative works or modified versions, and any portions |
| 40 | * thereof, and that both notices appear in supporting documentation. |
| 41 | * |
| 42 | * CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS" |
| 43 | * CONDITION. CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND FOR |
| 44 | * ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE. |
| 45 | * |
| 46 | * Carnegie Mellon requests users of this software to return to |
| 47 | * |
| 48 | * Software Distribution Coordinator or Software.Distribution@CS.CMU.EDU |
| 49 | * School of Computer Science |
| 50 | * Carnegie Mellon University |
| 51 | * Pittsburgh PA 15213-3890 |
| 52 | * |
| 53 | * any improvements or extensions that they make and grant Carnegie Mellon |
| 54 | * the rights to redistribute these changes. |
| 55 | */ |
| 56 | /* |
| 57 | */ |
| 58 | /* |
| 59 | * File: task.h |
| 60 | * Author: Avadis Tevanian, Jr. |
| 61 | * |
| 62 | * This file contains the structure definitions for tasks. |
| 63 | * |
| 64 | */ |
| 65 | /* |
| 66 | * Copyright (c) 1993 The University of Utah and |
| 67 | * the Computer Systems Laboratory (CSL). All rights reserved. |
| 68 | * |
| 69 | * Permission to use, copy, modify and distribute this software and its |
| 70 | * documentation is hereby granted, provided that both the copyright |
| 71 | * notice and this permission notice appear in all copies of the |
| 72 | * software, derivative works or modified versions, and any portions |
| 73 | * thereof, and that both notices appear in supporting documentation. |
| 74 | * |
| 75 | * THE UNIVERSITY OF UTAH AND CSL ALLOW FREE USE OF THIS SOFTWARE IN ITS "AS |
| 76 | * IS" CONDITION. THE UNIVERSITY OF UTAH AND CSL DISCLAIM ANY LIABILITY OF |
| 77 | * ANY KIND FOR ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE. |
| 78 | * |
| 79 | * CSL requests users of this software to return to csl-dist@cs.utah.edu any |
| 80 | * improvements that they make and grant CSL redistribution rights. |
| 81 | * |
| 82 | */ |
| 83 | /* |
| 84 | * NOTICE: This file was modified by McAfee Research in 2004 to introduce |
| 85 | * support for mandatory and extensible security protections. This notice |
| 86 | * is included in support of clause 2.2 (b) of the Apple Public License, |
| 87 | * Version 2.0. |
| 88 | * Copyright (c) 2005 SPARTA, Inc. |
| 89 | */ |
| 90 | |
| 91 | #ifndef _KERN_TASK_H_ |
| 92 | #define _KERN_TASK_H_ |
| 93 | |
| 94 | #include <kern/kern_types.h> |
| 95 | #include <kern/task_ref.h> |
| 96 | #include <mach/mach_types.h> |
| 97 | #include <sys/cdefs.h> |
| 98 | |
| 99 | #ifdef XNU_KERNEL_PRIVATE |
| 100 | #include <kern/btlog.h> |
| 101 | #include <kern/kern_cdata.h> |
| 102 | #include <mach/sfi_class.h> |
| 103 | #include <kern/counter.h> |
| 104 | #include <kern/cs_blobs.h> |
| 105 | #include <kern/queue.h> |
| 106 | #include <kern/recount.h> |
| 107 | #include <sys/kern_sysctl.h> |
| 108 | #if CONFIG_EXCLAVES |
| 109 | #include <mach/exclaves.h> |
| 110 | #endif /* CONFIG_EXCLAVES */ |
| 111 | #endif /* XNU_KERNEL_PRIVATE */ |
| 112 | |
| 113 | #ifdef MACH_KERNEL_PRIVATE |
| 114 | #include <mach/boolean.h> |
| 115 | #include <mach/port.h> |
| 116 | #include <mach/time_value.h> |
| 117 | #include <mach/message.h> |
| 118 | #include <mach/mach_param.h> |
| 119 | #include <mach/task_info.h> |
| 120 | #include <mach/exception_types.h> |
| 121 | #include <mach/vm_statistics.h> |
| 122 | #include <machine/task.h> |
| 123 | |
| 124 | #include <kern/cpu_data.h> |
| 125 | #include <kern/queue.h> |
| 126 | #include <kern/exception.h> |
| 127 | #include <kern/locks.h> |
| 128 | #include <security/_label.h> |
| 129 | #include <ipc/ipc_port.h> |
| 130 | |
| 131 | #include <kern/thread.h> |
| 132 | #include <mach/coalition.h> |
| 133 | #include <stdatomic.h> |
| 134 | #include <os/refcnt.h> |
| 135 | |
| 136 | #if CONFIG_DEFERRED_RECLAIM |
| 137 | typedef struct vm_deferred_reclamation_metadata_s *vm_deferred_reclamation_metadata_t; |
| 138 | #endif /* CONFIG_DEFFERED_RECLAIM */ |
| 139 | |
| 140 | struct _cpu_time_qos_stats { |
| 141 | uint64_t cpu_time_qos_default; |
| 142 | uint64_t cpu_time_qos_maintenance; |
| 143 | uint64_t cpu_time_qos_background; |
| 144 | uint64_t cpu_time_qos_utility; |
| 145 | uint64_t cpu_time_qos_legacy; |
| 146 | uint64_t cpu_time_qos_user_initiated; |
| 147 | uint64_t cpu_time_qos_user_interactive; |
| 148 | }; |
| 149 | |
| 150 | struct task_writes_counters { |
| 151 | uint64_t task_immediate_writes; |
| 152 | uint64_t task_deferred_writes; |
| 153 | uint64_t task_invalidated_writes; |
| 154 | uint64_t task_metadata_writes; |
| 155 | }; |
| 156 | |
| 157 | struct task_watchports; |
| 158 | #include <bank/bank_internal.h> |
| 159 | |
| 160 | #ifdef MACH_BSD |
| 161 | struct proc; |
| 162 | struct proc_ro; |
| 163 | #endif |
| 164 | |
| 165 | struct task { |
| 166 | /* Synchronization/destruction information */ |
| 167 | decl_lck_mtx_data(, lock); /* Task's lock */ |
| 168 | os_refcnt_t ref_count; /* Number of references to me */ |
| 169 | |
| 170 | #if DEVELOPMENT || DEBUG |
| 171 | struct os_refgrp *ref_group; |
| 172 | lck_spin_t ref_group_lock; |
| 173 | #endif /* DEVELOPMENT || DEBUG */ |
| 174 | |
| 175 | bool active; /* Task has not been terminated */ |
| 176 | bool ipc_active; /* IPC with the task ports is allowed */ |
| 177 | bool halting; /* Task is being halted */ |
| 178 | bool message_app_suspended; /* Let iokit know when pidsuspended */ |
| 179 | |
| 180 | /* Virtual timers */ |
| 181 | uint32_t vtimers; |
| 182 | uint32_t loadTag; /* dext ID used for logging identity */ |
| 183 | |
| 184 | /* Globally uniqueid to identify tasks and corpses */ |
| 185 | uint64_t task_uniqueid; |
| 186 | |
| 187 | /* Miscellaneous */ |
| 188 | vm_map_t XNU_PTRAUTH_SIGNED_PTR("task.map" ) map; /* Address space description */ |
| 189 | queue_chain_t tasks; /* global list of tasks */ |
| 190 | struct task_watchports *watchports; /* watchports passed in spawn */ |
| 191 | turnstile_inheritor_t returnwait_inheritor; /* inheritor for task_wait */ |
| 192 | |
| 193 | #if defined(CONFIG_SCHED_MULTIQ) |
| 194 | sched_group_t sched_group; |
| 195 | #endif /* defined(CONFIG_SCHED_MULTIQ) */ |
| 196 | |
| 197 | /* Threads in this task */ |
| 198 | queue_head_t threads; |
| 199 | struct restartable_ranges *t_rr_ranges; |
| 200 | |
| 201 | processor_set_t pset_hint; |
| 202 | struct affinity_space *affinity_space; |
| 203 | |
| 204 | int thread_count; |
| 205 | uint32_t active_thread_count; |
| 206 | int suspend_count; /* Internal scheduling only */ |
| 207 | #ifdef CONFIG_TASK_SUSPEND_STATS |
| 208 | struct task_suspend_stats_s t_suspend_stats; /* suspension statistics for this task */ |
| 209 | task_suspend_source_array_t t_suspend_sources; /* array of suspender debug info for this task */ |
| 210 | #endif /* CONFIG_TASK_SUSPEND_STATS */ |
| 211 | |
| 212 | /* User-visible scheduling information */ |
| 213 | integer_t user_stop_count; /* outstanding stops */ |
| 214 | integer_t legacy_stop_count; /* outstanding legacy stops */ |
| 215 | |
| 216 | int16_t priority; /* base priority for threads */ |
| 217 | int16_t max_priority; /* maximum priority for threads */ |
| 218 | |
| 219 | integer_t importance; /* priority offset (BSD 'nice' value) */ |
| 220 | |
| 221 | #define task_is_immovable(task) \ |
| 222 | !!(task_get_control_port_options(task) & TASK_CONTROL_PORT_IMMOVABLE) |
| 223 | #define task_is_pinned(task) \ |
| 224 | !!(task_get_control_port_options(task) & TASK_CONTROL_PORT_PINNED) |
| 225 | |
| 226 | /* Statistics */ |
| 227 | uint64_t total_runnable_time; |
| 228 | |
| 229 | struct recount_task tk_recount; |
| 230 | |
| 231 | /* IPC structures */ |
| 232 | decl_lck_mtx_data(, itk_lock_data); |
| 233 | /* |
| 234 | * Different flavors of task port. |
| 235 | * These flavors TASK_FLAVOR_* are defined in mach_types.h |
| 236 | */ |
| 237 | struct ipc_port * XNU_PTRAUTH_SIGNED_PTR("task.itk_task_ports" ) itk_task_ports[TASK_SELF_PORT_COUNT]; |
| 238 | struct ipc_port * XNU_PTRAUTH_SIGNED_PTR("task.itk_settable_self" ) itk_settable_self; /* a send right */ |
| 239 | struct ipc_port * XNU_PTRAUTH_SIGNED_PTR("task.itk_self" ) itk_self; /* immovable/pinned task port, does not hold right */ |
| 240 | struct exception_action exc_actions[EXC_TYPES_COUNT]; |
| 241 | /* a send right each valid element */ |
| 242 | struct ipc_port * XNU_PTRAUTH_SIGNED_PTR("task.itk_host" ) itk_host; /* a send right */ |
| 243 | struct ipc_port * XNU_PTRAUTH_SIGNED_PTR("task.itk_bootstrap" ) itk_bootstrap; /* a send right */ |
| 244 | struct ipc_port * XNU_PTRAUTH_SIGNED_PTR("task.itk_debug_control" ) itk_debug_control; /* send right for debugmode communications */ |
| 245 | struct ipc_port * XNU_PTRAUTH_SIGNED_PTR("task.itk_task_access" ) itk_task_access; /* and another send right */ |
| 246 | struct ipc_port * XNU_PTRAUTH_SIGNED_PTR("task.itk_resume" ) itk_resume; /* a receive right to resume this task */ |
| 247 | struct ipc_port * XNU_PTRAUTH_SIGNED_PTR("task.itk_registered" ) itk_registered[TASK_PORT_REGISTER_MAX]; |
| 248 | /* all send rights */ |
| 249 | ipc_port_t * XNU_PTRAUTH_SIGNED_PTR("task.itk_dyld_notify" ) itk_dyld_notify; /* lazy send rights array of size DYLD_MAX_PROCESS_INFO_NOTIFY_COUNT */ |
| 250 | #if CONFIG_PROC_RESOURCE_LIMITS |
| 251 | struct ipc_port * XNU_PTRAUTH_SIGNED_PTR("task.itk_resource_notify" ) itk_resource_notify; /* a send right to the resource notify port */ |
| 252 | #endif /* CONFIG_PROC_RESOURCE_LIMITS */ |
| 253 | struct ipc_space * XNU_PTRAUTH_SIGNED_PTR("task.itk_space" ) itk_space; |
| 254 | |
| 255 | ledger_t ledger; |
| 256 | /* Synchronizer ownership information */ |
| 257 | queue_head_t semaphore_list; /* list of owned semaphores */ |
| 258 | int semaphores_owned; /* number of semaphores owned */ |
| 259 | |
| 260 | unsigned int priv_flags; /* privilege resource flags */ |
| 261 | #define VM_BACKING_STORE_PRIV 0x1 |
| 262 | |
| 263 | MACHINE_TASK |
| 264 | |
| 265 | counter_t faults; /* faults counter */ |
| 266 | counter_t pageins; /* pageins counter */ |
| 267 | counter_t cow_faults; /* copy on write fault counter */ |
| 268 | counter_t messages_sent; /* messages sent counter */ |
| 269 | counter_t messages_received; /* messages received counter */ |
| 270 | uint32_t decompressions; /* decompression counter */ |
| 271 | uint32_t syscalls_mach; /* mach system call counter */ |
| 272 | uint32_t syscalls_unix; /* unix system call counter */ |
| 273 | uint32_t c_switch; /* total context switches */ |
| 274 | uint32_t p_switch; /* total processor switches */ |
| 275 | uint32_t ps_switch; /* total pset switches */ |
| 276 | |
| 277 | #ifdef MACH_BSD |
| 278 | struct proc_ro * bsd_info_ro; |
| 279 | #endif |
| 280 | kcdata_descriptor_t corpse_info; |
| 281 | uint64_t crashed_thread_id; |
| 282 | queue_chain_t corpse_tasks; |
| 283 | #ifdef CONFIG_MACF |
| 284 | struct label * crash_label; |
| 285 | #endif |
| 286 | volatile uint32_t t_flags; /* general-purpose task flags protected by task_lock (TL) */ |
| 287 | #define TF_NONE 0 |
| 288 | #define TF_64B_ADDR 0x00000001 /* task has 64-bit addressing */ |
| 289 | #define TF_64B_DATA 0x00000002 /* task has 64-bit data registers */ |
| 290 | #define TF_CPUMON_WARNING 0x00000004 /* task has at least one thread in CPU usage warning zone */ |
| 291 | #define TF_WAKEMON_WARNING 0x00000008 /* task is in wakeups monitor warning zone */ |
| 292 | #define TF_TELEMETRY (TF_CPUMON_WARNING | TF_WAKEMON_WARNING) /* task is a telemetry participant */ |
| 293 | #define TF_GPU_DENIED 0x00000010 /* task is not allowed to access the GPU */ |
| 294 | #define TF_PENDING_CORPSE 0x00000040 /* task corpse has not been reported yet */ |
| 295 | #define TF_CORPSE_FORK 0x00000080 /* task is a forked corpse */ |
| 296 | #define TF_CA_CLIENT_WI 0x00000800 /* task has CA_CLIENT work interval */ |
| 297 | #define TF_DARKWAKE_MODE 0x00001000 /* task is in darkwake mode */ |
| 298 | #define TF_NO_SMT 0x00002000 /* task threads must not be paired with SMT threads */ |
| 299 | #define TF_SYS_VERSION_COMPAT 0x00008000 /* shim task accesses to OS version data (macOS - app compatibility) */ |
| 300 | #define TF_TECS 0x00020000 /* task threads must enable CPU security */ |
| 301 | #if defined(__x86_64__) |
| 302 | #define TF_INSN_COPY_OPTOUT 0x00040000 /* task threads opt out of unhandled-fault instruction stream collection */ |
| 303 | #endif |
| 304 | #define TF_COALITION_MEMBER 0x00080000 /* task is a member of a coalition */ |
| 305 | #define TF_NO_CORPSE_FORKING 0x00100000 /* do not fork a corpse for this task */ |
| 306 | #define TF_USE_PSET_HINT_CLUSTER_TYPE 0x00200000 /* bind task to task->pset_hint->pset_cluster_type */ |
| 307 | #define TF_DYLD_ALL_IMAGE_FINAL 0x00400000 /* all_image_info_addr can no longer be changed */ |
| 308 | #define TF_HASPROC 0x00800000 /* task points to a proc */ |
| 309 | #define TF_HAS_REPLY_PORT_TELEMETRY 0x10000000 /* Rate limit telemetry for reply port security semantics violations rdar://100244531 */ |
| 310 | #define TF_HAS_EXCEPTION_TELEMETRY 0x20000000 /* Rate limit telemetry for exception identity violations rdar://100729339 */ |
| 311 | #define TF_GAME_MODE 0x40000000 /* Set the game mode bit for CLPC */ |
| 312 | |
| 313 | /* |
| 314 | * WARNING: These TF_ and TFRO_ flags are NOT automatically inherited by a child of fork |
| 315 | * If you believe something should be inherited, you must manually inherit the flags in `task_create_internal` |
| 316 | */ |
| 317 | |
| 318 | /* |
| 319 | * RO-protected flags: |
| 320 | */ |
| 321 | #define TFRO_CORPSE 0x00000020 /* task is a corpse */ |
| 322 | #define TFRO_HARDENED 0x00000100 /* task is a hardened runtime binary */ |
| 323 | #if XNU_TARGET_OS_OSX |
| 324 | #define TFRO_MACH_HARDENING_OPT_OUT 0x00000200 /* task might load third party plugins on macOS and should be opted out of mach hardening */ |
| 325 | #endif /* XNU_TARGET_OS_OSX */ |
| 326 | #define TFRO_PLATFORM 0x00000400 /* task is a platform binary */ |
| 327 | #define TFRO_FILTER_MSG 0x00004000 /* task calls into message filter callback before sending a message */ |
| 328 | #define TFRO_PAC_EXC_FATAL 0x00010000 /* task is marked a corpse if a PAC exception occurs */ |
| 329 | #define TFRO_JIT_EXC_FATAL 0x00020000 /* kill the task on access violations from privileged JIT code */ |
| 330 | #define TFRO_PAC_ENFORCE_USER_STATE 0x01000000 /* Enforce user and kernel signed thread state */ |
| 331 | #if CONFIG_EXCLAVES |
| 332 | #define TFRO_HAS_KD_ACCESS 0x02000000 /* Access to the kernel exclave resource domain */ |
| 333 | #endif /* CONFIG_EXCLAVES */ |
| 334 | |
| 335 | /* |
| 336 | * Task is running within a 64-bit address space. |
| 337 | */ |
| 338 | #define task_has_64Bit_addr(task) \ |
| 339 | (((task)->t_flags & TF_64B_ADDR) != 0) |
| 340 | #define task_set_64Bit_addr(task) \ |
| 341 | ((task)->t_flags |= TF_64B_ADDR) |
| 342 | #define task_clear_64Bit_addr(task) \ |
| 343 | ((task)->t_flags &= ~TF_64B_ADDR) |
| 344 | |
| 345 | /* |
| 346 | * Task is using 64-bit machine state. |
| 347 | */ |
| 348 | #define task_has_64Bit_data(task) \ |
| 349 | (((task)->t_flags & TF_64B_DATA) != 0) |
| 350 | #define task_set_64Bit_data(task) \ |
| 351 | ((task)->t_flags |= TF_64B_DATA) |
| 352 | #define task_clear_64Bit_data(task) \ |
| 353 | ((task)->t_flags &= ~TF_64B_DATA) |
| 354 | |
| 355 | #define task_corpse_pending_report(task) \ |
| 356 | (((task)->t_flags & TF_PENDING_CORPSE) != 0) |
| 357 | |
| 358 | #define task_set_corpse_pending_report(task) \ |
| 359 | ((task)->t_flags |= TF_PENDING_CORPSE) |
| 360 | |
| 361 | #define task_clear_corpse_pending_report(task) \ |
| 362 | ((task)->t_flags &= ~TF_PENDING_CORPSE) |
| 363 | |
| 364 | #define task_is_a_corpse_fork(task) \ |
| 365 | (((task)->t_flags & TF_CORPSE_FORK) != 0) |
| 366 | |
| 367 | #define task_set_coalition_member(task) \ |
| 368 | ((task)->t_flags |= TF_COALITION_MEMBER) |
| 369 | |
| 370 | #define task_clear_coalition_member(task) \ |
| 371 | ((task)->t_flags &= ~TF_COALITION_MEMBER) |
| 372 | |
| 373 | #define task_is_coalition_member(task) \ |
| 374 | (((task)->t_flags & TF_COALITION_MEMBER) != 0) |
| 375 | |
| 376 | #define task_has_proc(task) \ |
| 377 | (((task)->t_flags & TF_HASPROC) != 0) |
| 378 | |
| 379 | #define task_set_has_proc(task) \ |
| 380 | ((task)->t_flags |= TF_HASPROC) |
| 381 | |
| 382 | #define task_clear_has_proc(task) \ |
| 383 | ((task)->t_flags &= ~TF_HASPROC) |
| 384 | |
| 385 | #define task_has_reply_port_telemetry(task) \ |
| 386 | (((task)->t_flags & TF_HAS_REPLY_PORT_TELEMETRY) != 0) |
| 387 | |
| 388 | #define task_set_reply_port_telemetry(task) \ |
| 389 | ((task)->t_flags |= TF_HAS_REPLY_PORT_TELEMETRY) |
| 390 | |
| 391 | #define task_has_exception_telemetry(task) \ |
| 392 | (((task)->t_flags & TF_HAS_EXCEPTION_TELEMETRY) != 0) |
| 393 | |
| 394 | #define task_set_exception_telemetry(task) \ |
| 395 | ((task)->t_flags |= TF_HAS_EXCEPTION_TELEMETRY) |
| 396 | |
| 397 | uint32_t t_procflags; /* general-purpose task flags protected by proc_lock (PL) */ |
| 398 | #define TPF_NONE 0 |
| 399 | #define TPF_DID_EXEC 0x00000001 /* task has been execed to a new task */ |
| 400 | #define TPF_EXEC_COPY 0x00000002 /* task is the new copy of an exec */ |
| 401 | |
| 402 | #define task_did_exec_internal(task) \ |
| 403 | (((task)->t_procflags & TPF_DID_EXEC) != 0) |
| 404 | |
| 405 | #define task_is_exec_copy_internal(task) \ |
| 406 | (((task)->t_procflags & TPF_EXEC_COPY) != 0) |
| 407 | |
| 408 | mach_vm_address_t all_image_info_addr; /* dyld __all_image_info */ |
| 409 | mach_vm_size_t all_image_info_size; /* section location and size */ |
| 410 | |
| 411 | #if CONFIG_CPU_COUNTERS |
| 412 | #define TASK_KPC_FORCED_ALL_CTRS 0x2 /* Bit in "t_kpc" signifying this task forced all counters */ |
| 413 | uint32_t t_kpc; /* kpc flags */ |
| 414 | #endif /* CONFIG_CPU_COUNTERS */ |
| 415 | |
| 416 | bool pidsuspended; /* pid_suspend called; no threads can execute */ |
| 417 | bool frozen; /* frozen; private resident pages committed to swap */ |
| 418 | bool changing_freeze_state; /* in the process of freezing or thawing */ |
| 419 | bool is_large_corpse; |
| 420 | uint16_t policy_ru_cpu :4, |
| 421 | policy_ru_cpu_ext :4, |
| 422 | applied_ru_cpu :4, |
| 423 | applied_ru_cpu_ext :4; |
| 424 | uint8_t rusage_cpu_flags; |
| 425 | uint8_t rusage_cpu_percentage; /* Task-wide CPU limit percentage */ |
| 426 | uint8_t rusage_cpu_perthr_percentage; /* Per-thread CPU limit percentage */ |
| 427 | #if MACH_ASSERT |
| 428 | int8_t suspends_outstanding; /* suspends this task performed in excess of resumes */ |
| 429 | #endif |
| 430 | uint8_t t_returnwaitflags; |
| 431 | #define TWF_NONE 0 |
| 432 | #define TRW_LRETURNWAIT 0x01 /* task is waiting for fork/posix_spawn/exec to complete */ |
| 433 | #define TRW_LRETURNWAITER 0x02 /* task is waiting for TRW_LRETURNWAIT to get cleared */ |
| 434 | #define TRW_LEXEC_COMPLETE 0x04 /* thread should call exec complete */ |
| 435 | |
| 436 | #if CONFIG_EXCLAVES |
| 437 | uint8_t t_exclave_state; |
| 438 | #define TES_NONE 0 |
| 439 | #define TES_CONCLAVE_TAINTED 0x01 /* Task has talked to conclave, xnu has tainted the process */ |
| 440 | #define TES_CONCLAVE_UNTAINTABLE 0x02 /* Task can not be tainted by xnu when it talks to conclave */ |
| 441 | #endif /* CONFIG_EXCLAVES */ |
| 442 | |
| 443 | #if __has_feature(ptrauth_calls) |
| 444 | bool shared_region_auth_remapped; /* authenticated sections ready for use */ |
| 445 | char *shared_region_id; /* determines which ptr auth key to use */ |
| 446 | #endif /* __has_feature(ptrauth_calls) */ |
| 447 | struct vm_shared_region *shared_region; |
| 448 | |
| 449 | uint64_t rusage_cpu_interval; /* Task-wide CPU limit interval */ |
| 450 | uint64_t rusage_cpu_perthr_interval; /* Per-thread CPU limit interval */ |
| 451 | uint64_t rusage_cpu_deadline; |
| 452 | thread_call_t rusage_cpu_callt; |
| 453 | #if CONFIG_TASKWATCH |
| 454 | queue_head_t task_watchers; /* app state watcher threads */ |
| 455 | int num_taskwatchers; |
| 456 | int watchapplying; |
| 457 | #endif /* CONFIG_TASKWATCH */ |
| 458 | |
| 459 | struct bank_task *bank_context; /* pointer to per task bank structure */ |
| 460 | |
| 461 | #if IMPORTANCE_INHERITANCE |
| 462 | struct ipc_importance_task *task_imp_base; /* Base of IPC importance chain */ |
| 463 | #endif /* IMPORTANCE_INHERITANCE */ |
| 464 | |
| 465 | vm_extmod_statistics_data_t extmod_statistics; |
| 466 | |
| 467 | struct task_requested_policy requested_policy; |
| 468 | struct task_effective_policy effective_policy; |
| 469 | |
| 470 | /* |
| 471 | * Can be merged with imp_donor bits, once the IMPORTANCE_INHERITANCE macro goes away. |
| 472 | */ |
| 473 | uint32_t low_mem_notified_warn :1, /* warning low memory notification is sent to the task */ |
| 474 | low_mem_notified_critical :1, /* critical low memory notification is sent to the task */ |
| 475 | purged_memory_warn :1, /* purgeable memory of the task is purged for warning level pressure */ |
| 476 | purged_memory_critical :1, /* purgeable memory of the task is purged for critical level pressure */ |
| 477 | low_mem_privileged_listener :1, /* if set, task would like to know about pressure changes before other tasks on the system */ |
| 478 | mem_notify_reserved :27; /* reserved for future use */ |
| 479 | |
| 480 | uint32_t memlimit_is_active :1, /* if set, use active attributes, otherwise use inactive attributes */ |
| 481 | memlimit_is_fatal :1, /* if set, exceeding current memlimit will prove fatal to the task */ |
| 482 | memlimit_active_exc_resource :1, /* if set, suppress exc_resource exception when task exceeds active memory limit */ |
| 483 | memlimit_inactive_exc_resource :1, /* if set, suppress exc_resource exception when task exceeds inactive memory limit */ |
| 484 | memlimit_attrs_reserved :28; /* reserved for future use */ |
| 485 | |
| 486 | io_stat_info_t task_io_stats; |
| 487 | |
| 488 | struct task_writes_counters task_writes_counters_internal; |
| 489 | struct task_writes_counters task_writes_counters_external; |
| 490 | |
| 491 | /* |
| 492 | * The cpu_time_qos_stats fields are protected by the task lock |
| 493 | */ |
| 494 | struct _cpu_time_qos_stats cpu_time_eqos_stats; |
| 495 | struct _cpu_time_qos_stats cpu_time_rqos_stats; |
| 496 | |
| 497 | /* Statistics accumulated for terminated threads from this task */ |
| 498 | uint32_t task_timer_wakeups_bin_1; |
| 499 | uint32_t task_timer_wakeups_bin_2; |
| 500 | uint64_t task_gpu_ns; |
| 501 | |
| 502 | uint8_t task_can_transfer_memory_ownership; |
| 503 | #if DEVELOPMENT || DEBUG |
| 504 | uint8_t task_no_footprint_for_debug; |
| 505 | #endif |
| 506 | uint8_t task_objects_disowning; |
| 507 | uint8_t task_objects_disowned; |
| 508 | /* # of purgeable volatile VM objects owned by this task: */ |
| 509 | int task_volatile_objects; |
| 510 | /* # of purgeable but not volatile VM objects owned by this task: */ |
| 511 | int task_nonvolatile_objects; |
| 512 | int task_owned_objects; |
| 513 | queue_head_t task_objq; |
| 514 | decl_lck_mtx_data(, task_objq_lock); /* protects "task_objq" */ |
| 515 | |
| 516 | unsigned int task_thread_limit:16; |
| 517 | #if __arm64__ |
| 518 | unsigned int :1; |
| 519 | unsigned int :1; |
| 520 | unsigned int :1; |
| 521 | #endif /* __arm64__ */ |
| 522 | unsigned int :1; |
| 523 | unsigned int task_has_crossed_thread_limit:1; |
| 524 | unsigned int task_rr_in_flight:1; /* a t_rr_synchronzie() is in flight */ |
| 525 | /* |
| 526 | * A task's coalition set is "adopted" in task_create_internal |
| 527 | * and unset in task_deallocate_internal, so each array member |
| 528 | * can be referenced without the task lock. |
| 529 | * Note: these fields are protected by coalition->lock, |
| 530 | * not the task lock. |
| 531 | */ |
| 532 | coalition_t coalition[COALITION_NUM_TYPES]; |
| 533 | queue_chain_t task_coalition[COALITION_NUM_TYPES]; |
| 534 | uint64_t dispatchqueue_offset; |
| 535 | |
| 536 | #if DEVELOPMENT || DEBUG |
| 537 | boolean_t task_unnested; |
| 538 | int task_disconnected_count; |
| 539 | #endif |
| 540 | |
| 541 | #if HYPERVISOR |
| 542 | void * XNU_PTRAUTH_SIGNED_PTR("task.hv_task_target" ) hv_task_target; /* hypervisor virtual machine object associated with this task */ |
| 543 | #endif /* HYPERVISOR */ |
| 544 | |
| 545 | #if CONFIG_SECLUDED_MEMORY |
| 546 | uint8_t task_can_use_secluded_mem; |
| 547 | uint8_t task_could_use_secluded_mem; |
| 548 | uint8_t task_could_also_use_secluded_mem; |
| 549 | uint8_t task_suppressed_secluded; |
| 550 | #endif /* CONFIG_SECLUDED_MEMORY */ |
| 551 | |
| 552 | task_exc_guard_behavior_t task_exc_guard; |
| 553 | mach_vm_address_t ; |
| 554 | |
| 555 | queue_head_t io_user_clients; |
| 556 | |
| 557 | #if CONFIG_FREEZE |
| 558 | queue_head_t task_frozen_cseg_q; /* queue of csegs frozen to NAND */ |
| 559 | #endif /* CONFIG_FREEZE */ |
| 560 | boolean_t donates_own_pages; /* pages land on the special Q (only swappable pages on iPadOS, early swap on macOS) */ |
| 561 | uint32_t task_shared_region_slide; /* cached here to avoid locking during telemetry */ |
| 562 | #if CONFIG_PHYS_WRITE_ACCT |
| 563 | uint64_t task_fs_metadata_writes; |
| 564 | #endif /* CONFIG_PHYS_WRITE_ACCT */ |
| 565 | uuid_t task_shared_region_uuid; |
| 566 | #if CONFIG_MEMORYSTATUS |
| 567 | uint64_t memstat_dirty_start; /* last abstime transition into the dirty band or last call to task_ledger_settle_dirty_time while dirty */ |
| 568 | #endif /* CONFIG_MEMORYSTATUS */ |
| 569 | vmobject_list_output_t corpse_vmobject_list; |
| 570 | uint64_t corpse_vmobject_list_size; |
| 571 | #if CONFIG_DEFERRED_RECLAIM |
| 572 | vm_deferred_reclamation_metadata_t deferred_reclamation_metadata; /* Protected by the task lock */ |
| 573 | #endif /* CONFIG_DEFERRED_RECLAIM */ |
| 574 | |
| 575 | #if CONFIG_EXCLAVES |
| 576 | void * XNU_PTRAUTH_SIGNED_PTR("task.conclave" ) conclave; |
| 577 | void * XNU_PTRAUTH_SIGNED_PTR("task.exclave_crash_info" ) exclave_crash_info; |
| 578 | uint32_t exclave_crash_info_length; |
| 579 | #endif /* CONFIG_EXCLAVES */ |
| 580 | }; |
| 581 | |
| 582 | ZONE_DECLARE_ID(ZONE_ID_PROC_TASK, void *); |
| 583 | extern zone_t proc_task_zone; |
| 584 | |
| 585 | extern task_control_port_options_t task_get_control_port_options(task_t task); |
| 586 | extern void task_set_control_port_options(task_t task, task_control_port_options_t opts); |
| 587 | |
| 588 | /* |
| 589 | * EXC_GUARD default delivery behavior for optional Mach port and VM guards. |
| 590 | * Applied to new tasks at creation time. |
| 591 | */ |
| 592 | extern task_exc_guard_behavior_t task_exc_guard_default; |
| 593 | extern size_t proc_and_task_size; |
| 594 | extern void *get_bsdtask_info(task_t t); |
| 595 | extern void *task_get_proc_raw(task_t task); |
| 596 | static inline void |
| 597 | task_require(struct task *task) |
| 598 | { |
| 599 | zone_id_require(zone_id: ZONE_ID_PROC_TASK, elem_size: proc_and_task_size, addr: task_get_proc_raw(task)); |
| 600 | } |
| 601 | |
| 602 | /* |
| 603 | * task_lock() and task_unlock() need to be callable from the `bsd/` tree of |
| 604 | * XNU and are therefore promoted to full functions instead of macros so that |
| 605 | * they can be linked against. |
| 606 | * |
| 607 | * We provide `extern` declarations here for consumers of `task.h` in `osfmk/`, |
| 608 | * then separately provide `inline` definitions in `task.c`. Together with the |
| 609 | * `BUILD_LTO=1` build argument, this guarantees these functions are always |
| 610 | * inlined regardless of whether called from the `osfmk/` tree or `bsd/` tree. |
| 611 | */ |
| 612 | extern void task_lock(task_t); |
| 613 | extern void task_unlock(task_t); |
| 614 | |
| 615 | #define task_lock_assert_owned(task) LCK_MTX_ASSERT(&(task)->lock, LCK_MTX_ASSERT_OWNED) |
| 616 | #define task_lock_try(task) lck_mtx_try_lock(&(task)->lock) |
| 617 | |
| 618 | #define task_objq_lock_init(task) lck_mtx_init(&(task)->task_objq_lock, &vm_object_lck_grp, &vm_object_lck_attr) |
| 619 | #define task_objq_lock_destroy(task) lck_mtx_destroy(&(task)->task_objq_lock, &vm_object_lck_grp) |
| 620 | #define task_objq_lock(task) lck_mtx_lock(&(task)->task_objq_lock) |
| 621 | #define task_objq_lock_assert_owned(task) LCK_MTX_ASSERT(&(task)->task_objq_lock, LCK_MTX_ASSERT_OWNED) |
| 622 | #define task_objq_lock_try(task) lck_mtx_try_lock(&(task)->task_objq_lock) |
| 623 | #define task_objq_unlock(task) lck_mtx_unlock(&(task)->task_objq_lock) |
| 624 | |
| 625 | #define itk_lock_init(task) lck_mtx_init(&(task)->itk_lock_data, &ipc_lck_grp, &ipc_lck_attr) |
| 626 | #define itk_lock_destroy(task) lck_mtx_destroy(&(task)->itk_lock_data, &ipc_lck_grp) |
| 627 | #define itk_lock(task) lck_mtx_lock(&(task)->itk_lock_data) |
| 628 | #define itk_unlock(task) lck_mtx_unlock(&(task)->itk_lock_data) |
| 629 | |
| 630 | /* task clear return wait flags */ |
| 631 | #define TCRW_CLEAR_INITIAL_WAIT 0x1 |
| 632 | #define TCRW_CLEAR_FINAL_WAIT 0x2 |
| 633 | #define TCRW_CLEAR_EXEC_COMPLETE 0x4 |
| 634 | #define TCRW_CLEAR_ALL_WAIT (TCRW_CLEAR_INITIAL_WAIT | TCRW_CLEAR_FINAL_WAIT) |
| 635 | |
| 636 | /* Initialize task module */ |
| 637 | extern void task_init(void); |
| 638 | |
| 639 | /* coalition_init() calls this to initialize ledgers before task_init() */ |
| 640 | extern void init_task_ledgers(void); |
| 641 | |
| 642 | extern task_t current_task(void) __pure2; |
| 643 | |
| 644 | extern bool task_is_driver(task_t task); |
| 645 | extern uint32_t task_ro_flags_get(task_t task); |
| 646 | extern void task_ro_flags_set(task_t task, uint32_t flags); |
| 647 | extern void task_ro_flags_clear(task_t task, uint32_t flags); |
| 648 | |
| 649 | extern lck_attr_t task_lck_attr; |
| 650 | extern lck_grp_t task_lck_grp; |
| 651 | |
| 652 | struct task_watchport_elem { |
| 653 | task_t twe_task; |
| 654 | ipc_port_t twe_port; /* (Space lock) */ |
| 655 | ipc_port_t XNU_PTRAUTH_SIGNED_PTR("twe_pdrequest" ) twe_pdrequest; |
| 656 | }; |
| 657 | |
| 658 | struct task_watchports { |
| 659 | os_refcnt_t tw_refcount; /* (Space lock) */ |
| 660 | task_t tw_task; /* (Space lock) & tw_refcount == 0 */ |
| 661 | thread_t tw_thread; /* (Space lock) & tw_refcount == 0 */ |
| 662 | uint32_t tw_elem_array_count; /* (Space lock) */ |
| 663 | struct task_watchport_elem tw_elem[]; /* (Space lock) & (Portlock) & (mq lock) */ |
| 664 | }; |
| 665 | |
| 666 | #define task_watchports_retain(x) (os_ref_retain(&(x)->tw_refcount)) |
| 667 | #define task_watchports_release(x) (os_ref_release(&(x)->tw_refcount)) |
| 668 | |
| 669 | #define task_watchport_elem_init(elem, task, port) \ |
| 670 | do { \ |
| 671 | (elem)->twe_task = (task); \ |
| 672 | (elem)->twe_port = (port); \ |
| 673 | (elem)->twe_pdrequest = IP_NULL; \ |
| 674 | } while(0) |
| 675 | |
| 676 | #define task_watchport_elem_clear(elem) task_watchport_elem_init((elem), NULL, NULL) |
| 677 | |
| 678 | extern void |
| 679 | task_add_turnstile_watchports( |
| 680 | task_t task, |
| 681 | thread_t thread, |
| 682 | ipc_port_t *portwatch_ports, |
| 683 | uint32_t portwatch_count); |
| 684 | |
| 685 | extern void |
| 686 | task_watchport_elem_deallocate( |
| 687 | struct task_watchport_elem *watchport_elem); |
| 688 | |
| 689 | extern boolean_t |
| 690 | task_has_watchports(task_t task); |
| 691 | |
| 692 | void |
| 693 | task_dyld_process_info_update_helper( |
| 694 | task_t task, |
| 695 | size_t active_count, |
| 696 | vm_map_address_t magic_addr, |
| 697 | ipc_port_t *release_ports, |
| 698 | size_t release_count); |
| 699 | |
| 700 | extern kern_return_t |
| 701 | task_suspend2_mig( |
| 702 | task_t task, |
| 703 | task_suspension_token_t *suspend_token); |
| 704 | |
| 705 | extern kern_return_t |
| 706 | task_suspend2_external( |
| 707 | task_t task, |
| 708 | task_suspension_token_t *suspend_token); |
| 709 | |
| 710 | extern kern_return_t |
| 711 | task_resume2_mig( |
| 712 | task_suspension_token_t suspend_token); |
| 713 | |
| 714 | extern kern_return_t |
| 715 | task_resume2_external( |
| 716 | task_suspension_token_t suspend_token); |
| 717 | |
| 718 | extern void |
| 719 | task_suspension_token_deallocate_grp( |
| 720 | task_suspension_token_t suspend_token, |
| 721 | task_grp_t grp); |
| 722 | |
| 723 | extern ipc_port_t |
| 724 | convert_task_to_port_with_flavor( |
| 725 | task_t task, |
| 726 | mach_task_flavor_t flavor, |
| 727 | task_grp_t grp); |
| 728 | |
| 729 | extern task_t current_task_early(void) __pure2; |
| 730 | |
| 731 | #else /* MACH_KERNEL_PRIVATE */ |
| 732 | |
| 733 | __BEGIN_DECLS |
| 734 | |
| 735 | extern task_t current_task(void) __pure2; |
| 736 | |
| 737 | extern bool task_is_driver(task_t task); |
| 738 | |
| 739 | #define TF_NONE 0 |
| 740 | |
| 741 | #define TWF_NONE 0 |
| 742 | #define TRW_LRETURNWAIT 0x01 /* task is waiting for fork/posix_spawn/exec to complete */ |
| 743 | #define TRW_LRETURNWAITER 0x02 /* task is waiting for TRW_LRETURNWAIT to get cleared */ |
| 744 | #define TRW_LEXEC_COMPLETE 0x04 /* thread should call exec complete */ |
| 745 | |
| 746 | /* task clear return wait flags */ |
| 747 | #define TCRW_CLEAR_INITIAL_WAIT 0x1 |
| 748 | #define TCRW_CLEAR_FINAL_WAIT 0x2 |
| 749 | #define TCRW_CLEAR_EXEC_COMPLETE 0x4 |
| 750 | #define TCRW_CLEAR_ALL_WAIT (TCRW_CLEAR_INITIAL_WAIT | TCRW_CLEAR_FINAL_WAIT) |
| 751 | |
| 752 | |
| 753 | #define TPF_NONE 0 |
| 754 | #define TPF_EXEC_COPY 0x00000002 /* task is the new copy of an exec */ |
| 755 | |
| 756 | |
| 757 | __END_DECLS |
| 758 | |
| 759 | #endif /* MACH_KERNEL_PRIVATE */ |
| 760 | |
| 761 | __BEGIN_DECLS |
| 762 | |
| 763 | #ifdef KERNEL_PRIVATE |
| 764 | extern boolean_t task_is_app_suspended(task_t task); |
| 765 | extern bool task_is_exotic(task_t task); |
| 766 | extern bool task_is_alien(task_t task); |
| 767 | #endif /* KERNEL_PRIVATE */ |
| 768 | |
| 769 | #ifdef XNU_KERNEL_PRIVATE |
| 770 | |
| 771 | /* Hold all threads in a task, Wait for task to stop running, just to get off CPU */ |
| 772 | extern kern_return_t task_hold_and_wait( |
| 773 | task_t task); |
| 774 | |
| 775 | /* Release hold on all threads in a task */ |
| 776 | extern kern_return_t task_release( |
| 777 | task_t task); |
| 778 | |
| 779 | /* Suspend/resume a task where the kernel owns the suspend count */ |
| 780 | extern kern_return_t task_suspend_internal_locked( task_t task); |
| 781 | extern kern_return_t task_suspend_internal( task_t task); |
| 782 | extern kern_return_t task_resume_internal_locked( task_t task); |
| 783 | extern kern_return_t task_resume_internal( task_t task); |
| 784 | |
| 785 | /* Suspends a task by placing a hold on its threads */ |
| 786 | extern kern_return_t task_pidsuspend( |
| 787 | task_t task); |
| 788 | |
| 789 | /* Resumes a previously paused task */ |
| 790 | extern kern_return_t task_pidresume( |
| 791 | task_t task); |
| 792 | |
| 793 | extern kern_return_t task_send_trace_memory( |
| 794 | task_t task, |
| 795 | uint32_t pid, |
| 796 | uint64_t uniqueid); |
| 797 | |
| 798 | extern void task_remove_turnstile_watchports( |
| 799 | task_t task); |
| 800 | |
| 801 | extern void task_transfer_turnstile_watchports( |
| 802 | task_t old_task, |
| 803 | task_t new_task, |
| 804 | thread_t new_thread); |
| 805 | |
| 806 | extern kern_return_t |
| 807 | task_violated_guard(mach_exception_code_t, mach_exception_subcode_t, void *, bool); |
| 808 | |
| 809 | #if DEVELOPMENT || DEBUG |
| 810 | |
| 811 | extern kern_return_t task_disconnect_page_mappings( |
| 812 | task_t task); |
| 813 | #endif /* DEVELOPMENT || DEBUG */ |
| 814 | |
| 815 | extern void tasks_system_suspend(boolean_t suspend); |
| 816 | |
| 817 | #if CONFIG_FREEZE |
| 818 | |
| 819 | /* Freeze a task's resident pages */ |
| 820 | extern kern_return_t task_freeze( |
| 821 | task_t task, |
| 822 | uint32_t *purgeable_count, |
| 823 | uint32_t *wired_count, |
| 824 | uint32_t *clean_count, |
| 825 | uint32_t *dirty_count, |
| 826 | uint32_t dirty_budget, |
| 827 | uint32_t *shared_count, |
| 828 | int *freezer_error_code, |
| 829 | boolean_t eval_only); |
| 830 | |
| 831 | /* Thaw a currently frozen task */ |
| 832 | extern kern_return_t task_thaw( |
| 833 | task_t task); |
| 834 | |
| 835 | typedef enum { |
| 836 | CREDIT_TO_SWAP = 1, |
| 837 | DEBIT_FROM_SWAP = 2 |
| 838 | } freezer_acct_op_t; |
| 839 | |
| 840 | extern void task_update_frozen_to_swap_acct( |
| 841 | task_t task, |
| 842 | int64_t amount, |
| 843 | freezer_acct_op_t op); |
| 844 | |
| 845 | #endif /* CONFIG_FREEZE */ |
| 846 | |
| 847 | /* Halt all other threads in the current task */ |
| 848 | extern kern_return_t task_start_halt( |
| 849 | task_t task); |
| 850 | |
| 851 | /* Wait for other threads to halt and free halting task resources */ |
| 852 | extern void task_complete_halt( |
| 853 | task_t task); |
| 854 | |
| 855 | extern kern_return_t task_terminate_internal( |
| 856 | task_t task); |
| 857 | |
| 858 | struct proc_ro; |
| 859 | typedef struct proc_ro *proc_ro_t; |
| 860 | |
| 861 | extern kern_return_t task_create_internal( |
| 862 | task_t parent_task, |
| 863 | proc_ro_t proc_ro, |
| 864 | coalition_t *parent_coalitions, |
| 865 | boolean_t inherit_memory, |
| 866 | boolean_t is_64bit, |
| 867 | boolean_t is_64bit_data, |
| 868 | uint32_t t_flags, |
| 869 | uint32_t t_flags_ro, |
| 870 | uint32_t procflags, |
| 871 | uint8_t t_returnwaitflags, |
| 872 | task_t child_task); |
| 873 | |
| 874 | extern kern_return_t task_set_special_port_internal( |
| 875 | task_t task, |
| 876 | int which, |
| 877 | ipc_port_t port); |
| 878 | |
| 879 | extern kern_return_t task_set_security_tokens( |
| 880 | task_t task, |
| 881 | security_token_t sec_token, |
| 882 | audit_token_t audit_token, |
| 883 | host_priv_t host_priv); |
| 884 | |
| 885 | extern kern_return_t task_info( |
| 886 | task_t task, |
| 887 | task_flavor_t flavor, |
| 888 | task_info_t task_info_out, |
| 889 | mach_msg_type_number_t *task_info_count); |
| 890 | |
| 891 | /* |
| 892 | * Additional fields that aren't exposed through `task_power_info` but needed |
| 893 | * by clients of `task_power_info_locked`. |
| 894 | */ |
| 895 | struct { |
| 896 | uint64_t ; |
| 897 | uint64_t ; |
| 898 | uint64_t ; |
| 899 | uint64_t ; |
| 900 | uint64_t ; |
| 901 | uint64_t ; |
| 902 | uint64_t ; |
| 903 | uint64_t ; |
| 904 | uint64_t ; |
| 905 | uint64_t ; |
| 906 | uint64_t ; |
| 907 | }; |
| 908 | |
| 909 | void task_power_info_locked( |
| 910 | task_t task, |
| 911 | task_power_info_t info, |
| 912 | gpu_energy_data_t gpu_energy, |
| 913 | task_power_info_v2_t infov2, |
| 914 | struct task_power_info_extra *); |
| 915 | |
| 916 | extern uint64_t task_gpu_utilisation( |
| 917 | task_t task); |
| 918 | |
| 919 | extern void task_update_cpu_time_qos_stats( |
| 920 | task_t task, |
| 921 | uint64_t *eqos_stats, |
| 922 | uint64_t *rqos_stats); |
| 923 | |
| 924 | extern void task_vtimer_set( |
| 925 | task_t task, |
| 926 | integer_t which); |
| 927 | |
| 928 | extern void task_vtimer_clear( |
| 929 | task_t task, |
| 930 | integer_t which); |
| 931 | |
| 932 | extern void task_vtimer_update( |
| 933 | task_t task, |
| 934 | integer_t which, |
| 935 | uint32_t *microsecs); |
| 936 | |
| 937 | #define TASK_VTIMER_USER 0x01 |
| 938 | #define TASK_VTIMER_PROF 0x02 |
| 939 | #define TASK_VTIMER_RLIM 0x04 |
| 940 | |
| 941 | extern void task_set_64bit( |
| 942 | task_t task, |
| 943 | boolean_t is_64bit, |
| 944 | boolean_t is_64bit_data); |
| 945 | |
| 946 | extern bool task_get_64bit_addr( |
| 947 | task_t task); |
| 948 | |
| 949 | extern bool task_get_64bit_data( |
| 950 | task_t task); |
| 951 | |
| 952 | extern void task_set_platform_binary( |
| 953 | task_t task, |
| 954 | boolean_t is_platform); |
| 955 | |
| 956 | #if XNU_TARGET_OS_OSX |
| 957 | #if DEVELOPMENT || DEBUG |
| 958 | /* Disables task identity security hardening (*_set_exception_ports policy) |
| 959 | * for all tasks if amfi_get_out_of_my_way is set. */ |
| 960 | extern bool AMFI_bootarg_disable_mach_hardening; |
| 961 | #endif /* DEVELOPMENT || DEBUG */ |
| 962 | extern void task_disable_mach_hardening( |
| 963 | task_t task); |
| 964 | |
| 965 | extern bool task_opted_out_mach_hardening( |
| 966 | task_t task); |
| 967 | #endif /* XNU_TARGET_OS_OSX */ |
| 968 | |
| 969 | extern boolean_t task_get_platform_binary( |
| 970 | task_t task); |
| 971 | |
| 972 | extern void |
| 973 | task_set_hardened_runtime( |
| 974 | task_t task, |
| 975 | bool is_hardened); |
| 976 | |
| 977 | extern boolean_t |
| 978 | task_is_hardened_binary( |
| 979 | task_t task); |
| 980 | |
| 981 | extern boolean_t task_is_a_corpse( |
| 982 | task_t task); |
| 983 | |
| 984 | extern boolean_t task_is_ipc_active( |
| 985 | task_t task); |
| 986 | |
| 987 | extern void task_set_corpse( |
| 988 | task_t task); |
| 989 | |
| 990 | extern void task_set_exc_guard_ctrl_port_default( |
| 991 | task_t task, |
| 992 | thread_t main_thread, |
| 993 | const char *name, |
| 994 | unsigned int namelen, |
| 995 | boolean_t is_simulated, |
| 996 | uint32_t platform, |
| 997 | uint32_t sdk); |
| 998 | |
| 999 | extern void task_set_immovable_pinned(task_t task); |
| 1000 | |
| 1001 | extern bool task_set_ca_client_wi( |
| 1002 | task_t task, |
| 1003 | boolean_t ca_client_wi); |
| 1004 | |
| 1005 | extern kern_return_t task_set_dyld_info( |
| 1006 | task_t task, |
| 1007 | mach_vm_address_t addr, |
| 1008 | mach_vm_size_t size); |
| 1009 | |
| 1010 | extern void ( |
| 1011 | task_t task, |
| 1012 | mach_vm_address_t addr); |
| 1013 | |
| 1014 | extern void task_set_uniqueid(task_t task); |
| 1015 | |
| 1016 | /* Get number of activations in a task */ |
| 1017 | extern int get_task_numacts( |
| 1018 | task_t task); |
| 1019 | |
| 1020 | extern bool task_donates_own_pages( |
| 1021 | task_t task); |
| 1022 | |
| 1023 | struct label; |
| 1024 | extern kern_return_t task_collect_crash_info( |
| 1025 | task_t task, |
| 1026 | #if CONFIG_MACF |
| 1027 | struct label *crash_label, |
| 1028 | #endif |
| 1029 | int is_corpse_fork); |
| 1030 | void task_wait_till_threads_terminate_locked(task_t task); |
| 1031 | |
| 1032 | /* JMM - should just be temporary (implementation in bsd_kern still) */ |
| 1033 | extern void set_bsdtask_info(task_t, void *); |
| 1034 | extern uint32_t set_task_loadTag(task_t task, uint32_t loadTag); |
| 1035 | extern vm_map_t get_task_map_reference(task_t); |
| 1036 | extern vm_map_t swap_task_map(task_t, thread_t, vm_map_t); |
| 1037 | extern pmap_t get_task_pmap(task_t); |
| 1038 | extern uint64_t get_task_resident_size(task_t); |
| 1039 | extern uint64_t get_task_compressed(task_t); |
| 1040 | extern uint64_t get_task_resident_max(task_t); |
| 1041 | extern uint64_t (task_t); |
| 1042 | #if CONFIG_LEDGER_INTERVAL_MAX |
| 1043 | extern uint64_t (task_t, int reset); |
| 1044 | #endif /* CONFIG_FOOTPRINT_INTERVAL_MAX */ |
| 1045 | extern uint64_t (task_t); |
| 1046 | extern uint64_t (task_t); |
| 1047 | extern uint64_t get_task_purgeable_size(task_t); |
| 1048 | extern uint64_t get_task_cpu_time(task_t); |
| 1049 | extern uint64_t get_task_dispatchqueue_offset(task_t); |
| 1050 | extern uint64_t get_task_dispatchqueue_serialno_offset(task_t); |
| 1051 | extern uint64_t get_task_dispatchqueue_label_offset(task_t); |
| 1052 | extern uint64_t get_task_uniqueid(task_t task); |
| 1053 | extern int get_task_version(task_t task); |
| 1054 | |
| 1055 | extern uint64_t get_task_internal(task_t); |
| 1056 | extern uint64_t get_task_internal_compressed(task_t); |
| 1057 | extern uint64_t get_task_purgeable_nonvolatile(task_t); |
| 1058 | extern uint64_t get_task_purgeable_nonvolatile_compressed(task_t); |
| 1059 | extern uint64_t get_task_iokit_mapped(task_t); |
| 1060 | extern uint64_t get_task_alternate_accounting(task_t); |
| 1061 | extern uint64_t get_task_alternate_accounting_compressed(task_t); |
| 1062 | extern uint64_t get_task_memory_region_count(task_t); |
| 1063 | extern uint64_t get_task_page_table(task_t); |
| 1064 | #if CONFIG_FREEZE |
| 1065 | extern uint64_t get_task_frozen_to_swap(task_t); |
| 1066 | #endif |
| 1067 | extern uint64_t get_task_network_nonvolatile(task_t); |
| 1068 | extern uint64_t get_task_network_nonvolatile_compressed(task_t); |
| 1069 | extern uint64_t get_task_wired_mem(task_t); |
| 1070 | extern uint32_t get_task_loadTag(task_t task); |
| 1071 | |
| 1072 | extern uint64_t (task_t task); |
| 1073 | extern uint64_t (task_t task); |
| 1074 | extern uint64_t (task_t task); |
| 1075 | extern uint64_t (task_t task); |
| 1076 | extern uint64_t (task_t task); |
| 1077 | extern uint64_t (task_t task); |
| 1078 | extern uint64_t (task_t task); |
| 1079 | extern uint64_t (task_t task); |
| 1080 | |
| 1081 | extern kern_return_t (int, int *); |
| 1082 | extern kern_return_t (task_t, int, int *, boolean_t, boolean_t); |
| 1083 | extern kern_return_t (task_t task, int *limit_mb); |
| 1084 | #if DEBUG || DEVELOPMENT |
| 1085 | #if CONFIG_MEMORYSTATUS |
| 1086 | extern kern_return_t task_set_diag_footprint_limit_internal(task_t, uint64_t, uint64_t *); |
| 1087 | extern kern_return_t task_get_diag_footprint_limit_internal(task_t, uint64_t *, bool *); |
| 1088 | extern kern_return_t task_set_diag_footprint_limit(task_t task, uint64_t new_limit_mb, uint64_t *old_limit_mb); |
| 1089 | #endif /* CONFIG_MEMORYSTATUS */ |
| 1090 | #endif /* DEBUG || DEVELOPMENT */ |
| 1091 | |
| 1092 | extern security_token_t *task_get_sec_token(task_t task); |
| 1093 | extern void task_set_sec_token(task_t task, security_token_t *token); |
| 1094 | extern audit_token_t *task_get_audit_token(task_t task); |
| 1095 | extern void task_set_audit_token(task_t task, audit_token_t *token); |
| 1096 | extern void task_set_tokens(task_t task, security_token_t *sec_token, audit_token_t *audit_token); |
| 1097 | extern boolean_t task_is_privileged(task_t task); |
| 1098 | extern uint8_t *task_get_mach_trap_filter_mask(task_t task); |
| 1099 | extern void task_set_mach_trap_filter_mask(task_t task, uint8_t *mask); |
| 1100 | extern uint8_t *task_get_mach_kobj_filter_mask(task_t task); |
| 1101 | extern void task_set_mach_kobj_filter_mask(task_t task, uint8_t *mask); |
| 1102 | extern mach_vm_address_t task_get_all_image_info_addr(task_t task); |
| 1103 | |
| 1104 | /* Jetsam memlimit attributes */ |
| 1105 | extern boolean_t task_get_memlimit_is_active(task_t task); |
| 1106 | extern boolean_t task_get_memlimit_is_fatal(task_t task); |
| 1107 | extern void task_set_memlimit_is_active(task_t task, boolean_t memlimit_is_active); |
| 1108 | extern void task_set_memlimit_is_fatal(task_t task, boolean_t memlimit_is_fatal); |
| 1109 | extern boolean_t task_has_triggered_exc_resource(task_t task, boolean_t memlimit_is_active); |
| 1110 | extern void task_mark_has_triggered_exc_resource(task_t task, boolean_t memlimit_is_active); |
| 1111 | |
| 1112 | extern uint64_t task_get_dirty_start(task_t task); |
| 1113 | extern void task_set_dirty_start(task_t task, uint64_t start); |
| 1114 | |
| 1115 | extern void task_set_thread_limit(task_t task, uint16_t thread_limit); |
| 1116 | #if CONFIG_PROC_RESOURCE_LIMITS |
| 1117 | extern kern_return_t task_set_port_space_limits(task_t task, uint32_t soft_limit, uint32_t hard_limit); |
| 1118 | #endif /* CONFIG_PROC_RESOURCE_LIMITS */ |
| 1119 | extern void task_port_space_ast(task_t task); |
| 1120 | |
| 1121 | #if XNU_TARGET_OS_OSX |
| 1122 | extern boolean_t task_has_system_version_compat_enabled(task_t task); |
| 1123 | extern void task_set_system_version_compat_enabled(task_t task, boolean_t enable_system_version_compat); |
| 1124 | #endif |
| 1125 | |
| 1126 | extern boolean_t is_kerneltask(task_t task); |
| 1127 | extern boolean_t is_corpsefork(task_t task); |
| 1128 | |
| 1129 | extern kern_return_t check_actforsig(task_t task, thread_t thread, int setast); |
| 1130 | |
| 1131 | extern kern_return_t machine_task_get_state( |
| 1132 | task_t task, |
| 1133 | int flavor, |
| 1134 | thread_state_t state, |
| 1135 | mach_msg_type_number_t *state_count); |
| 1136 | |
| 1137 | extern kern_return_t machine_task_set_state( |
| 1138 | task_t task, |
| 1139 | int flavor, |
| 1140 | thread_state_t state, |
| 1141 | mach_msg_type_number_t state_count); |
| 1142 | |
| 1143 | extern void machine_task_terminate(task_t task); |
| 1144 | |
| 1145 | extern kern_return_t machine_task_process_signature(task_t task, uint32_t platform, uint32_t sdk, char const **error_msg); |
| 1146 | |
| 1147 | struct _task_ledger_indices { |
| 1148 | int cpu_time; |
| 1149 | int tkm_private; |
| 1150 | int tkm_shared; |
| 1151 | int phys_mem; |
| 1152 | int wired_mem; |
| 1153 | int conclave_mem; |
| 1154 | int internal; |
| 1155 | int iokit_mapped; |
| 1156 | int external; |
| 1157 | int reusable; |
| 1158 | int alternate_accounting; |
| 1159 | int alternate_accounting_compressed; |
| 1160 | int page_table; |
| 1161 | int ; |
| 1162 | int internal_compressed; |
| 1163 | int purgeable_volatile; |
| 1164 | int purgeable_nonvolatile; |
| 1165 | int purgeable_volatile_compressed; |
| 1166 | int purgeable_nonvolatile_compressed; |
| 1167 | int ; |
| 1168 | int ; |
| 1169 | int ; |
| 1170 | int ; |
| 1171 | int network_volatile; |
| 1172 | int network_nonvolatile; |
| 1173 | int network_volatile_compressed; |
| 1174 | int network_nonvolatile_compressed; |
| 1175 | int ; |
| 1176 | int ; |
| 1177 | int ; |
| 1178 | int ; |
| 1179 | int ; |
| 1180 | int ; |
| 1181 | int ; |
| 1182 | int ; |
| 1183 | int ; |
| 1184 | int ; |
| 1185 | int ; |
| 1186 | int ; |
| 1187 | int platform_idle_wakeups; |
| 1188 | int interrupt_wakeups; |
| 1189 | #if CONFIG_SCHED_SFI |
| 1190 | int sfi_wait_times[MAX_SFI_CLASS_ID]; |
| 1191 | #endif /* CONFIG_SCHED_SFI */ |
| 1192 | int cpu_time_billed_to_me; |
| 1193 | int cpu_time_billed_to_others; |
| 1194 | int physical_writes; |
| 1195 | int logical_writes; |
| 1196 | int logical_writes_to_external; |
| 1197 | int energy_billed_to_me; |
| 1198 | int energy_billed_to_others; |
| 1199 | #if CONFIG_MEMORYSTATUS |
| 1200 | int memorystatus_dirty_time; |
| 1201 | #endif /* CONFIG_MEMORYSTATUS */ |
| 1202 | #if DEBUG || DEVELOPMENT |
| 1203 | int pages_grabbed; |
| 1204 | int pages_grabbed_kern; |
| 1205 | int pages_grabbed_iopl; |
| 1206 | int pages_grabbed_upl; |
| 1207 | #endif |
| 1208 | #if CONFIG_FREEZE |
| 1209 | int frozen_to_swap; |
| 1210 | #endif /* CONFIG_FREEZE */ |
| 1211 | #if CONFIG_PHYS_WRITE_ACCT |
| 1212 | int fs_metadata_writes; |
| 1213 | #endif /* CONFIG_PHYS_WRITE_ACCT */ |
| 1214 | int swapins; |
| 1215 | }; |
| 1216 | |
| 1217 | /* |
| 1218 | * Many of the task ledger entries use a reduced feature set |
| 1219 | * (specifically they just use LEDGER_ENTRY_ALLOW_PANIC_ON_NEGATIVE) |
| 1220 | * and are stored in a smaller entry structure. |
| 1221 | * That structure is an implementation detail of the ledger. |
| 1222 | * But on PPL systems, the task ledger's memory is managed by the PPL |
| 1223 | * and it has to determine the size of the task ledger at compile time. |
| 1224 | * This define specifies the number of small entries so the PPL can |
| 1225 | * properly determine the ledger's size. |
| 1226 | * |
| 1227 | * If you add a new entry with only the |
| 1228 | * LEDGER_ENTRY_ALLOW_PANIC_ON_NEGATIVE | LEDGER_ENTRY_ALLOW_INACTIVE |
| 1229 | * flags, you need to increment this count. |
| 1230 | * Otherwise, PPL systems will panic at boot. |
| 1231 | */ |
| 1232 | #if DEVELOPMENT || DEBUG |
| 1233 | #define TASK_LEDGER_NUM_SMALL_INDICES 33 |
| 1234 | #else |
| 1235 | #define TASK_LEDGER_NUM_SMALL_INDICES 29 |
| 1236 | #endif /* DEVELOPMENT || DEBUG */ |
| 1237 | extern struct _task_ledger_indices task_ledgers; |
| 1238 | |
| 1239 | /* requires task to be unlocked, returns a referenced thread */ |
| 1240 | thread_t task_findtid(task_t task, uint64_t tid); |
| 1241 | int pid_from_task(task_t task); |
| 1242 | |
| 1243 | extern kern_return_t task_wakeups_monitor_ctl(task_t task, uint32_t *rate_hz, int32_t *flags); |
| 1244 | extern kern_return_t task_cpu_usage_monitor_ctl(task_t task, uint32_t *flags); |
| 1245 | extern void task_rollup_accounting_info(task_t new_task, task_t parent_task); |
| 1246 | extern kern_return_t task_io_monitor_ctl(task_t task, uint32_t *flags); |
| 1247 | extern void task_set_did_exec_flag(task_t task); |
| 1248 | extern void task_clear_exec_copy_flag(task_t task); |
| 1249 | extern boolean_t task_is_exec_copy(task_t); |
| 1250 | extern boolean_t task_did_exec(task_t task); |
| 1251 | extern boolean_t task_is_active(task_t task); |
| 1252 | extern boolean_t task_is_halting(task_t task); |
| 1253 | extern void task_clear_return_wait(task_t task, uint32_t flags); |
| 1254 | extern void task_wait_to_return(void) __attribute__((noreturn)); |
| 1255 | extern event_t task_get_return_wait_event(task_t task); |
| 1256 | |
| 1257 | extern void task_bank_reset(task_t task); |
| 1258 | extern void task_bank_init(task_t task); |
| 1259 | |
| 1260 | #if CONFIG_MEMORYSTATUS |
| 1261 | extern void task_ledger_settle_dirty_time(task_t t); |
| 1262 | #endif /* CONFIG_MEMORYSTATUS */ |
| 1263 | |
| 1264 | #if CONFIG_ARCADE |
| 1265 | extern void task_prep_arcade(task_t task, thread_t thread); |
| 1266 | #endif /* CONFIG_ARCADE */ |
| 1267 | |
| 1268 | extern int task_pid(task_t task); |
| 1269 | |
| 1270 | #if __has_feature(ptrauth_calls) |
| 1271 | char *task_get_vm_shared_region_id_and_jop_pid(task_t task, uint64_t *); |
| 1272 | void task_set_shared_region_id(task_t task, char *id); |
| 1273 | #endif /* __has_feature(ptrauth_calls) */ |
| 1274 | |
| 1275 | extern boolean_t task_has_assertions(task_t task); |
| 1276 | /* End task_policy */ |
| 1277 | |
| 1278 | extern void task_set_gpu_denied(task_t task, boolean_t denied); |
| 1279 | extern boolean_t task_is_gpu_denied(task_t task); |
| 1280 | |
| 1281 | extern void task_set_game_mode(task_t task, bool enabled); |
| 1282 | /* returns true if update must be pushed to coalition (Automatically handled by task_set_game_mode) */ |
| 1283 | extern bool task_set_game_mode_locked(task_t task, bool enabled); |
| 1284 | extern bool task_get_game_mode(task_t task); |
| 1285 | |
| 1286 | extern queue_head_t * task_io_user_clients(task_t task); |
| 1287 | extern void task_set_message_app_suspended(task_t task, boolean_t enable); |
| 1288 | |
| 1289 | extern void task_copy_fields_for_exec(task_t dst_task, task_t src_task); |
| 1290 | |
| 1291 | extern void task_copy_vmobjects(task_t task, vm_object_query_t query, size_t len, size_t *num); |
| 1292 | extern void task_get_owned_vmobjects(task_t task, size_t buffer_size, vmobject_list_output_t buffer, size_t* output_size, size_t* entries); |
| 1293 | extern void task_store_owned_vmobject_info(task_t to_task, task_t from_task); |
| 1294 | |
| 1295 | extern void task_set_filter_msg_flag(task_t task, boolean_t flag); |
| 1296 | extern boolean_t task_get_filter_msg_flag(task_t task); |
| 1297 | |
| 1298 | #if __has_feature(ptrauth_calls) |
| 1299 | extern bool task_is_pac_exception_fatal(task_t task); |
| 1300 | extern void task_set_pac_exception_fatal_flag(task_t task); |
| 1301 | #endif /*__has_feature(ptrauth_calls)*/ |
| 1302 | |
| 1303 | extern bool task_is_jit_exception_fatal(task_t task); |
| 1304 | extern void task_set_jit_exception_fatal_flag(task_t task); |
| 1305 | |
| 1306 | extern bool task_needs_user_signed_thread_state(task_t task); |
| 1307 | extern void task_set_tecs(task_t task); |
| 1308 | extern void task_get_corpse_vmobject_list(task_t task, vmobject_list_output_t* list, size_t* list_size); |
| 1309 | |
| 1310 | extern boolean_t task_corpse_forking_disabled(task_t task); |
| 1311 | |
| 1312 | void __attribute__((noinline)) SENDING_NOTIFICATION__THIS_PROCESS_HAS_TOO_MANY_MACH_PORTS(task_t task, |
| 1313 | uint32_t current_size, uint32_t soft_limit, uint32_t hard_limit); |
| 1314 | |
| 1315 | extern int get_task_cdhash(task_t task, char cdhash[CS_CDHASH_LEN]); |
| 1316 | |
| 1317 | extern boolean_t kdp_task_is_locked(task_t task); |
| 1318 | |
| 1319 | /* Kernel side prototypes for MIG routines */ |
| 1320 | extern kern_return_t task_get_exception_ports( |
| 1321 | task_t task, |
| 1322 | exception_mask_t exception_mask, |
| 1323 | exception_mask_array_t masks, |
| 1324 | mach_msg_type_number_t *CountCnt, |
| 1325 | exception_port_array_t ports, |
| 1326 | exception_behavior_array_t behaviors, |
| 1327 | thread_state_flavor_array_t flavors); |
| 1328 | |
| 1329 | #if CONFIG_EXCLAVES |
| 1330 | int task_add_conclave(task_t task, void *, int64_t, const char *task_conclave_id); |
| 1331 | kern_return_t task_inherit_conclave(task_t old_task, task_t new_task, void *vnode, int64_t off); |
| 1332 | kern_return_t task_launch_conclave(mach_port_name_t port); |
| 1333 | void task_clear_conclave(task_t task); |
| 1334 | void task_stop_conclave(task_t task, bool gather_crash_bt); |
| 1335 | kern_return_t task_stop_conclave_upcall(void); |
| 1336 | kern_return_t task_stop_conclave_upcall_complete(void); |
| 1337 | kern_return_t task_suspend_conclave_upcall(uint64_t *, size_t); |
| 1338 | struct xnuupcalls_conclavesharedbuffer_s; |
| 1339 | kern_return_t task_crash_info_conclave_upcall(task_t task, |
| 1340 | const struct xnuupcalls_conclavesharedbuffer_s *shared_buf, uint32_t length); |
| 1341 | typedef struct exclaves_resource exclaves_resource_t; |
| 1342 | exclaves_resource_t *task_get_conclave(task_t task); |
| 1343 | void task_set_conclave_untaintable(task_t task); |
| 1344 | void task_add_conclave_crash_info(task_t task, void *crash_info_ptr); |
| 1345 | //Changing this would also warrant a change in ConclaveSharedBuffer |
| 1346 | #define CONCLAVE_CRASH_BUFFER_PAGECOUNT 2 |
| 1347 | |
| 1348 | #endif /* CONFIG_EXCLAVES */ |
| 1349 | |
| 1350 | #endif /* XNU_KERNEL_PRIVATE */ |
| 1351 | #ifdef KERNEL_PRIVATE |
| 1352 | |
| 1353 | extern void *get_bsdtask_info(task_t); |
| 1354 | extern void *get_bsdthreadtask_info(thread_t); |
| 1355 | extern void task_bsdtask_kill(task_t); |
| 1356 | extern vm_map_t get_task_map(task_t); |
| 1357 | extern ledger_t get_task_ledger(task_t); |
| 1358 | |
| 1359 | extern boolean_t get_task_pidsuspended(task_t); |
| 1360 | extern boolean_t get_task_suspended(task_t); |
| 1361 | extern boolean_t get_task_frozen(task_t); |
| 1362 | |
| 1363 | /* |
| 1364 | * Flavors of convert_task_to_port. XNU callers get convert_task_to_port_kernel, |
| 1365 | * external callers get convert_task_to_port_external. |
| 1366 | */ |
| 1367 | extern ipc_port_t convert_task_to_port(task_t); |
| 1368 | extern ipc_port_t convert_task_to_port_kernel(task_t); |
| 1369 | extern ipc_port_t convert_task_to_port_external(task_t); |
| 1370 | extern ipc_port_t convert_task_to_port_pinned(task_t); |
| 1371 | |
| 1372 | extern ipc_port_t convert_task_read_to_port(task_t); |
| 1373 | extern ipc_port_t convert_task_read_to_port_kernel(task_read_t); |
| 1374 | extern ipc_port_t convert_task_read_to_port_external(task_t); |
| 1375 | |
| 1376 | extern ipc_port_t convert_task_inspect_to_port(task_inspect_t); |
| 1377 | extern ipc_port_t convert_task_name_to_port(task_name_t); |
| 1378 | |
| 1379 | extern ipc_port_t convert_corpse_to_port_and_nsrequest(task_t task); |
| 1380 | |
| 1381 | extern ipc_port_t convert_task_suspension_token_to_port(task_suspension_token_t task); |
| 1382 | /* Convert from a port (in this case, an SO right to a task's resume port) to a task. */ |
| 1383 | extern task_suspension_token_t convert_port_to_task_suspension_token(ipc_port_t port); |
| 1384 | |
| 1385 | extern void task_suspension_send_once(ipc_port_t port); |
| 1386 | |
| 1387 | #define TASK_WRITE_IMMEDIATE 0x1 |
| 1388 | #define TASK_WRITE_DEFERRED 0x2 |
| 1389 | #define TASK_WRITE_INVALIDATED 0x4 |
| 1390 | #define TASK_WRITE_METADATA 0x8 |
| 1391 | extern void task_update_logical_writes(task_t task, uint32_t io_size, int flags, void *vp); |
| 1392 | |
| 1393 | __enum_decl(task_balance_flags_t, uint8_t, { |
| 1394 | TASK_BALANCE_CREDIT = 0x1, |
| 1395 | TASK_BALANCE_DEBIT = 0x2, |
| 1396 | }); |
| 1397 | |
| 1398 | __enum_decl(task_physical_write_flavor_t, uint8_t, { |
| 1399 | TASK_PHYSICAL_WRITE_METADATA = 0x1, |
| 1400 | }); |
| 1401 | extern void task_update_physical_writes(task_t task, task_physical_write_flavor_t flavor, |
| 1402 | uint64_t io_size, task_balance_flags_t flags); |
| 1403 | |
| 1404 | #if CONFIG_SECLUDED_MEMORY |
| 1405 | extern void task_set_can_use_secluded_mem( |
| 1406 | task_t task, |
| 1407 | boolean_t can_use_secluded_mem); |
| 1408 | extern void task_set_could_use_secluded_mem( |
| 1409 | task_t task, |
| 1410 | boolean_t could_use_secluded_mem); |
| 1411 | extern void task_set_could_also_use_secluded_mem( |
| 1412 | task_t task, |
| 1413 | boolean_t could_also_use_secluded_mem); |
| 1414 | extern boolean_t task_can_use_secluded_mem( |
| 1415 | task_t task, |
| 1416 | boolean_t is_allocate); |
| 1417 | extern boolean_t task_could_use_secluded_mem(task_t task); |
| 1418 | extern boolean_t task_could_also_use_secluded_mem(task_t task); |
| 1419 | #endif /* CONFIG_SECLUDED_MEMORY */ |
| 1420 | |
| 1421 | extern void task_set_darkwake_mode(task_t, boolean_t); |
| 1422 | extern boolean_t task_get_darkwake_mode(task_t); |
| 1423 | |
| 1424 | #if __arm64__ |
| 1425 | extern void (task_t task); |
| 1426 | extern void (task_t task); |
| 1427 | extern void (task_t task); |
| 1428 | #endif /* __arm64__ */ |
| 1429 | |
| 1430 | #if CONFIG_MACF |
| 1431 | extern struct label *get_task_crash_label(task_t task); |
| 1432 | extern void set_task_crash_label(task_t task, struct label *label); |
| 1433 | #endif /* CONFIG_MACF */ |
| 1434 | |
| 1435 | #endif /* KERNEL_PRIVATE */ |
| 1436 | |
| 1437 | extern task_t kernel_task; |
| 1438 | |
| 1439 | extern void task_name_deallocate_mig( |
| 1440 | task_name_t task_name); |
| 1441 | |
| 1442 | extern void task_policy_set_deallocate_mig( |
| 1443 | task_policy_set_t task_policy_set); |
| 1444 | |
| 1445 | extern void task_policy_get_deallocate_mig( |
| 1446 | task_policy_get_t task_policy_get); |
| 1447 | |
| 1448 | extern void task_inspect_deallocate_mig( |
| 1449 | task_inspect_t task_inspect); |
| 1450 | |
| 1451 | extern void task_read_deallocate_mig( |
| 1452 | task_read_t task_read); |
| 1453 | |
| 1454 | extern void task_suspension_token_deallocate( |
| 1455 | task_suspension_token_t token); |
| 1456 | |
| 1457 | extern boolean_t (void); |
| 1458 | extern void (boolean_t newval); |
| 1459 | extern void (ledger_t ledger, |
| 1460 | ledger_amount_t *ledger_resident, |
| 1461 | ledger_amount_t *ledger_compressed); |
| 1462 | extern void task_set_memory_ownership_transfer( |
| 1463 | task_t task, |
| 1464 | boolean_t value); |
| 1465 | |
| 1466 | #if DEVELOPMENT || DEBUG |
| 1467 | extern void task_set_no_footprint_for_debug( |
| 1468 | task_t task, |
| 1469 | boolean_t value); |
| 1470 | extern int task_get_no_footprint_for_debug( |
| 1471 | task_t task); |
| 1472 | #endif /* DEVELOPMENT || DEBUG */ |
| 1473 | |
| 1474 | #ifdef KERNEL_PRIVATE |
| 1475 | extern kern_return_t task_get_suspend_stats(task_t task, task_suspend_stats_t stats); |
| 1476 | extern kern_return_t task_get_suspend_stats_kdp(task_t task, task_suspend_stats_t stats); |
| 1477 | #endif /* KERNEL_PRIVATE*/ |
| 1478 | |
| 1479 | #ifdef XNU_KERNEL_PRIVATE |
| 1480 | extern kern_return_t task_get_suspend_sources(task_t task, task_suspend_source_array_t sources); |
| 1481 | extern kern_return_t task_get_suspend_sources_kdp(task_t task, task_suspend_source_array_t sources); |
| 1482 | #endif /* XNU_KERNEL_PRIVATE */ |
| 1483 | |
| 1484 | #if CONFIG_ROSETTA |
| 1485 | extern bool task_is_translated(task_t task); |
| 1486 | #endif |
| 1487 | |
| 1488 | |
| 1489 | |
| 1490 | #ifdef MACH_KERNEL_PRIVATE |
| 1491 | void task_procname(task_t task, char *buf, int size); |
| 1492 | void task_best_name(task_t task, char *buf, size_t size); |
| 1493 | #endif /* MACH_KERNEL_PRIVATE */ |
| 1494 | |
| 1495 | __END_DECLS |
| 1496 | |
| 1497 | #endif /* _KERN_TASK_H_ */ |
| 1498 | |