| File: | obj-scan-build/../kern/processor.c |
| Location: | line 927, column 4 |
| Description: | Array access (from variable 'threads') results in a null pointer dereference |
| 1 | /* | |||
| 2 | * Mach Operating System | |||
| 3 | * Copyright (c) 1993-1988 Carnegie Mellon University | |||
| 4 | * All Rights Reserved. | |||
| 5 | * | |||
| 6 | * Permission to use, copy, modify and distribute this software and its | |||
| 7 | * documentation is hereby granted, provided that both the copyright | |||
| 8 | * notice and this permission notice appear in all copies of the | |||
| 9 | * software, derivative works or modified versions, and any portions | |||
| 10 | * thereof, and that both notices appear in supporting documentation. | |||
| 11 | * | |||
| 12 | * CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS" | |||
| 13 | * CONDITION. CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND FOR | |||
| 14 | * ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE. | |||
| 15 | * | |||
| 16 | * Carnegie Mellon requests users of this software to return to | |||
| 17 | * | |||
| 18 | * Software Distribution Coordinator or Software.Distribution@CS.CMU.EDU | |||
| 19 | * School of Computer Science | |||
| 20 | * Carnegie Mellon University | |||
| 21 | * Pittsburgh PA 15213-3890 | |||
| 22 | * | |||
| 23 | * any improvements or extensions that they make and grant Carnegie Mellon | |||
| 24 | * the rights to redistribute these changes. | |||
| 25 | */ | |||
| 26 | /* | |||
| 27 | * processor.c: processor and processor_set manipulation routines. | |||
| 28 | */ | |||
| 29 | ||||
| 30 | #include <string.h> | |||
| 31 | ||||
| 32 | #include <mach/boolean.h> | |||
| 33 | #include <mach/policy.h> | |||
| 34 | #include <mach/processor_info.h> | |||
| 35 | #include <mach/vm_param.h> | |||
| 36 | #include <kern/cpu_number.h> | |||
| 37 | #include <kern/debug.h> | |||
| 38 | #include <kern/kalloc.h> | |||
| 39 | #include <kern/lock.h> | |||
| 40 | #include <kern/host.h> | |||
| 41 | #include <kern/ipc_tt.h> | |||
| 42 | #include <kern/processor.h> | |||
| 43 | #include <kern/sched.h> | |||
| 44 | #include <kern/task.h> | |||
| 45 | #include <kern/thread.h> | |||
| 46 | #include <kern/ipc_host.h> | |||
| 47 | #include <ipc/ipc_port.h> | |||
| 48 | ||||
| 49 | #if MACH_HOST0 | |||
| 50 | #include <kern/slab.h> | |||
| 51 | struct kmem_cache pset_cache; | |||
| 52 | #endif /* MACH_HOST */ | |||
| 53 | ||||
| 54 | ||||
| 55 | /* | |||
| 56 | * Exported variables. | |||
| 57 | */ | |||
| 58 | struct processor_set default_pset; | |||
| 59 | struct processor processor_array[NCPUS1]; | |||
| 60 | ||||
| 61 | queue_head_t all_psets; | |||
| 62 | int all_psets_count; | |||
| 63 | decl_simple_lock_data(, all_psets_lock); | |||
| 64 | ||||
| 65 | processor_t master_processor; | |||
| 66 | processor_t processor_ptr[NCPUS1]; | |||
| 67 | ||||
| 68 | /* | |||
| 69 | * Forward declarations. | |||
| 70 | */ | |||
| 71 | void quantum_set(processor_set_t); | |||
| 72 | void pset_init(processor_set_t); | |||
| 73 | void processor_init(processor_t, int); | |||
| 74 | ||||
| 75 | /* | |||
| 76 | * Bootstrap the processor/pset system so the scheduler can run. | |||
| 77 | */ | |||
| 78 | void pset_sys_bootstrap(void) | |||
| 79 | { | |||
| 80 | int i; | |||
| 81 | ||||
| 82 | pset_init(&default_pset); | |||
| 83 | default_pset.empty = FALSE((boolean_t) 0); | |||
| 84 | for (i = 0; i < NCPUS1; i++) { | |||
| 85 | /* | |||
| 86 | * Initialize processor data structures. | |||
| 87 | * Note that cpu_to_processor(i) is processor_ptr[i]. | |||
| 88 | */ | |||
| 89 | processor_ptr[i] = &processor_array[i]; | |||
| 90 | processor_init(processor_ptr[i], i); | |||
| 91 | } | |||
| 92 | master_processor = cpu_to_processor(master_cpu)(processor_ptr[master_cpu]); | |||
| 93 | queue_init(&all_psets)((&all_psets)->next = (&all_psets)->prev = & all_psets); | |||
| 94 | simple_lock_init(&all_psets_lock); | |||
| 95 | queue_enter(&all_psets, &default_pset, processor_set_t, all_psets){ register queue_entry_t prev; prev = (&all_psets)->prev ; if ((&all_psets) == prev) { (&all_psets)->next = (queue_entry_t) (&default_pset); } else { ((processor_set_t )prev)->all_psets.next = (queue_entry_t)(&default_pset ); } (&default_pset)->all_psets.prev = prev; (&default_pset )->all_psets.next = &all_psets; (&all_psets)->prev = (queue_entry_t) &default_pset; }; | |||
| 96 | all_psets_count = 1; | |||
| 97 | default_pset.active = TRUE((boolean_t) 1); | |||
| 98 | default_pset.empty = FALSE((boolean_t) 0); | |||
| 99 | ||||
| 100 | /* | |||
| 101 | * Note: the default_pset has a max_priority of BASEPRI_USER. | |||
| 102 | * Internal kernel threads override this in kernel_thread. | |||
| 103 | */ | |||
| 104 | } | |||
| 105 | ||||
| 106 | #if MACH_HOST0 | |||
| 107 | /* | |||
| 108 | * Rest of pset system initializations. | |||
| 109 | */ | |||
| 110 | void pset_sys_init(void) | |||
| 111 | { | |||
| 112 | int i; | |||
| 113 | processor_t processor; | |||
| 114 | ||||
| 115 | /* | |||
| 116 | * Allocate the cache for processor sets. | |||
| 117 | */ | |||
| 118 | kmem_cache_init(&pset_cache, "processor_set", | |||
| 119 | sizeof(struct processor_set), 0, NULL((void *) 0), NULL((void *) 0), NULL((void *) 0), 0); | |||
| 120 | ||||
| 121 | /* | |||
| 122 | * Give each processor a control port. | |||
| 123 | * The master processor already has one. | |||
| 124 | */ | |||
| 125 | for (i = 0; i < NCPUS1; i++) { | |||
| 126 | processor = cpu_to_processor(i)(processor_ptr[i]); | |||
| 127 | if (processor != master_processor && | |||
| 128 | machine_slot[i].is_cpu) | |||
| 129 | { | |||
| 130 | ipc_processor_init(processor); | |||
| 131 | } | |||
| 132 | } | |||
| 133 | } | |||
| 134 | #endif /* MACH_HOST */ | |||
| 135 | ||||
| 136 | /* | |||
| 137 | * Initialize the given processor_set structure. | |||
| 138 | */ | |||
| 139 | ||||
| 140 | void pset_init( | |||
| 141 | processor_set_t pset) | |||
| 142 | { | |||
| 143 | int i; | |||
| 144 | ||||
| 145 | simple_lock_init(&pset->runq.lock); | |||
| 146 | pset->runq.low = 0; | |||
| 147 | pset->runq.count = 0; | |||
| 148 | for (i = 0; i < NRQS50; i++) { | |||
| 149 | queue_init(&(pset->runq.runq[i]))((&(pset->runq.runq[i]))->next = (&(pset->runq .runq[i]))->prev = &(pset->runq.runq[i])); | |||
| 150 | } | |||
| 151 | queue_init(&pset->idle_queue)((&pset->idle_queue)->next = (&pset->idle_queue )->prev = &pset->idle_queue); | |||
| 152 | pset->idle_count = 0; | |||
| 153 | simple_lock_init(&pset->idle_lock); | |||
| 154 | queue_init(&pset->processors)((&pset->processors)->next = (&pset->processors )->prev = &pset->processors); | |||
| 155 | pset->processor_count = 0; | |||
| 156 | pset->empty = TRUE((boolean_t) 1); | |||
| 157 | queue_init(&pset->tasks)((&pset->tasks)->next = (&pset->tasks)->prev = &pset->tasks); | |||
| 158 | pset->task_count = 0; | |||
| 159 | queue_init(&pset->threads)((&pset->threads)->next = (&pset->threads)-> prev = &pset->threads); | |||
| 160 | pset->thread_count = 0; | |||
| 161 | pset->ref_count = 1; | |||
| 162 | simple_lock_init(&pset->ref_lock); | |||
| 163 | queue_init(&pset->all_psets)((&pset->all_psets)->next = (&pset->all_psets )->prev = &pset->all_psets); | |||
| 164 | pset->active = FALSE((boolean_t) 0); | |||
| 165 | simple_lock_init(&pset->lock); | |||
| 166 | pset->pset_self = IP_NULL((ipc_port_t) ((ipc_object_t) 0)); | |||
| 167 | pset->pset_name_self = IP_NULL((ipc_port_t) ((ipc_object_t) 0)); | |||
| 168 | pset->max_priority = BASEPRI_USER25; | |||
| 169 | #if MACH_FIXPRI1 | |||
| 170 | pset->policies = POLICY_TIMESHARE1; | |||
| 171 | #endif /* MACH_FIXPRI */ | |||
| 172 | pset->set_quantum = min_quantum; | |||
| 173 | #if NCPUS1 > 1 | |||
| 174 | pset->quantum_adj_index = 0; | |||
| 175 | simple_lock_init(&pset->quantum_adj_lock); | |||
| 176 | ||||
| 177 | for (i = 0; i <= NCPUS1; i++) { | |||
| 178 | pset->machine_quantum[i] = min_quantum; | |||
| 179 | } | |||
| 180 | #endif /* NCPUS > 1 */ | |||
| 181 | pset->mach_factor = 0; | |||
| 182 | pset->load_average = 0; | |||
| 183 | pset->sched_load = SCHED_SCALE128; /* i.e. 1 */ | |||
| 184 | } | |||
| 185 | ||||
| 186 | /* | |||
| 187 | * Initialize the given processor structure for the processor in | |||
| 188 | * the slot specified by slot_num. | |||
| 189 | */ | |||
| 190 | ||||
| 191 | void processor_init( | |||
| 192 | processor_t pr, | |||
| 193 | int slot_num) | |||
| 194 | { | |||
| 195 | int i; | |||
| 196 | ||||
| 197 | simple_lock_init(&pr->runq.lock); | |||
| 198 | pr->runq.low = 0; | |||
| 199 | pr->runq.count = 0; | |||
| 200 | for (i = 0; i < NRQS50; i++) { | |||
| 201 | queue_init(&(pr->runq.runq[i]))((&(pr->runq.runq[i]))->next = (&(pr->runq.runq [i]))->prev = &(pr->runq.runq[i])); | |||
| 202 | } | |||
| 203 | queue_init(&pr->processor_queue)((&pr->processor_queue)->next = (&pr->processor_queue )->prev = &pr->processor_queue); | |||
| 204 | pr->state = PROCESSOR_OFF_LINE0; | |||
| 205 | pr->next_thread = THREAD_NULL((thread_t) 0); | |||
| 206 | pr->idle_thread = THREAD_NULL((thread_t) 0); | |||
| 207 | pr->quantum = 0; | |||
| 208 | pr->first_quantum = FALSE((boolean_t) 0); | |||
| 209 | pr->last_quantum = 0; | |||
| 210 | pr->processor_set = PROCESSOR_SET_NULL((processor_set_t) 0); | |||
| 211 | pr->processor_set_next = PROCESSOR_SET_NULL((processor_set_t) 0); | |||
| 212 | queue_init(&pr->processors)((&pr->processors)->next = (&pr->processors) ->prev = &pr->processors); | |||
| 213 | simple_lock_init(&pr->lock); | |||
| 214 | pr->processor_self = IP_NULL((ipc_port_t) ((ipc_object_t) 0)); | |||
| 215 | pr->slot_num = slot_num; | |||
| 216 | } | |||
| 217 | ||||
| 218 | /* | |||
| 219 | * pset_remove_processor() removes a processor from a processor_set. | |||
| 220 | * It can only be called on the current processor. Caller must | |||
| 221 | * hold lock on current processor and processor set. | |||
| 222 | */ | |||
| 223 | ||||
| 224 | void pset_remove_processor( | |||
| 225 | processor_set_t pset, | |||
| 226 | processor_t processor) | |||
| 227 | { | |||
| 228 | if (pset != processor->processor_set) | |||
| 229 | panic("pset_remove_processor: wrong pset"); | |||
| 230 | ||||
| 231 | queue_remove(&pset->processors, processor, processor_t, processors){ queue_entry_t next, prev; next = (processor)->processors .next; prev = (processor)->processors.prev; if ((&pset ->processors) == next) (&pset->processors)->prev = prev; else ((processor_t)next)->processors.prev = prev; if ((&pset->processors) == prev) (&pset->processors )->next = next; else ((processor_t)prev)->processors.next = next; }; | |||
| 232 | processor->processor_set = PROCESSOR_SET_NULL((processor_set_t) 0); | |||
| 233 | pset->processor_count--; | |||
| 234 | quantum_set(pset); | |||
| 235 | } | |||
| 236 | ||||
| 237 | /* | |||
| 238 | * pset_add_processor() adds a processor to a processor_set. | |||
| 239 | * It can only be called on the current processor. Caller must | |||
| 240 | * hold lock on curent processor and on pset. No reference counting on | |||
| 241 | * processors. Processor reference to pset is implicit. | |||
| 242 | */ | |||
| 243 | ||||
| 244 | void pset_add_processor( | |||
| 245 | processor_set_t pset, | |||
| 246 | processor_t processor) | |||
| 247 | { | |||
| 248 | queue_enter(&pset->processors, processor, processor_t, processors){ register queue_entry_t prev; prev = (&pset->processors )->prev; if ((&pset->processors) == prev) { (&pset ->processors)->next = (queue_entry_t) (processor); } else { ((processor_t)prev)->processors.next = (queue_entry_t)( processor); } (processor)->processors.prev = prev; (processor )->processors.next = &pset->processors; (&pset-> processors)->prev = (queue_entry_t) processor; }; | |||
| 249 | processor->processor_set = pset; | |||
| 250 | pset->processor_count++; | |||
| 251 | quantum_set(pset); | |||
| 252 | } | |||
| 253 | ||||
| 254 | /* | |||
| 255 | * pset_remove_task() removes a task from a processor_set. | |||
| 256 | * Caller must hold locks on pset and task. Pset reference count | |||
| 257 | * is not decremented; caller must explicitly pset_deallocate. | |||
| 258 | */ | |||
| 259 | ||||
| 260 | void pset_remove_task( | |||
| 261 | processor_set_t pset, | |||
| 262 | task_t task) | |||
| 263 | { | |||
| 264 | if (pset != task->processor_set) | |||
| 265 | return; | |||
| 266 | ||||
| 267 | queue_remove(&pset->tasks, task, task_t, pset_tasks){ queue_entry_t next, prev; next = (task)->pset_tasks.next ; prev = (task)->pset_tasks.prev; if ((&pset->tasks ) == next) (&pset->tasks)->prev = prev; else ((task_t )next)->pset_tasks.prev = prev; if ((&pset->tasks) == prev) (&pset->tasks)->next = next; else ((task_t)prev )->pset_tasks.next = next; }; | |||
| 268 | task->processor_set = PROCESSOR_SET_NULL((processor_set_t) 0); | |||
| 269 | pset->task_count--; | |||
| 270 | } | |||
| 271 | ||||
| 272 | /* | |||
| 273 | * pset_add_task() adds a task to a processor_set. | |||
| 274 | * Caller must hold locks on pset and task. Pset references to | |||
| 275 | * tasks are implicit. | |||
| 276 | */ | |||
| 277 | ||||
| 278 | void pset_add_task( | |||
| 279 | processor_set_t pset, | |||
| 280 | task_t task) | |||
| 281 | { | |||
| 282 | queue_enter(&pset->tasks, task, task_t, pset_tasks){ register queue_entry_t prev; prev = (&pset->tasks)-> prev; if ((&pset->tasks) == prev) { (&pset->tasks )->next = (queue_entry_t) (task); } else { ((task_t)prev)-> pset_tasks.next = (queue_entry_t)(task); } (task)->pset_tasks .prev = prev; (task)->pset_tasks.next = &pset->tasks ; (&pset->tasks)->prev = (queue_entry_t) task; }; | |||
| 283 | task->processor_set = pset; | |||
| 284 | pset->task_count++; | |||
| 285 | } | |||
| 286 | ||||
| 287 | /* | |||
| 288 | * pset_remove_thread() removes a thread from a processor_set. | |||
| 289 | * Caller must hold locks on pset and thread. Pset reference count | |||
| 290 | * is not decremented; caller must explicitly pset_deallocate. | |||
| 291 | */ | |||
| 292 | ||||
| 293 | void pset_remove_thread( | |||
| 294 | processor_set_t pset, | |||
| 295 | thread_t thread) | |||
| 296 | { | |||
| 297 | queue_remove(&pset->threads, thread, thread_t, pset_threads){ queue_entry_t next, prev; next = (thread)->pset_threads. next; prev = (thread)->pset_threads.prev; if ((&pset-> threads) == next) (&pset->threads)->prev = prev; else ((thread_t)next)->pset_threads.prev = prev; if ((&pset ->threads) == prev) (&pset->threads)->next = next ; else ((thread_t)prev)->pset_threads.next = next; }; | |||
| 298 | thread->processor_set = PROCESSOR_SET_NULL((processor_set_t) 0); | |||
| 299 | pset->thread_count--; | |||
| 300 | } | |||
| 301 | ||||
| 302 | /* | |||
| 303 | * pset_add_thread() adds a thread to a processor_set. | |||
| 304 | * Caller must hold locks on pset and thread. Pset references to | |||
| 305 | * threads are implicit. | |||
| 306 | */ | |||
| 307 | ||||
| 308 | void pset_add_thread( | |||
| 309 | processor_set_t pset, | |||
| 310 | thread_t thread) | |||
| 311 | { | |||
| 312 | queue_enter(&pset->threads, thread, thread_t, pset_threads){ register queue_entry_t prev; prev = (&pset->threads) ->prev; if ((&pset->threads) == prev) { (&pset-> threads)->next = (queue_entry_t) (thread); } else { ((thread_t )prev)->pset_threads.next = (queue_entry_t)(thread); } (thread )->pset_threads.prev = prev; (thread)->pset_threads.next = &pset->threads; (&pset->threads)->prev = ( queue_entry_t) thread; }; | |||
| 313 | thread->processor_set = pset; | |||
| 314 | pset->thread_count++; | |||
| 315 | } | |||
| 316 | ||||
| 317 | /* | |||
| 318 | * thread_change_psets() changes the pset of a thread. Caller must | |||
| 319 | * hold locks on both psets and thread. The old pset must be | |||
| 320 | * explicitly pset_deallocat()'ed by caller. | |||
| 321 | */ | |||
| 322 | ||||
| 323 | void thread_change_psets( | |||
| 324 | thread_t thread, | |||
| 325 | processor_set_t old_pset, | |||
| 326 | processor_set_t new_pset) | |||
| 327 | { | |||
| 328 | queue_remove(&old_pset->threads, thread, thread_t, pset_threads){ queue_entry_t next, prev; next = (thread)->pset_threads. next; prev = (thread)->pset_threads.prev; if ((&old_pset ->threads) == next) (&old_pset->threads)->prev = prev; else ((thread_t)next)->pset_threads.prev = prev; if ((&old_pset->threads) == prev) (&old_pset->threads )->next = next; else ((thread_t)prev)->pset_threads.next = next; }; | |||
| 329 | old_pset->thread_count--; | |||
| 330 | queue_enter(&new_pset->threads, thread, thread_t, pset_threads){ register queue_entry_t prev; prev = (&new_pset->threads )->prev; if ((&new_pset->threads) == prev) { (& new_pset->threads)->next = (queue_entry_t) (thread); } else { ((thread_t)prev)->pset_threads.next = (queue_entry_t)(thread ); } (thread)->pset_threads.prev = prev; (thread)->pset_threads .next = &new_pset->threads; (&new_pset->threads )->prev = (queue_entry_t) thread; }; | |||
| 331 | thread->processor_set = new_pset; | |||
| 332 | new_pset->thread_count++; | |||
| 333 | } | |||
| 334 | ||||
| 335 | /* | |||
| 336 | * pset_deallocate: | |||
| 337 | * | |||
| 338 | * Remove one reference to the processor set. Destroy processor_set | |||
| 339 | * if this was the last reference. | |||
| 340 | */ | |||
| 341 | void pset_deallocate( | |||
| 342 | processor_set_t pset) | |||
| 343 | { | |||
| 344 | if (pset == PROCESSOR_SET_NULL((processor_set_t) 0)) | |||
| 345 | return; | |||
| 346 | ||||
| 347 | pset_ref_lock(pset); | |||
| 348 | if (--pset->ref_count > 0) { | |||
| 349 | pset_ref_unlock(pset); | |||
| 350 | return; | |||
| 351 | } | |||
| 352 | #if !MACH_HOST0 | |||
| 353 | panic("pset_deallocate: default_pset destroyed"); | |||
| 354 | #endif /* !MACH_HOST */ | |||
| 355 | ||||
| 356 | #if MACH_HOST0 | |||
| 357 | /* | |||
| 358 | * Reference count is zero, however the all_psets list | |||
| 359 | * holds an implicit reference and may make new ones. | |||
| 360 | * Its lock also dominates the pset lock. To check for this, | |||
| 361 | * temporarily restore one reference, and then lock the | |||
| 362 | * other structures in the right order. | |||
| 363 | */ | |||
| 364 | pset->ref_count = 1; | |||
| 365 | pset_ref_unlock(pset); | |||
| 366 | ||||
| 367 | simple_lock(&all_psets_lock); | |||
| 368 | pset_ref_lock(pset); | |||
| 369 | if (--pset->ref_count > 0) { | |||
| 370 | /* | |||
| 371 | * Made an extra reference. | |||
| 372 | */ | |||
| 373 | pset_ref_unlock(pset); | |||
| 374 | simple_unlock(&all_psets_lock); | |||
| 375 | return; | |||
| 376 | } | |||
| 377 | ||||
| 378 | /* | |||
| 379 | * Ok to destroy pset. Make a few paranoia checks. | |||
| 380 | */ | |||
| 381 | ||||
| 382 | if ((pset == &default_pset) || (pset->thread_count > 0) || | |||
| 383 | (pset->task_count > 0) || pset->processor_count > 0) { | |||
| 384 | panic("pset_deallocate: destroy default or active pset"); | |||
| 385 | } | |||
| 386 | /* | |||
| 387 | * Remove from all_psets queue. | |||
| 388 | */ | |||
| 389 | queue_remove(&all_psets, pset, processor_set_t, all_psets){ queue_entry_t next, prev; next = (pset)->all_psets.next; prev = (pset)->all_psets.prev; if ((&all_psets) == next ) (&all_psets)->prev = prev; else ((processor_set_t)next )->all_psets.prev = prev; if ((&all_psets) == prev) (& all_psets)->next = next; else ((processor_set_t)prev)-> all_psets.next = next; }; | |||
| 390 | all_psets_count--; | |||
| 391 | ||||
| 392 | pset_ref_unlock(pset); | |||
| 393 | simple_unlock(&all_psets_lock); | |||
| 394 | ||||
| 395 | /* | |||
| 396 | * That's it, free data structure. | |||
| 397 | */ | |||
| 398 | kmem_cache_free(&pset_cache, (vm_offset_t)pset); | |||
| 399 | #endif /* MACH_HOST */ | |||
| 400 | } | |||
| 401 | ||||
| 402 | /* | |||
| 403 | * pset_reference: | |||
| 404 | * | |||
| 405 | * Add one reference to the processor set. | |||
| 406 | */ | |||
| 407 | void pset_reference( | |||
| 408 | processor_set_t pset) | |||
| 409 | { | |||
| 410 | pset_ref_lock(pset); | |||
| 411 | pset->ref_count++; | |||
| 412 | pset_ref_unlock(pset); | |||
| 413 | } | |||
| 414 | ||||
| 415 | kern_return_t | |||
| 416 | processor_info( | |||
| 417 | processor_t processor, | |||
| 418 | int flavor, | |||
| 419 | host_t *host, | |||
| 420 | processor_info_t info, | |||
| 421 | natural_t *count) | |||
| 422 | { | |||
| 423 | int slot_num, state; | |||
| 424 | processor_basic_info_t basic_info; | |||
| 425 | ||||
| 426 | if (processor == PROCESSOR_NULL((processor_t) 0)) | |||
| 427 | return KERN_INVALID_ARGUMENT4; | |||
| 428 | ||||
| 429 | if (flavor != PROCESSOR_BASIC_INFO1 || | |||
| 430 | *count < PROCESSOR_BASIC_INFO_COUNT(sizeof(processor_basic_info_data_t)/sizeof(integer_t))) | |||
| 431 | return KERN_FAILURE5; | |||
| 432 | ||||
| 433 | basic_info = (processor_basic_info_t) info; | |||
| 434 | ||||
| 435 | slot_num = processor->slot_num; | |||
| 436 | basic_info->cpu_type = machine_slot[slot_num].cpu_type; | |||
| 437 | basic_info->cpu_subtype = machine_slot[slot_num].cpu_subtype; | |||
| 438 | state = processor->state; | |||
| 439 | if (state == PROCESSOR_SHUTDOWN5 || state == PROCESSOR_OFF_LINE0) | |||
| 440 | basic_info->running = FALSE((boolean_t) 0); | |||
| 441 | else | |||
| 442 | basic_info->running = TRUE((boolean_t) 1); | |||
| 443 | basic_info->slot_num = slot_num; | |||
| 444 | if (processor == master_processor) | |||
| 445 | basic_info->is_master = TRUE((boolean_t) 1); | |||
| 446 | else | |||
| 447 | basic_info->is_master = FALSE((boolean_t) 0); | |||
| 448 | ||||
| 449 | *count = PROCESSOR_BASIC_INFO_COUNT(sizeof(processor_basic_info_data_t)/sizeof(integer_t)); | |||
| 450 | *host = &realhost; | |||
| 451 | return KERN_SUCCESS0; | |||
| 452 | } | |||
| 453 | ||||
| 454 | kern_return_t processor_start( | |||
| 455 | processor_t processor) | |||
| 456 | { | |||
| 457 | if (processor == PROCESSOR_NULL((processor_t) 0)) | |||
| 458 | return KERN_INVALID_ARGUMENT4; | |||
| 459 | #if NCPUS1 > 1 | |||
| 460 | return cpu_start(processor->slot_num); | |||
| 461 | #else /* NCPUS > 1 */ | |||
| 462 | return KERN_FAILURE5; | |||
| 463 | #endif /* NCPUS > 1 */ | |||
| 464 | } | |||
| 465 | ||||
| 466 | kern_return_t processor_exit( | |||
| 467 | processor_t processor) | |||
| 468 | { | |||
| 469 | if (processor == PROCESSOR_NULL((processor_t) 0)) | |||
| 470 | return KERN_INVALID_ARGUMENT4; | |||
| 471 | ||||
| 472 | #if NCPUS1 > 1 | |||
| 473 | return processor_shutdown(processor); | |||
| 474 | #else /* NCPUS > 1 */ | |||
| 475 | return KERN_FAILURE5; | |||
| 476 | #endif /* NCPUS > 1 */ | |||
| 477 | } | |||
| 478 | ||||
| 479 | kern_return_t | |||
| 480 | processor_control( | |||
| 481 | processor_t processor, | |||
| 482 | processor_info_t info, | |||
| 483 | natural_t count) | |||
| 484 | { | |||
| 485 | if (processor == PROCESSOR_NULL((processor_t) 0)) | |||
| 486 | return KERN_INVALID_ARGUMENT4; | |||
| 487 | ||||
| 488 | #if NCPUS1 > 1 | |||
| 489 | return cpu_control(processor->slot_num, (int *)info, count); | |||
| 490 | #else /* NCPUS > 1 */ | |||
| 491 | return KERN_FAILURE5; | |||
| 492 | #endif /* NCPUS > 1 */ | |||
| 493 | } | |||
| 494 | ||||
| 495 | /* | |||
| 496 | * Precalculate the appropriate system quanta based on load. The | |||
| 497 | * index into machine_quantum is the number of threads on the | |||
| 498 | * processor set queue. It is limited to the number of processors in | |||
| 499 | * the set. | |||
| 500 | */ | |||
| 501 | ||||
| 502 | void quantum_set( | |||
| 503 | processor_set_t pset) | |||
| 504 | { | |||
| 505 | #if NCPUS1 > 1 | |||
| 506 | int i, ncpus; | |||
| 507 | ||||
| 508 | ncpus = pset->processor_count; | |||
| 509 | ||||
| 510 | for ( i=1 ; i <= ncpus ; i++) { | |||
| 511 | pset->machine_quantum[i] = | |||
| 512 | ((min_quantum * ncpus) + (i/2)) / i ; | |||
| 513 | } | |||
| 514 | pset->machine_quantum[0] = 2 * pset->machine_quantum[1]; | |||
| 515 | ||||
| 516 | i = ((pset->runq.count > pset->processor_count) ? | |||
| 517 | pset->processor_count : pset->runq.count); | |||
| 518 | pset->set_quantum = pset->machine_quantum[i]; | |||
| 519 | #else /* NCPUS > 1 */ | |||
| 520 | default_pset.set_quantum = min_quantum; | |||
| 521 | #endif /* NCPUS > 1 */ | |||
| 522 | } | |||
| 523 | ||||
| 524 | #if MACH_HOST0 | |||
| 525 | /* | |||
| 526 | * processor_set_create: | |||
| 527 | * | |||
| 528 | * Create and return a new processor set. | |||
| 529 | */ | |||
| 530 | ||||
| 531 | kern_return_t | |||
| 532 | processor_set_create( | |||
| 533 | host_t host, | |||
| 534 | processor_set_t *new_set, | |||
| 535 | processor_set_t *new_name) | |||
| 536 | { | |||
| 537 | processor_set_t pset; | |||
| 538 | ||||
| 539 | if (host == HOST_NULL((host_t)0)) | |||
| 540 | return KERN_INVALID_ARGUMENT4; | |||
| 541 | ||||
| 542 | pset = (processor_set_t) kmem_cache_alloc(&pset_cache); | |||
| 543 | pset_init(pset); | |||
| 544 | pset_reference(pset); /* for new_set out argument */ | |||
| 545 | pset_reference(pset); /* for new_name out argument */ | |||
| 546 | ipc_pset_init(pset); | |||
| 547 | pset->active = TRUE((boolean_t) 1); | |||
| 548 | ||||
| 549 | simple_lock(&all_psets_lock); | |||
| 550 | queue_enter(&all_psets, pset, processor_set_t, all_psets){ register queue_entry_t prev; prev = (&all_psets)->prev ; if ((&all_psets) == prev) { (&all_psets)->next = (queue_entry_t) (pset); } else { ((processor_set_t)prev)-> all_psets.next = (queue_entry_t)(pset); } (pset)->all_psets .prev = prev; (pset)->all_psets.next = &all_psets; (& all_psets)->prev = (queue_entry_t) pset; }; | |||
| 551 | all_psets_count++; | |||
| 552 | simple_unlock(&all_psets_lock); | |||
| 553 | ||||
| 554 | ipc_pset_enable(pset); | |||
| 555 | ||||
| 556 | *new_set = pset; | |||
| 557 | *new_name = pset; | |||
| 558 | return KERN_SUCCESS0; | |||
| 559 | } | |||
| 560 | ||||
| 561 | /* | |||
| 562 | * processor_set_destroy: | |||
| 563 | * | |||
| 564 | * destroy a processor set. Any tasks, threads or processors | |||
| 565 | * currently assigned to it are reassigned to the default pset. | |||
| 566 | */ | |||
| 567 | kern_return_t processor_set_destroy( | |||
| 568 | processor_set_t pset) | |||
| 569 | { | |||
| 570 | queue_entry_t elem; | |||
| 571 | queue_head_t *list; | |||
| 572 | ||||
| 573 | if (pset == PROCESSOR_SET_NULL((processor_set_t) 0) || pset == &default_pset) | |||
| 574 | return KERN_INVALID_ARGUMENT4; | |||
| 575 | ||||
| 576 | /* | |||
| 577 | * Handle multiple termination race. First one through sets | |||
| 578 | * active to FALSE and disables ipc access. | |||
| 579 | */ | |||
| 580 | pset_lock(pset); | |||
| 581 | if (!(pset->active)) { | |||
| 582 | pset_unlock(pset); | |||
| 583 | return KERN_FAILURE5; | |||
| 584 | } | |||
| 585 | ||||
| 586 | pset->active = FALSE((boolean_t) 0); | |||
| 587 | ipc_pset_disable(pset); | |||
| 588 | ||||
| 589 | ||||
| 590 | /* | |||
| 591 | * Now reassign everything in this set to the default set. | |||
| 592 | */ | |||
| 593 | ||||
| 594 | if (pset->task_count > 0) { | |||
| 595 | list = &pset->tasks; | |||
| 596 | while (!queue_empty(list)(((list)) == (((list)->next)))) { | |||
| 597 | elem = queue_first(list)((list)->next); | |||
| 598 | task_reference((task_t) elem); | |||
| 599 | pset_unlock(pset); | |||
| 600 | task_assign((task_t) elem, &default_pset, FALSE((boolean_t) 0)); | |||
| 601 | task_deallocate((task_t) elem); | |||
| 602 | pset_lock(pset); | |||
| 603 | } | |||
| 604 | } | |||
| 605 | ||||
| 606 | if (pset->thread_count > 0) { | |||
| 607 | list = &pset->threads; | |||
| 608 | while (!queue_empty(list)(((list)) == (((list)->next)))) { | |||
| 609 | elem = queue_first(list)((list)->next); | |||
| 610 | thread_reference((thread_t) elem); | |||
| 611 | pset_unlock(pset); | |||
| 612 | thread_assign((thread_t) elem, &default_pset); | |||
| 613 | thread_deallocate((thread_t) elem); | |||
| 614 | pset_lock(pset); | |||
| 615 | } | |||
| 616 | } | |||
| 617 | ||||
| 618 | if (pset->processor_count > 0) { | |||
| 619 | list = &pset->processors; | |||
| 620 | while(!queue_empty(list)(((list)) == (((list)->next)))) { | |||
| 621 | elem = queue_first(list)((list)->next); | |||
| 622 | pset_unlock(pset); | |||
| 623 | processor_assign((processor_t) elem, &default_pset, TRUE((boolean_t) 1)); | |||
| 624 | pset_lock(pset); | |||
| 625 | } | |||
| 626 | } | |||
| 627 | ||||
| 628 | pset_unlock(pset); | |||
| 629 | ||||
| 630 | /* | |||
| 631 | * Destroy ipc state. | |||
| 632 | */ | |||
| 633 | ipc_pset_terminate(pset); | |||
| 634 | ||||
| 635 | /* | |||
| 636 | * Deallocate pset's reference to itself. | |||
| 637 | */ | |||
| 638 | pset_deallocate(pset); | |||
| 639 | return KERN_SUCCESS0; | |||
| 640 | } | |||
| 641 | ||||
| 642 | #else /* MACH_HOST */ | |||
| 643 | ||||
| 644 | kern_return_t | |||
| 645 | processor_set_create( | |||
| 646 | host_t host, | |||
| 647 | processor_set_t *new_set, | |||
| 648 | processor_set_t *new_name) | |||
| 649 | { | |||
| 650 | return KERN_FAILURE5; | |||
| 651 | } | |||
| 652 | ||||
| 653 | kern_return_t processor_set_destroy( | |||
| 654 | processor_set_t pset) | |||
| 655 | { | |||
| 656 | return KERN_FAILURE5; | |||
| 657 | } | |||
| 658 | ||||
| 659 | #endif /* MACH_HOST */ | |||
| 660 | ||||
| 661 | kern_return_t | |||
| 662 | processor_get_assignment( | |||
| 663 | processor_t processor, | |||
| 664 | processor_set_t *pset) | |||
| 665 | { | |||
| 666 | int state; | |||
| 667 | ||||
| 668 | state = processor->state; | |||
| 669 | if (state == PROCESSOR_SHUTDOWN5 || state == PROCESSOR_OFF_LINE0) | |||
| 670 | return KERN_FAILURE5; | |||
| 671 | ||||
| 672 | *pset = processor->processor_set; | |||
| 673 | pset_reference(*pset); | |||
| 674 | return KERN_SUCCESS0; | |||
| 675 | } | |||
| 676 | ||||
| 677 | kern_return_t | |||
| 678 | processor_set_info( | |||
| 679 | processor_set_t pset, | |||
| 680 | int flavor, | |||
| 681 | host_t *host, | |||
| 682 | processor_set_info_t info, | |||
| 683 | natural_t *count) | |||
| 684 | { | |||
| 685 | if (pset == PROCESSOR_SET_NULL((processor_set_t) 0)) | |||
| 686 | return KERN_INVALID_ARGUMENT4; | |||
| 687 | ||||
| 688 | if (flavor == PROCESSOR_SET_BASIC_INFO1) { | |||
| 689 | processor_set_basic_info_t basic_info; | |||
| 690 | ||||
| 691 | if (*count < PROCESSOR_SET_BASIC_INFO_COUNT(sizeof(processor_set_basic_info_data_t)/sizeof(integer_t))) | |||
| 692 | return KERN_FAILURE5; | |||
| 693 | ||||
| 694 | basic_info = (processor_set_basic_info_t) info; | |||
| 695 | ||||
| 696 | pset_lock(pset); | |||
| 697 | basic_info->processor_count = pset->processor_count; | |||
| 698 | basic_info->task_count = pset->task_count; | |||
| 699 | basic_info->thread_count = pset->thread_count; | |||
| 700 | basic_info->mach_factor = pset->mach_factor; | |||
| 701 | basic_info->load_average = pset->load_average; | |||
| 702 | pset_unlock(pset); | |||
| 703 | ||||
| 704 | *count = PROCESSOR_SET_BASIC_INFO_COUNT(sizeof(processor_set_basic_info_data_t)/sizeof(integer_t)); | |||
| 705 | *host = &realhost; | |||
| 706 | return KERN_SUCCESS0; | |||
| 707 | } | |||
| 708 | else if (flavor == PROCESSOR_SET_SCHED_INFO2) { | |||
| 709 | processor_set_sched_info_t sched_info; | |||
| 710 | ||||
| 711 | if (*count < PROCESSOR_SET_SCHED_INFO_COUNT(sizeof(processor_set_sched_info_data_t)/sizeof(integer_t))) | |||
| 712 | return KERN_FAILURE5; | |||
| 713 | ||||
| 714 | sched_info = (processor_set_sched_info_t) info; | |||
| 715 | ||||
| 716 | pset_lock(pset); | |||
| 717 | #if MACH_FIXPRI1 | |||
| 718 | sched_info->policies = pset->policies; | |||
| 719 | #else /* MACH_FIXPRI */ | |||
| 720 | sched_info->policies = POLICY_TIMESHARE1; | |||
| 721 | #endif /* MACH_FIXPRI */ | |||
| 722 | sched_info->max_priority = pset->max_priority; | |||
| 723 | pset_unlock(pset); | |||
| 724 | ||||
| 725 | *count = PROCESSOR_SET_SCHED_INFO_COUNT(sizeof(processor_set_sched_info_data_t)/sizeof(integer_t)); | |||
| 726 | *host = &realhost; | |||
| 727 | return KERN_SUCCESS0; | |||
| 728 | } | |||
| 729 | ||||
| 730 | *host = HOST_NULL((host_t)0); | |||
| 731 | return KERN_INVALID_ARGUMENT4; | |||
| 732 | } | |||
| 733 | ||||
| 734 | /* | |||
| 735 | * processor_set_max_priority: | |||
| 736 | * | |||
| 737 | * Specify max priority permitted on processor set. This affects | |||
| 738 | * newly created and assigned threads. Optionally change existing | |||
| 739 | * ones. | |||
| 740 | */ | |||
| 741 | kern_return_t | |||
| 742 | processor_set_max_priority( | |||
| 743 | processor_set_t pset, | |||
| 744 | int max_priority, | |||
| 745 | boolean_t change_threads) | |||
| 746 | { | |||
| 747 | if (pset == PROCESSOR_SET_NULL((processor_set_t) 0) || invalid_pri(max_priority)(((max_priority) < 0) || ((max_priority) >= 50))) | |||
| 748 | return KERN_INVALID_ARGUMENT4; | |||
| 749 | ||||
| 750 | pset_lock(pset); | |||
| 751 | pset->max_priority = max_priority; | |||
| 752 | ||||
| 753 | if (change_threads) { | |||
| 754 | queue_head_t *list; | |||
| 755 | thread_t thread; | |||
| 756 | ||||
| 757 | list = &pset->threads; | |||
| 758 | queue_iterate(list, thread, thread_t, pset_threads)for ((thread) = (thread_t) ((list)->next); !(((list)) == ( (queue_entry_t)(thread))); (thread) = (thread_t) ((&(thread )->pset_threads)->next)) { | |||
| 759 | if (thread->max_priority < max_priority) | |||
| 760 | thread_max_priority(thread, pset, max_priority); | |||
| 761 | } | |||
| 762 | } | |||
| 763 | ||||
| 764 | pset_unlock(pset); | |||
| 765 | ||||
| 766 | return KERN_SUCCESS0; | |||
| 767 | } | |||
| 768 | ||||
| 769 | /* | |||
| 770 | * processor_set_policy_enable: | |||
| 771 | * | |||
| 772 | * Allow indicated policy on processor set. | |||
| 773 | */ | |||
| 774 | ||||
| 775 | kern_return_t | |||
| 776 | processor_set_policy_enable( | |||
| 777 | processor_set_t pset, | |||
| 778 | int policy) | |||
| 779 | { | |||
| 780 | if ((pset == PROCESSOR_SET_NULL((processor_set_t) 0)) || invalid_policy(policy)(((policy) <= 0) || ((policy) > 2))) | |||
| 781 | return KERN_INVALID_ARGUMENT4; | |||
| 782 | ||||
| 783 | #if MACH_FIXPRI1 | |||
| 784 | pset_lock(pset); | |||
| 785 | pset->policies |= policy; | |||
| 786 | pset_unlock(pset); | |||
| 787 | ||||
| 788 | return KERN_SUCCESS0; | |||
| 789 | #else /* MACH_FIXPRI */ | |||
| 790 | if (policy == POLICY_TIMESHARE1) | |||
| 791 | return KERN_SUCCESS0; | |||
| 792 | else | |||
| 793 | return KERN_FAILURE5; | |||
| 794 | #endif /* MACH_FIXPRI */ | |||
| 795 | } | |||
| 796 | ||||
| 797 | /* | |||
| 798 | * processor_set_policy_disable: | |||
| 799 | * | |||
| 800 | * Forbid indicated policy on processor set. Time sharing cannot | |||
| 801 | * be forbidden. | |||
| 802 | */ | |||
| 803 | ||||
| 804 | kern_return_t | |||
| 805 | processor_set_policy_disable( | |||
| 806 | processor_set_t pset, | |||
| 807 | int policy, | |||
| 808 | boolean_t change_threads) | |||
| 809 | { | |||
| 810 | if ((pset == PROCESSOR_SET_NULL((processor_set_t) 0)) || policy == POLICY_TIMESHARE1 || | |||
| 811 | invalid_policy(policy)(((policy) <= 0) || ((policy) > 2))) | |||
| 812 | return KERN_INVALID_ARGUMENT4; | |||
| 813 | ||||
| 814 | #if MACH_FIXPRI1 | |||
| 815 | pset_lock(pset); | |||
| 816 | ||||
| 817 | /* | |||
| 818 | * Check if policy enabled. Disable if so, then handle | |||
| 819 | * change_threads. | |||
| 820 | */ | |||
| 821 | if (pset->policies & policy) { | |||
| 822 | pset->policies &= ~policy; | |||
| 823 | ||||
| 824 | if (change_threads) { | |||
| 825 | queue_head_t *list; | |||
| 826 | thread_t thread; | |||
| 827 | ||||
| 828 | list = &pset->threads; | |||
| 829 | queue_iterate(list, thread, thread_t, pset_threads)for ((thread) = (thread_t) ((list)->next); !(((list)) == ( (queue_entry_t)(thread))); (thread) = (thread_t) ((&(thread )->pset_threads)->next)) { | |||
| 830 | if (thread->policy == policy) | |||
| 831 | thread_policy(thread, POLICY_TIMESHARE1, 0); | |||
| 832 | } | |||
| 833 | } | |||
| 834 | } | |||
| 835 | pset_unlock(pset); | |||
| 836 | #endif /* MACH_FIXPRI */ | |||
| 837 | ||||
| 838 | return KERN_SUCCESS0; | |||
| 839 | } | |||
| 840 | ||||
| 841 | #define THING_TASK0 0 | |||
| 842 | #define THING_THREAD1 1 | |||
| 843 | ||||
| 844 | /* | |||
| 845 | * processor_set_things: | |||
| 846 | * | |||
| 847 | * Common internals for processor_set_{threads,tasks} | |||
| 848 | */ | |||
| 849 | kern_return_t | |||
| 850 | processor_set_things( | |||
| 851 | processor_set_t pset, | |||
| 852 | mach_port_t **thing_list, | |||
| 853 | natural_t *count, | |||
| 854 | int type) | |||
| 855 | { | |||
| 856 | unsigned int actual; /* this many things */ | |||
| 857 | int i; | |||
| 858 | ||||
| 859 | vm_size_t size, size_needed; | |||
| 860 | vm_offset_t addr; | |||
| 861 | ||||
| 862 | if (pset == PROCESSOR_SET_NULL((processor_set_t) 0)) | |||
| 863 | return KERN_INVALID_ARGUMENT4; | |||
| 864 | ||||
| 865 | size = 0; addr = 0; | |||
| 866 | ||||
| 867 | for (;;) { | |||
| 868 | pset_lock(pset); | |||
| 869 | if (!pset->active) { | |||
| 870 | pset_unlock(pset); | |||
| 871 | return KERN_FAILURE5; | |||
| 872 | } | |||
| 873 | ||||
| 874 | if (type == THING_TASK0) | |||
| 875 | actual = pset->task_count; | |||
| 876 | else | |||
| 877 | actual = pset->thread_count; | |||
| 878 | ||||
| 879 | /* do we have the memory we need? */ | |||
| 880 | ||||
| 881 | size_needed = actual * sizeof(mach_port_t); | |||
| 882 | if (size_needed <= size) | |||
| 883 | break; | |||
| 884 | ||||
| 885 | /* unlock the pset and allocate more memory */ | |||
| 886 | pset_unlock(pset); | |||
| 887 | ||||
| 888 | if (size != 0) | |||
| 889 | kfree(addr, size); | |||
| 890 | ||||
| 891 | assert(size_needed > 0)({ if (!(size_needed > 0)) Assert("size_needed > 0", "../kern/processor.c" , 891); }); | |||
| 892 | size = size_needed; | |||
| 893 | ||||
| 894 | addr = kalloc(size); | |||
| 895 | if (addr == 0) | |||
| 896 | return KERN_RESOURCE_SHORTAGE6; | |||
| 897 | } | |||
| 898 | ||||
| 899 | /* OK, have memory and the processor_set is locked & active */ | |||
| 900 | ||||
| 901 | switch (type) { | |||
| 902 | case THING_TASK0: { | |||
| 903 | task_t *tasks = (task_t *) addr; | |||
| 904 | task_t task; | |||
| 905 | ||||
| 906 | for (i = 0, task = (task_t) queue_first(&pset->tasks)((&pset->tasks)->next); | |||
| 907 | i < actual; | |||
| 908 | i++, task = (task_t) queue_next(&task->pset_tasks)((&task->pset_tasks)->next)) { | |||
| 909 | /* take ref for convert_task_to_port */ | |||
| 910 | task_reference(task); | |||
| 911 | tasks[i] = task; | |||
| 912 | } | |||
| 913 | assert(queue_end(&pset->tasks, (queue_entry_t) task))({ if (!(((&pset->tasks) == ((queue_entry_t) task)))) Assert ("queue_end(&pset->tasks, (queue_entry_t) task)", "../kern/processor.c" , 913); }); | |||
| 914 | break; | |||
| 915 | } | |||
| 916 | ||||
| 917 | case THING_THREAD1: { | |||
| 918 | thread_t *threads = (thread_t *) addr; | |||
| 919 | thread_t thread; | |||
| 920 | ||||
| 921 | for (i = 0, thread = (thread_t) queue_first(&pset->threads)((&pset->threads)->next); | |||
| 922 | i < actual; | |||
| 923 | i++, | |||
| 924 | thread = (thread_t) queue_next(&thread->pset_threads)((&thread->pset_threads)->next)) { | |||
| 925 | /* take ref for convert_thread_to_port */ | |||
| 926 | thread_reference(thread); | |||
| 927 | threads[i] = thread; | |||
| ||||
| 928 | } | |||
| 929 | assert(queue_end(&pset->threads, (queue_entry_t) thread))({ if (!(((&pset->threads) == ((queue_entry_t) thread) ))) Assert("queue_end(&pset->threads, (queue_entry_t) thread)" , "../kern/processor.c", 929); }); | |||
| 930 | break; | |||
| 931 | } | |||
| 932 | } | |||
| 933 | ||||
| 934 | /* can unlock processor set now that we have the task/thread refs */ | |||
| 935 | pset_unlock(pset); | |||
| 936 | ||||
| 937 | if (actual == 0) { | |||
| 938 | /* no things, so return null pointer and deallocate memory */ | |||
| 939 | *thing_list = 0; | |||
| 940 | *count = 0; | |||
| 941 | ||||
| 942 | if (size != 0) | |||
| 943 | kfree(addr, size); | |||
| 944 | } else { | |||
| 945 | /* if we allocated too much, must copy */ | |||
| 946 | ||||
| 947 | if (size_needed < size) { | |||
| 948 | vm_offset_t newaddr; | |||
| 949 | ||||
| 950 | newaddr = kalloc(size_needed); | |||
| 951 | if (newaddr == 0) { | |||
| 952 | switch (type) { | |||
| 953 | case THING_TASK0: { | |||
| 954 | task_t *tasks = (task_t *) addr; | |||
| 955 | ||||
| 956 | for (i = 0; i < actual; i++) | |||
| 957 | task_deallocate(tasks[i]); | |||
| 958 | break; | |||
| 959 | } | |||
| 960 | ||||
| 961 | case THING_THREAD1: { | |||
| 962 | thread_t *threads = (thread_t *) addr; | |||
| 963 | ||||
| 964 | for (i = 0; i < actual; i++) | |||
| 965 | thread_deallocate(threads[i]); | |||
| 966 | break; | |||
| 967 | } | |||
| 968 | } | |||
| 969 | kfree(addr, size); | |||
| 970 | return KERN_RESOURCE_SHORTAGE6; | |||
| 971 | } | |||
| 972 | ||||
| 973 | memcpy((void *) newaddr, (void *) addr, size_needed); | |||
| 974 | kfree(addr, size); | |||
| 975 | addr = newaddr; | |||
| 976 | } | |||
| 977 | ||||
| 978 | *thing_list = (mach_port_t *) addr; | |||
| 979 | *count = actual; | |||
| 980 | ||||
| 981 | /* do the conversion that Mig should handle */ | |||
| 982 | ||||
| 983 | switch (type) { | |||
| 984 | case THING_TASK0: { | |||
| 985 | task_t *tasks = (task_t *) addr; | |||
| 986 | ||||
| 987 | for (i = 0; i < actual; i++) | |||
| 988 | ((mach_port_t *) tasks)[i] = | |||
| 989 | (mach_port_t)convert_task_to_port(tasks[i]); | |||
| 990 | break; | |||
| 991 | } | |||
| 992 | ||||
| 993 | case THING_THREAD1: { | |||
| 994 | thread_t *threads = (thread_t *) addr; | |||
| 995 | ||||
| 996 | for (i = 0; i < actual; i++) | |||
| 997 | ((mach_port_t *) threads)[i] = | |||
| 998 | (mach_port_t)convert_thread_to_port(threads[i]); | |||
| 999 | break; | |||
| 1000 | } | |||
| 1001 | } | |||
| 1002 | } | |||
| 1003 | ||||
| 1004 | return KERN_SUCCESS0; | |||
| 1005 | } | |||
| 1006 | ||||
| 1007 | ||||
| 1008 | /* | |||
| 1009 | * processor_set_tasks: | |||
| 1010 | * | |||
| 1011 | * List all tasks in the processor set. | |||
| 1012 | */ | |||
| 1013 | kern_return_t | |||
| 1014 | processor_set_tasks( | |||
| 1015 | processor_set_t pset, | |||
| 1016 | task_array_t *task_list, | |||
| 1017 | natural_t *count) | |||
| 1018 | { | |||
| 1019 | return processor_set_things(pset, task_list, count, THING_TASK0); | |||
| 1020 | } | |||
| 1021 | ||||
| 1022 | /* | |||
| 1023 | * processor_set_threads: | |||
| 1024 | * | |||
| 1025 | * List all threads in the processor set. | |||
| 1026 | */ | |||
| 1027 | kern_return_t | |||
| 1028 | processor_set_threads( | |||
| 1029 | processor_set_t pset, | |||
| 1030 | thread_array_t *thread_list, | |||
| 1031 | natural_t *count) | |||
| 1032 | { | |||
| 1033 | return processor_set_things(pset, thread_list, count, THING_THREAD1); | |||
| ||||
| 1034 | } |