1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
|
/*
* Mach Operating System
* Copyright (c) 1993-1987 Carnegie Mellon University
* All Rights Reserved.
*
* Permission to use, copy, modify and distribute this software and its
* documentation is hereby granted, provided that both the copyright
* notice and this permission notice appear in all copies of the
* software, derivative works or modified versions, and any portions
* thereof, and that both notices appear in supporting documentation.
*
* CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS"
* CONDITION. CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND FOR
* ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE.
*
* Carnegie Mellon requests users of this software to return to
*
* Software Distribution Coordinator or Software.Distribution@CS.CMU.EDU
* School of Computer Science
* Carnegie Mellon University
* Pittsburgh PA 15213-3890
*
* any improvements or extensions that they make and grant Carnegie Mellon
* the rights to redistribute these changes.
*/
/*
* File: kern/kalloc.c
* Author: Avadis Tevanian, Jr.
* Date: 1985
*
* General kernel memory allocator. This allocator is designed
* to be used by the kernel to manage dynamic memory fast.
*/
#include <mach.h>
#include <cthreads.h> /* for spin locks */
#include <malloc.h> /* for malloc_hook/free_hook */
#include "wiring.h"
static void init_hook (void);
static void *malloc_hook (size_t size, const void *caller);
static void free_hook (void *ptr, const void *caller);
/* GNU libc 2.14 defines this macro to declare hook variables as volatile.
Define it as empty for older libc versions. */
#ifndef __MALLOC_HOOK_VOLATILE
# define __MALLOC_HOOK_VOLATILE
#endif
void (*__MALLOC_HOOK_VOLATILE __malloc_initialize_hook) (void) = init_hook;
#define DEBUG
/*
* All allocations of size less than kalloc_max are rounded to the
* next highest power of 2.
*/
vm_size_t kalloc_max; /* max before we use vm_allocate */
#define MINSIZE 4 /* minimum allocation size */
struct free_list {
spin_lock_t lock;
vm_offset_t head; /* head of free list */
#ifdef DEBUG
int count;
#endif /*DEBUG*/
};
#define KLIST_MAX 13
/* sizes: 4, 8, 16, 32, 64,
128, 256, 512, 1024,
2048, 4096, 8192, 16384 */
struct free_list kfree_list[KLIST_MAX];
spin_lock_t kget_space_lock;
vm_offset_t kalloc_next_space = 0;
vm_offset_t kalloc_end_of_space = 0;
vm_size_t kalloc_wasted_space = 0;
boolean_t kalloc_initialized = FALSE;
/*
* Initialize the memory allocator. This should be called only
* once on a system wide basis (i.e. first processor to get here
* does the initialization).
*
* This initializes all of the zones.
*/
void kalloc_init(void)
{
vm_offset_t min, max;
vm_size_t size;
register int i;
/*
* Support free lists for items up to vm_page_size or
* 16Kbytes, whichever is less.
*/
if (vm_page_size > 16*1024)
kalloc_max = 16*1024;
else
kalloc_max = vm_page_size;
for (i = 0; i < KLIST_MAX; i++) {
spin_lock_init(&kfree_list[i].lock);
kfree_list[i].head = 0;
}
spin_lock_init(&kget_space_lock);
/*
* Do not allocate memory at address 0.
*/
kalloc_next_space = vm_page_size;
kalloc_end_of_space = vm_page_size;
}
/*
* Contiguous space allocator for items of less than a page size.
*/
vm_offset_t kget_space(vm_offset_t size)
{
vm_size_t space_to_add;
vm_offset_t new_space = 0;
vm_offset_t addr;
spin_lock(&kget_space_lock);
while (kalloc_next_space + size > kalloc_end_of_space) {
/*
* Add at least one page to allocation area.
*/
space_to_add = round_page(size);
if (new_space == 0) {
/*
* Unlock and allocate memory.
* Try to make it contiguous with the last
* allocation area.
*/
spin_unlock(&kget_space_lock);
new_space = kalloc_end_of_space;
if (vm_map(mach_task_self(),
&new_space, space_to_add, (vm_offset_t) 0, TRUE,
MEMORY_OBJECT_NULL, (vm_offset_t) 0, FALSE,
VM_PROT_DEFAULT, VM_PROT_ALL, VM_INHERIT_DEFAULT)
!= KERN_SUCCESS)
return 0;
wire_memory(new_space, space_to_add,
VM_PROT_READ|VM_PROT_WRITE);
spin_lock(&kget_space_lock);
continue;
}
/*
* Memory was allocated in a previous iteration.
* Check whether the new region is contiguous with the
* old one.
*/
if (new_space != kalloc_end_of_space) {
/*
* Throw away the remainder of the old space,
* and start a new one.
*/
kalloc_wasted_space +=
kalloc_end_of_space - kalloc_next_space;
kalloc_next_space = new_space;
}
kalloc_end_of_space = new_space + space_to_add;
new_space = 0;
}
addr = kalloc_next_space;
kalloc_next_space += size;
spin_unlock(&kget_space_lock);
if (new_space != 0)
(void) vm_deallocate(mach_task_self(), new_space, space_to_add);
return addr;
}
void *kalloc(vm_size_t size)
{
register vm_size_t allocsize;
vm_offset_t addr;
register struct free_list *fl;
if (!kalloc_initialized) {
kalloc_init();
kalloc_initialized = TRUE;
}
/* compute the size of the block that we will actually allocate */
allocsize = size;
if (size < kalloc_max) {
allocsize = MINSIZE;
fl = kfree_list;
while (allocsize < size) {
allocsize <<= 1;
fl++;
}
}
/*
* If our size is still small enough, check the queue for that size
* and allocate.
*/
if (allocsize < kalloc_max) {
spin_lock(&fl->lock);
if ((addr = fl->head) != 0) {
fl->head = *(vm_offset_t *)addr;
#ifdef DEBUG
fl->count--;
#endif
spin_unlock(&fl->lock);
}
else {
spin_unlock(&fl->lock);
addr = kget_space(allocsize);
}
}
else {
if (vm_allocate(mach_task_self(), &addr, allocsize, TRUE)
!= KERN_SUCCESS)
addr = 0;
}
return (void *) addr;
}
void
kfree( void *data,
vm_size_t size)
{
register vm_size_t freesize;
register struct free_list *fl;
freesize = size;
if (size < kalloc_max) {
freesize = MINSIZE;
fl = kfree_list;
while (freesize < size) {
freesize <<= 1;
fl++;
}
}
if (freesize < kalloc_max) {
spin_lock(&fl->lock);
*(vm_offset_t *)data = fl->head;
fl->head = (vm_offset_t) data;
#ifdef DEBUG
fl->count++;
#endif
spin_unlock(&fl->lock);
}
else {
(void) vm_deallocate(mach_task_self(), (vm_offset_t)data, freesize);
}
}
static void
init_hook (void)
{
__malloc_hook = malloc_hook;
__free_hook = free_hook;
}
static void *
malloc_hook (size_t size, const void *caller)
{
return (void *) kalloc ((vm_size_t) size);
}
static void
free_hook (void *ptr, const void *caller)
{
/* Just ignore harmless attempts at cleanliness. */
/* panic("free not implemented"); */
}
void malloc_fork_prepare()
{
}
void malloc_fork_parent()
{
}
void malloc_fork_child()
{
}
|