summaryrefslogtreecommitdiff
path: root/kern
diff options
context:
space:
mode:
Diffstat (limited to 'kern')
-rw-r--r--kern/act.c2
-rw-r--r--kern/kalloc.c2
-rw-r--r--kern/processor.c2
-rw-r--r--kern/task.c2
-rw-r--r--kern/thread.c2
-rw-r--r--kern/zalloc.c29
-rw-r--r--kern/zalloc.h5
7 files changed, 30 insertions, 14 deletions
diff --git a/kern/act.c b/kern/act.c
index ffbce28..d0a03a3 100644
--- a/kern/act.c
+++ b/kern/act.c
@@ -69,7 +69,7 @@ global_act_init()
{
#ifndef ACT_STATIC_KLUDGE
act_zone = zinit(
- sizeof(struct Act),
+ sizeof(struct Act), 0,
ACT_MAX * sizeof(struct Act), /* XXX */
ACT_CHUNK * sizeof(struct Act),
0, "activations");
diff --git a/kern/kalloc.c b/kern/kalloc.c
index b24eaa3..4460d59 100644
--- a/kern/kalloc.c
+++ b/kern/kalloc.c
@@ -145,7 +145,7 @@ void kalloc_init()
if (size == MINSIZE) {
first_k_zone = i;
}
- k_zone[i] = zinit(size, k_zone_max[i] * size, size,
+ k_zone[i] = zinit(size, 0, k_zone_max[i] * size, size,
size >= PAGE_SIZE ? ZONE_COLLECTABLE : 0,
k_zone_name[i]);
}
diff --git a/kern/processor.c b/kern/processor.c
index 4035ad9..d645051 100644
--- a/kern/processor.c
+++ b/kern/processor.c
@@ -113,7 +113,7 @@ void pset_sys_init(void)
/*
* Allocate the zone for processor sets.
*/
- pset_zone = zinit(sizeof(struct processor_set), 128*PAGE_SIZE,
+ pset_zone = zinit(sizeof(struct processor_set), 0, 128*PAGE_SIZE,
PAGE_SIZE, 0, "processor sets");
/*
diff --git a/kern/task.c b/kern/task.c
index db61522..1c4c673 100644
--- a/kern/task.c
+++ b/kern/task.c
@@ -61,7 +61,7 @@ extern void eml_task_deallocate(task_t);
void task_init(void)
{
task_zone = zinit(
- sizeof(struct task),
+ sizeof(struct task), 0,
TASK_MAX * sizeof(struct task),
TASK_CHUNK * sizeof(struct task),
0, "tasks");
diff --git a/kern/thread.c b/kern/thread.c
index 8c08dc4..6ca91a7 100644
--- a/kern/thread.c
+++ b/kern/thread.c
@@ -299,7 +299,7 @@ void stack_privilege(
void thread_init(void)
{
thread_zone = zinit(
- sizeof(struct thread),
+ sizeof(struct thread), 0,
THREAD_MAX * sizeof(struct thread),
THREAD_CHUNK * sizeof(struct thread),
0, "threads");
diff --git a/kern/zalloc.c b/kern/zalloc.c
index dba6190..1f415f7 100644
--- a/kern/zalloc.c
+++ b/kern/zalloc.c
@@ -72,6 +72,9 @@ MACRO_BEGIN \
} \
MACRO_END
+#define ALIGN_SIZE_UP(size, align) \
+((size) = (((size) + ((align) - 1)) & ~((align) - 1)))
+
/*
* Support for garbage collection of unused zone pages:
*/
@@ -146,7 +149,7 @@ MACRO_BEGIN \
} \
MACRO_END
-static vm_offset_t zget_space();
+static vm_offset_t zget_space(vm_offset_t size, vm_size_t align);
decl_simple_lock_data(,zget_space_lock)
vm_offset_t zalloc_next_space;
@@ -182,8 +185,9 @@ int num_zones;
* are stored in a zone, which is initially a static structure that
* is initialized by zone_init.
*/
-zone_t zinit(size, max, alloc, memtype, name)
+zone_t zinit(size, align, max, alloc, memtype, name)
vm_size_t size; /* the size of an element */
+ vm_size_t align; /* alignment of elements */
vm_size_t max; /* maximum memory to use */
vm_size_t alloc; /* allocation size */
unsigned int memtype; /* flags specifying type of memory */
@@ -192,12 +196,11 @@ zone_t zinit(size, max, alloc, memtype, name)
register zone_t z;
if (zone_zone == ZONE_NULL)
- z = (zone_t) zget_space(sizeof(struct zone));
+ z = (zone_t) zget_space(sizeof(struct zone), 0);
else
z = (zone_t) zalloc(zone_zone);
if (z == ZONE_NULL)
panic("zinit");
-
if (alloc == 0)
alloc = PAGE_SIZE;
@@ -210,11 +213,18 @@ zone_t zinit(size, max, alloc, memtype, name)
if ((max = round_page(max)) < (alloc = round_page(alloc)))
max = alloc;
+ if (align > 0) {
+ if (align >= PAGE_SIZE)
+ panic("zinit");
+ ALIGN_SIZE_UP(size, align);
+ }
+
z->free_elements = 0;
z->cur_size = 0;
z->max_size = max;
z->elem_size = ((size-1) + sizeof(z->free_elements)) -
((size-1) % sizeof(z->free_elements));
+ z->align = align;
z->alloc_size = alloc;
z->type = memtype;
@@ -268,13 +278,18 @@ void zcram(zone_t zone, vm_offset_t newmem, vm_size_t size)
* of memory from zone_map.
*/
-static vm_offset_t zget_space(vm_offset_t size)
+static vm_offset_t zget_space(vm_offset_t size, vm_size_t align)
{
vm_offset_t new_space = 0;
vm_offset_t result;
vm_size_t space_to_add = 0; /*'=0' to quiet gcc warnings */
simple_lock(&zget_space_lock);
+ if (align > 0) {
+ assert(align < PAGE_SIZE);
+ ALIGN_SIZE_UP(zalloc_next_space, align);
+ }
+
while ((zalloc_next_space + size) > zalloc_end_of_space) {
/*
* Add at least one page to allocation area.
@@ -359,7 +374,7 @@ void zone_bootstrap()
zalloc_wasted_space = 0;
zone_zone = ZONE_NULL;
- zone_zone = zinit(sizeof(struct zone), 128 * sizeof(struct zone),
+ zone_zone = zinit(sizeof(struct zone), 0, 128 * sizeof(struct zone),
sizeof(struct zone), 0, "zones");
}
@@ -487,7 +502,7 @@ vm_offset_t zalloc(zone_t zone)
zone_lock(zone);
REMOVE_FROM_ZONE(zone, addr, vm_offset_t);
} else {
- addr = zget_space(zone->elem_size);
+ addr = zget_space(zone->elem_size, zone->align);
if (addr == 0)
panic("zalloc: zone %s exhausted",
zone->zone_name);
diff --git a/kern/zalloc.h b/kern/zalloc.h
index 9e44a6d..3f9bf21 100644
--- a/kern/zalloc.h
+++ b/kern/zalloc.h
@@ -59,6 +59,7 @@ struct zone {
vm_size_t cur_size; /* current memory utilization */
vm_size_t max_size; /* how large can this zone grow */
vm_size_t elem_size; /* size of an element */
+ vm_size_t align; /* alignment of elements */
vm_size_t alloc_size; /* size used for more memory */
boolean_t doing_alloc; /* is zone expanding now? */
char *zone_name; /* a name for the zone */
@@ -71,8 +72,8 @@ typedef struct zone *zone_t;
#define ZONE_NULL ((zone_t) 0)
/* Exported to everyone */
-zone_t zinit(vm_size_t size, vm_size_t max, vm_size_t alloc,
- unsigned int memtype, char *name);
+zone_t zinit(vm_size_t size, vm_size_t align, vm_size_t max,
+ vm_size_t alloc, unsigned int memtype, char *name);
vm_offset_t zalloc(zone_t zone);
vm_offset_t zget(zone_t zone);
void zfree(zone_t zone, vm_offset_t elem);