1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
|
[[!meta copyright="Copyright © 2012, 2013 Free Software Foundation, Inc."]]
[[!meta license="""[[!toggle id="license" text="GFDL 1.2+"]][[!toggleable
id="license" text="Permission is granted to copy, distribute and/or modify this
document under the terms of the GNU Free Documentation License, Version 1.2 or
any later version published by the Free Software Foundation; with no Invariant
Sections, no Front-Cover Texts, and no Back-Cover Texts. A copy of the license
is included in the section entitled [[GNU Free Documentation
License|/fdl]]."]]"""]]
[[!tag open_issue_documentation open_issue_gnumach]]
# IRC, freenode, #hurd, 2012-06-29
<henrikcozza> I do not understand what are the deficiencies of Mach, the
content I find on this is vague...
<antrik> the major problems are that the IPC architecture offers poor
performance; and that resource usage can not be properly accounted to the
right parties
<braunr> antrik: the more i study it, the more i think ipc isn't the
problem when it comes to performance, not directly
<braunr> i mean, the implementation is a bit heavy, yes, but it's fine
<braunr> the problems are resource accounting/scheduling and still too much
stuff inside kernel space
<braunr> and with a very good implementation, the performance problem would
come from crossing address spaces
<braunr> (and even more on SMP, i've been thinking about it lately, since
it would require syncing mmu state on each processor currently using an
address space being modified)
<antrik> braunr: the problem with Mach IPC is that it requires too many
indirections to ever be performant AIUI
<braunr> antrik: can you mention them ?
<antrik> the semantics are generally quite complex, compared to Coyotos for
example, or even Viengoos
<braunr> antrik: the semantics are related to the message format, which can
be simplified
<braunr> i think everybody agrees on that
<braunr> i'm more interested in the indirections
<antrik> but then it's not Mach IPC anymore :-)
<braunr> right
<braunr> 22:03 < braunr> i mean, the implementation is a bit heavy, yes,
but it's fine
<antrik> that's not an implementation issue
<braunr> that's what i meant by heavy :)
<braunr> well, yes and no
<braunr> Mach IPC have changed over time
<braunr> it would be newer Mach IPC ... :)
<antrik> the fact that data types are (supposed to be) transparent to the
kernel is a major part of the concept, not just an implementation detail
<antrik> but it's not just the message format
<braunr> transparent ?
<braunr> but they're not :/
<antrik> the option to buffer in the kernel also adds a lot of complexity
<braunr> buffer in the kernel ?
<braunr> ah you mean message queues
<braunr> yes
<antrik> braunr: eh? the kernel parses all the type headers during transfer
<braunr> yes, so it's not transparent at all
<antrik> maybe you have a different understanding of "transparent" ;-)
<braunr> i guess
<antrik> I think most of the other complex semantics are kinda related to
the in-kernel buffering...
<braunr> i fail to see why :/
<antrik> well, it allows ports rights to be destroyed while a message is in
transfer. a lot of semantics revolve around what happens in that case
<braunr> yes but it doesn't affect performance a lot
<antrik> sure it does. it requires a lot of extra code and indirections
<braunr> not a lot of it
<antrik> "a lot" is quite a relative term :-)
<antrik> compared to L4 for example, it *is* a lot
<braunr> and those indirections (i think you refer to more branching here)
are taken only when appropriate, and can be isolated, improved through
locality, etc..
<braunr> the features they add are also huge
<braunr> L4 is clearly insufficient
<braunr> all current L4 forks have added capabilities ..
<braunr> (that, with the formal verification, make se4L one of the
"hottest" recent system projects)
<braunr> seL4*
<antrik> yes, but with very few extra indirection I think... similar to
EROS (which claims to have IPC almost as efficient as the original L4)
<braunr> possibly
<antrik> I still fail to see much real benefit in formal verification :-)
<braunr> but compared to other problems, this added code is negligible
<braunr> antrik: for a microkernel, me too :/
<braunr> the kernel is already so small you can simply audit it :)
<antrik> no, it's not neglible, if you go from say two cache lines touched
per IPC (original L4) to dozens (Mach)
<antrik> every additional variable that needs to be touched to resolve some
indirection, check some condition adds significant overhead
<braunr> if you compare the dozens to the huge amount of inter processor
interrupt you get each time you change the kernel map, it's next to
nothing ..
<antrik> change the kernel map? not sure what you mean
<braunr> syncing address spaces on hundreds of processors each time you
send a message is a real scalability issue here (as an example), where
Mach to L4 IPC seem like microoptimization
<youpi> braunr: modify, you mean?
<braunr> yes
<youpi> (not switchp
<youpi> )
<braunr> but that's only one example
<braunr> yes, modify, not switch
<braunr> also, we could easily get rid of the ihash library
<braunr> making the message provide the address of the object associated to
a receive right
<braunr> so the only real indirection is the capability, like in other
systems, and yes, buffering adds a bit of complexity
<braunr> there are other optimizations that could be made in mach, like
merging structures to improve locality
<pinotree> "locality"?
<braunr> having rights close to their target port when there are only a few
<braunr> pinotree: locality of reference
<youpi> for cache efficiency
<antrik> hundreds of processors? let's stay realistic here :-)
<braunr> i am ..
<braunr> a microkernel based system is also a very good environment for RCU
<braunr> (i yet have to understand how liburcu actually works on linux)
<antrik> I'm not interested in systems for supercomputers. and I doubt
desktop machines will get that many independant cores any time soon. we
still lack software that could even romotely exploit that
<braunr> hum, the glibc build system ? :>
<braunr> lol
<youpi> we have done a survey over the nix linux distribution
<youpi> quite few packages actually benefit from a lot of cores
<youpi> and we already know them :)
<braunr> what i'm trying to say is that, whenever i think or even measure
system performance, both of the hurd and others, i never actually see the
IPC as being the real performance problem
<braunr> there are many other sources of overhead to overcome before
getting to IPC
<youpi> I completely agree
<braunr> and with the advent of SMP, it's even more important to focus on
contention
<antrik> (also, 8 cores aren't exactly a lot...)
<youpi> antrik: s/8/7/ , or even 6 ;)
<antrik> braunr: it depends a lot on the use case. most of the problems we
see in the Hurd are probably not directly related to IPC performance; but
I pretty sure some are
<antrik> (such as X being hardly usable with UNIX domain sockets)
<braunr> antrik: these have more to do with the way mach blocks than IPC
itself
<braunr> similar to the ext2 "sleep storm"
<antrik> a lot of overhead comes from managing ports (for for example),
which also mostly comes down to IPC performance
<braunr> antrik: yes, that's the main indirection
<braunr> antrik: but you need such management, and the related semantics in
the kernel interface
<braunr> (although i wonder if those should be moved away from the message
passing call)
<antrik> you mean a different interface for kernel calls than for IPC to
other processes? that would break transparency in a major way. not sure
we really want that...
<braunr> antrik: no
<braunr> antrik: i mean calls specific to right management
<antrik> admittedly, transparency for port management is only useful in
special cases such as rpctrace, and that probably could be served better
with dedicated debugging interfaces...
<braunr> antrik: i.e. not passing rights inside messages
<antrik> passing rights inside messages is quite essential for a capability
system. the problem with Mach IPC in regard to that is that the message
format allows way more flexibility than necessary in that regard...
<braunr> antrik: right
<braunr> antrik: i don't understand why passing rights inside messages is
important though
<braunr> antrik: essential even
<youpi> braunr: I guess he means you need at least one way to pass rights
<antrik> braunr: well, for one, you need to pass a reply port with each RPC
request...
<braunr> youpi: well, as he put, the message passing call is overpowered,
and this leads to many branches in the code
<braunr> antrik: the reply port is obvious, and can be optimized
<braunr> antrik: but the case i worry about is passing references to
objects between tasks
<braunr> antrik: rights and identities with the auth server for example
<braunr> antrik: well ok forget it, i just recall how it actually works :)
<braunr> antrik: don't forget we lack thread migration
<braunr> antrik: you may not think it's important, but to me, it's a major
improvement for RPC performance
<antrik> braunr: how can seL4 be the most interesting microkernel
then?... ;-)
<braunr> antrik: hm i don't know the details, but if it lacks thread
migration, something is wrong :p
<braunr> antrik: they should work on viengoos :)
<antrik> (BTW, AIUI thread migration is quite related to passive objects --
something Hurd folks never dared seriously consider...)
<braunr> i still don't know what passive objects are, or i have forgotten
it :/
<antrik> no own control threads
<braunr> hm, i'm still missing something
<braunr> what do you refer to by control thread ?
<braunr> with*
<antrik> i.e. no main loop etc.; only activated by incoming calls
<braunr> ok
<braunr> well, if i'm right, thomas bushnel himself wrote (recently) that
the ext2 "sleep" performance issue was expected to be solved with thread
migration
<braunr> so i guess they definitely considered having it
<antrik> braunr: don't know what the "sleep peformance issue" is...
<braunr> http://lists.gnu.org/archive/html/bug-hurd/2011-12/msg00032.html
<braunr> antrik: also, the last message in the thread,
http://lists.gnu.org/archive/html/bug-hurd/2011-12/msg00050.html
<braunr> antrik: do you consider having a reply port being an avoidable
overhead ?
<antrik> braunr: not sure. I don't remember hearing of any capability
system doing this kind of optimisation though; so I guess there are
reasons for that...
<braunr> antrik: yes me too, even more since neal talked about it on
viengoos
<antrik> I wonder whether thread management is also such a large overhead
with fully sync IPC, on L4 or EROS for example...
<braunr> antrik: it's still a very handy optimization for thread scheduling
<braunr> antrik: it makes solving priority inversions a lot easier
<antrik> actually, is thread scheduling a problem at all with a thread
activation approach like in Viengoos?
<braunr> antrik: thread activation is part of thread migration
<braunr> antrik: actually, i'd say they both refer to the same thing
<antrik> err... scheduler activation was the term I wanted to use
<braunr> same
<braunr> well
<braunr> scheduler activation is too vague to assert that
<braunr> antrik: do you refer to scheduler activations as described in
http://en.wikipedia.org/wiki/Scheduler_activations ?
<antrik> my understanding was that Viengoos still has traditional threads;
they just can get scheduled directly on incoming IPC
<antrik> braunr: that Wikipedia article is strange. it seems to use
"scheduler activations" as a synonym for N:M multithreading, which is not
at all how I understood it
<youpi> antrik: I used to try to keep a look at those pages, to fix such
wrong things, but left it
<braunr> antrik: that's why i ask
<antrik> IIRC Viengoos has a thread associated with each receive
buffer. after copying the message, the kernel would activate the
processes activation handler, which in turn could decide to directly
schedule the thead associated with the buffer
<antrik> or something along these lines
<braunr> antrik: that's similar to mach handoff
<youpi> antrik: generally enough, all the thread-related pages on wikipedia
are quite bogus
<antrik> nah, handoff just schedules the process; which is not useful, if
the right thread isn't activated in turn...
<braunr> antrik: but i think it's more than that, even in viengoos
<youpi> for instance, the french "thread" page was basically saying that
they were invented for GUIs to overlap computation with user interaction
<braunr> .. :)
<antrik> youpi: good to know...
<braunr> antrik: the "misunderstanding" comes from the fact that scheduler
activations is the way N:M threading was implemented on netbsd
<antrik> youpi: that's a refreshing take on the matter... ;-)
<braunr> antrik: i'll read the critique and viengoos doc/source again to be
sure about what we're talking :)
<braunr> antrik: as threading is a major issue in mach, and one of the
things i completely changed (and intend to change) in x15, whenever i get
to work on that again ..... :)
<braunr> antrik: interestingly, the paper about scheduler activations was
written (among others) by brian bershad, in 92, when he was actively
working on research around mach
<antrik> braunr: BTW, I have little doubt that making RPC first-class would
solve a number of problems... I just wonder how many others it would open
# IRC, freenode, #hurd, 2012-09-04
X15
<braunr> it was intended as a mach clone, but now that i have better
knowledge of both mach and the hurd, i don't want to retain mach
compatibility
<braunr> and unlike viengoos, it's not really experimental
<braunr> it's focused on memory and cpu scalability, and performance, with
techniques likes thread migration and rcu
<braunr> the design i have in mind is closer to what exists today, with
strong emphasis on scalability and performance, that's all
<braunr> and the reason the hurd can't be modified first is that my design
relies on some important design changes
<braunr> so there is a strong dependency on these mechanisms that requires
the kernel to exists first
## IRC, freenode, #hurd, 2012-09-06
In context of [[open_issues/multithreading]] and later [[open_issues/select]].
<gnu_srs> And you will address the design flaws or implementation faults
with x15?
<braunr> no
<braunr> i'll address the implementation details :p
<braunr> and some design issues like cpu and memory resource accounting
<braunr> but i won't implement generic resource containers
<braunr> assuming it's completed, my work should provide a hurd system on
par with modern monolithic systems
<braunr> (less performant of course, but performant, scalable, and with
about the same kinds of problems)
<braunr> for example, thread migration should be mandatory
<braunr> which would make client calls behave exactly like a userspace task
asking a service from the kernel
<braunr> you have to realize that, on a monolithic kernel, applications are
clients, and the kernel is a server
<braunr> and when performing a system call, the calling thread actually
services itself by running kernel code
<braunr> which is exactly what thread migration is for a multiserver system
<braunr> thread migration also implies sync IPC
<braunr> and sync IPC is inherently more performant because it only
requires one copy, no in kernel buffering
<braunr> sync ipc also avoids message floods, since client threads must run
server code
<gnu_srs> and this is not achievable with evolved gnumach and/or hurd?
<braunr> well that's not entirely true, because there is still a form of
async ipc, but it's a lot less likely
<braunr> it probably is
<braunr> but there are so many things to change i prefer starting from
scratch
<braunr> scalability itself probably requires a revamp of the hurd core
libraries
<braunr> and these libraries are like more than half of the hurd code
<braunr> mach ipc and vm are also very complicated
<braunr> it's better to get something new and simpler from the start
<gnu_srs> a major task nevertheless:-D
<braunr> at least with the vm, netbsd showed it's easier to achieve good
results from new code, as other mach vm based systems like freebsd
struggled to get as good
<braunr> well yes
<braunr> but at least it's not experimental
<braunr> everything i want to implement already exists, and is tested on
production systems
<braunr> it's just time to assemble those ideas and components together
into something that works
<braunr> you could see it as a qnx-like system with thread migration, the
global architecture of the hurd, and some improvements from linux like
rcu :)
### IRC, freenode, #hurd, 2012-09-07
<antrik> braunr: thread migration is tested on production systems?
<antrik> BTW, I don't think that generally increasing the priority of
servers is a good idea
<antrik> in most cases, IPC should actually be sync. slpz looked at it at
some point, and concluded that the implementation actually has a
fast-path for that case. I wonder what happens to scheduling in this case
-- is the receiver sheduled immediately? if not, that's something to
fix...
<braunr> antrik: qnx does something very close to thread migration, yes
<braunr> antrik: i agree increasing the priority isn't a good thing, but
it's the best of the quick and dirty ways to reduce message floods
<braunr> the problem isn't sync ipc in mach
<braunr> the problem is the notifications (in our cases the dead name
notifications) that are by nature async
<braunr> and a malicious program could send whatever it wants at the
fastest rate it can
<antrik> braunr: malicious programs can do any number of DOS attacks on the
Hurd; I don't see how increasing priority of system servers is relevant
in that context
<antrik> (BTW, I don't think dead name notifications are async by
nature... just like for most other IPC, the *usual* case is that a server
thread is actively waiting for the message when it's generated)
<braunr> antrik: it's async with respect to the client
<braunr> antrik: and malicious programs shouldn't be able to do that kind
of dos
<braunr> but this won't be fixed any time soon
<braunr> on the other hand, a higher priority helps servers not create too
many threads because of notifications, and that's a good thing
<braunr> gnu_srs: the "fix" for this will be to rewrite select so that it's
synchronous btw
<braunr> replacing dead name notifications with something like cancelling a
previously installed select request
<antrik> no idea what "async with respect to the client" means
<braunr> it means the client doesn't wait for anything
<antrik> what is the client? what scenario are you talking about? how does
it affect scheduling?
<braunr> for notifications, it's usually the kernel
<braunr> it doesn't directly affect scheduling
<braunr> it affects the amount of messages a hurd server has to take care
of
<braunr> and the more messages, the more threads
<braunr> i'm talking about event loops
<braunr> and non blocking (or very short) selects
<antrik> the amount of messages is always the same. the question is whether
they can be handled before more come in. which would be the case if be
default the receiver gets scheduled as soon as a message is sent...
<braunr> no
<braunr> scheduling handoff doesn't imply the thread will be ready to
service the next message by the time a client sends a new one
<braunr> the rate at which a message queue gets filled has nothing to do
with scheduling handoff
<antrik> I very much doubt rates come into play at all
<braunr> well they do
<antrik> in my understanding the problem is that a lot of messages are sent
before the receive ever has a chance to handle them. so no matter how
fast the receiver is, it looses
<braunr> a lot of non blocking selects means a lot of reply ports
destroyed, a lot of dead name notifications, and what i call message
floods at server side
<braunr> no
<braunr> it used to work fine with cthreads
<braunr> it doesn't any more with pthreads because pthreads are slightly
slower
<antrik> if the receiver gets a chance to do some work each time a message
arrives, in most cases it would be free to service the next request with
the same thread
<braunr> no, because that thread won't have finished soon enough
<antrik> no, it *never* worked fine. it might have been slighly less
terrible.
<braunr> ok it didn't work fine, it worked ok
<braunr> it's entirely a matter of rate here
<braunr> and that's the big problem, because it shouldn't
<antrik> I'm pretty sure the thread would finish before the time slice ends
in almost all cases
<braunr> no
<braunr> too much contention
<braunr> and in addition locking a contended spin lock depresses priority
<braunr> so servers really waste a lot of time because of that
<antrik> I doubt contention would be a problem if the server gets a chance
to handle each request before 100 others come in
<braunr> i don't see how this is related
<braunr> handling a request doesn't mean entirely processing it
<braunr> there is *no* relation between handoff and the rate of incoming
message rate
<braunr> unless you assume threads can always complete their task in some
fixed and low duration
<antrik> sure there is. we are talking about a single-processor system
here.
<braunr> which is definitely not the case
<braunr> i don't see what it changes
<antrik> I'm pretty sure notifications can generally be handled in a very
short time
<braunr> if the server thread is scheduled as soon as it gets a message, it
can also get preempted by the kernel before replying
<braunr> no, notifications can actually be very long
<braunr> hurd_thread_cancel calls condition_broadcast
<braunr> so if there are a lot of threads on that ..
<braunr> (this is one of the optimizations i have in mind for pthreads,
since it's possible to precisely select the target thread with a doubly
linked list)
<braunr> but even if that's the case, there is no guarantee
<braunr> you can't assume it will be "quick enough"
<antrik> there is no guarantee. but I'm pretty sure it will be "quick
enough" in the vast majority of cases. which is all it needs.
<braunr> ok
<braunr> that's also the idea behind raising server priorities
<antrik> braunr: so you are saying the storms are all caused by select(),
and once this is fixed, the problem should be mostly gone and the
workaround not necessary anymore?
<braunr> yes
<antrik> let's hope you are right :-)
<braunr> :)
<antrik> (I still think though that making hand-off scheduling default is
the right thing to do, and would improve performance in general...)
<braunr> sure
<braunr> well
<braunr> no it's just a hack ;p
<braunr> but it's a right one
<braunr> the right thing to do is a lot more complicated
<braunr> as roland wrote a long time ago, the hurd doesn't need dead-name
notifications, or any notification other than the no-sender (which can be
replaced by a synchronous close on fd like operation)
<antrik> well, yes... I still think the viengoos approach is promising. I
meant the right thing to do in the existing context ;-)
<braunr> better than this priority hack
<antrik> oh? you happen to have a link? never heard of that...
<braunr> i didn't want to do it initially, even resorting to priority
depression on trhead creation to work around the problem
<braunr> hm maybe it wasn't him, i can't manage to find it
<braunr> antrik:
http://lists.gnu.org/archive/html/l4-hurd/2003-09/msg00009.html
<braunr> "Long ago, in specifying the constraints of
<braunr> what the Hurd needs from an underlying IPC system/object model we
made it
<braunr> very clear that we only need no-senders notifications for object
<braunr> implementors (servers)"
<braunr> "We don't in general make use of dead-name notifications,
<braunr> which are the general kind of object death notification Mach
provides and
<braunr> what serves as task death notification."
<braunr> "In the places we do, it's to serve
<braunr> some particular quirky need (and mostly those are side effects of
Mach's
<braunr> decouplable RPCs) and not a semantic model we insist on having."
### IRC, freenode, #hurd, 2012-09-08
<antrik> The notion that seemed appropriate when we thought about these
issues for
<antrik> Fluke was that the "alert" facility be a feature of the IPC system
itself
<antrik> rather than another layer like the Hurd's io_interrupt protocol.
<antrik> braunr: funny, that's *exactly* what I was thinking when looking
at the io_interrupt mess :-)
<antrik> (and what ultimately convinced me that the Hurd could be much more
elegant with a custom-tailored kernel rather than building around Mach)
## IRC, freenode, #hurd, 2012-09-24
<braunr> my initial attempt was a mach clone
<braunr> but now i want a mach-like kernel, without compability
<lisporu> which new licence ?
<braunr> and some very important changes like sync ipc
<braunr> gplv3
<braunr> (or later)
<lisporu> cool 8)
<braunr> yes it is gplv2+ since i didn't take the time to read gplv3, but
now that i have, i can't use anything else for such a project: )
<lisporu> what is mach-like ? (how it is different from Pistachio like ?)
<braunr> l4 doesn't provide capabilities
<lisporu> hmmm..
<braunr> you need a userspace for that
<braunr> +server
<braunr> and it relies on complete external memory management
<lisporu> how much work is done ?
<braunr> my kernel will provide capabilities, similar to mach ports, but
simpler (less overhead)
<braunr> i want the primitives right
<braunr> like multiprocessor, synchronization, virtual memory, etc..
### IRC, freenode, #hurd, 2012-09-30
<braunr> for those interested, x15 is now a project of its own, with no
gnumach compability goal, and covered by gplv3+
### IRC, freenode, #hurd, 2012-12-31
<braunr> bits of news about x15: it can now create tasks, threads, vm_maps,
physical maps (cpu-specific page tables) for user tasks, and stack
tracing (in addition to symbol resolution when symbols are available)
were recently added
### IRC, freenode, #hurd, 2013-01-15
<braunr> Anarchos: as a side note, i'm currently working on a hurd clone
with a microkernel that takes a lot from mach but also completely changes
the ipc interface (making it not mach at all in the end)
<braunr> it's something between mach and qnx neutrino
<zacts> braunr: do you have a git repo of your new clone?
<braunr> http://git.sceen.net/rbraun/x15.git/
<zacts> neat
<braunr> it's far from complete
<braunr> and hasn't reached a status where it can be publically announced
<zacts> ok
<braunr> but progress has been constant so far, the ideas i'm using have
proven applicable on other systems, i don't expect the kind of design
issues that blocked HurdNG
<braunr> (also, my attempt doesn't aim at the same goals as hurdng did)
<braunr> (e.g. denial of service remains completely possible)
<zacts> so x15 will use the current hurd translators? you are only
replacing gnumach?
<braunr> that was the plan some years ago, but now that i know the hurd
better, i think the main issues are in the hurd, so there isn't much
point rewriting mach
<braunr> so, if the hurd needs a revamp, it's better to also make the
underlying interface better if possible
<braunr> zacts: in other words: it's a completely different beast
<zacts> ok
<braunr> the main goal is to create a hurd-like system that overcomes the
current major defficiencies, most of them being caused by old design
decisions
<zacts> like async ipc?
<braunr> yes
<Anarchos> time for a persistent hurd ? :)
<braunr> no way
<braunr> i don't see a point to persistence for a general purpose system
<braunr> and it easily kills performance
<braunr> on the other hand, it would be nice to have a truely scalable,
performant, and free microkernel based system
<braunr> (and posix compatible)
<braunr> there is currently none
<braunr> zacts: the projects focuses mostly on performance and scalability,
while also being very easy to understand and maintain (something i think
the current hurd has failed at :/)
<braunr> project*
<zacts> very cool
<braunr> i think so, but we'll have to wait for an end result :)
<braunr> what's currently blocking me is the IDL
<braunr> earlier research has shown that an idl must be optmized the same
way compilers are for the best performances
<braunr> i'm not sure i can write something good enough :/
<braunr> the first version will probably be very similar to mig, small and
unoptimized
### IRC, freenode, #hurd, 2013-01-18
<zacts> braunr: so how exactly do the goals of x15 differ from viengoos?
<braunr> zacts: viengoos is much more ambitious about the design
<braunr> tbh, i still don't clearly see how its half-sync ipc work
<braunr> x15 is much more mach-like, e.g. a hybrid microkernel with
scheduling and virtual memory in the kernel
<braunr> its goals are close to those of mach, adding increased scalability
and performance to the list
<zacts> that's neat
<braunr> that's different
<braunr> in a way, you could consider x15 is to mach what linux is to unix,
a clone with a "slightly" different interface
<zacts> ah, ok. cool!
<braunr> viengoos is rather a research project, with very interesting
goals, i think they're both neat :p
### IRC, freenode, #hurd, 2013-01-19
<braunr> for now, it provides kernel memory allocation and basic threading
<braunr> it already supports both i386 and amd64 processors (from i586
onwards), and basic smp
<zacts> oh wow
<zacts> how easily can it be ported to other archs?
<braunr> the current focus is smp load balancing, so that thread migration
is enabled during development
<braunr> hard to say
<braunr> everything that is arch-specific is cleanly separated, the same
way it is in mach and netbsd
<braunr> but the arch-specific interfaces aren't well defined yet because
there is only one (and incomplete) arch
|