1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
|
/***********************************************************************/
/* */
/* OCaml */
/* */
/* Xavier Leroy and Damien Doligez, INRIA Rocquencourt */
/* */
/* Copyright 1995 Institut National de Recherche en Informatique et */
/* en Automatique. All rights reserved. This file is distributed */
/* under the terms of the GNU Library General Public License, with */
/* the special exception on linking described in file ../../LICENSE. */
/* */
/***********************************************************************/
/* $Id$ */
#include "alloc.h"
#include "backtrace.h"
#include "callback.h"
#include "custom.h"
#include "fail.h"
#include "io.h"
#include "memory.h"
#include "misc.h"
#include "mlvalues.h"
#include "printexc.h"
#include "roots.h"
#include "signals.h"
#ifdef NATIVE_CODE
#include "stack.h"
#else
#include "stacks.h"
#endif
#include "sys.h"
#include "threads.h"
/* Initial size of bytecode stack when a thread is created (4 Ko) */
#define Thread_stack_size (Stack_size / 4)
/* Max computation time before rescheduling, in milliseconds */
#define Thread_timeout 50
/* OS-specific code */
#ifdef _WIN32
#include "st_win32.h"
#else
#include "st_posix.h"
#endif
/* The ML value describing a thread (heap-allocated) */
struct caml_thread_descr {
value ident; /* Unique integer ID */
value start_closure; /* The closure to start this thread */
value terminated; /* Triggered event for thread termination */
};
#define Ident(v) (((struct caml_thread_descr *)(v))->ident)
#define Start_closure(v) (((struct caml_thread_descr *)(v))->start_closure)
#define Terminated(v) (((struct caml_thread_descr *)(v))->terminated)
/* The infos on threads (allocated via malloc()) */
struct caml_thread_struct {
value descr; /* The heap-allocated descriptor (root) */
struct caml_thread_struct * next; /* Double linking of running threads */
struct caml_thread_struct * prev;
#ifdef NATIVE_CODE
char * top_of_stack; /* Top of stack for this thread (approx.) */
char * bottom_of_stack; /* Saved value of caml_bottom_of_stack */
uintnat last_retaddr; /* Saved value of caml_last_return_address */
value * gc_regs; /* Saved value of caml_gc_regs */
char * exception_pointer; /* Saved value of caml_exception_pointer */
struct caml__roots_block * local_roots; /* Saved value of local_roots */
struct longjmp_buffer * exit_buf; /* For thread exit */
#else
value * stack_low; /* The execution stack for this thread */
value * stack_high;
value * stack_threshold;
value * sp; /* Saved value of extern_sp for this thread */
value * trapsp; /* Saved value of trapsp for this thread */
struct caml__roots_block * local_roots; /* Saved value of local_roots */
struct longjmp_buffer * external_raise; /* Saved external_raise */
#endif
int backtrace_pos; /* Saved backtrace_pos */
code_t * backtrace_buffer; /* Saved backtrace_buffer */
value backtrace_last_exn; /* Saved backtrace_last_exn (root) */
};
typedef struct caml_thread_struct * caml_thread_t;
/* The "head" of the circular list of thread descriptors */
static caml_thread_t all_threads = NULL;
/* The descriptor for the currently executing thread */
static caml_thread_t curr_thread = NULL;
/* The master lock protecting the OCaml runtime system */
static st_masterlock caml_master_lock;
/* Whether the ``tick'' thread is already running */
static int caml_tick_thread_running = 0;
/* The thread identifier of the ``tick'' thread */
static st_thread_id caml_tick_thread_id;
/* The key used for storing the thread descriptor in the specific data
of the corresponding system thread. */
static st_tlskey thread_descriptor_key;
/* The key used for unlocking I/O channels on exceptions */
static st_tlskey last_channel_locked_key;
/* Identifier for next thread creation */
static intnat thread_next_ident = 0;
/* Forward declarations */
static value caml_threadstatus_new (void);
static void caml_threadstatus_terminate (value);
static st_retcode caml_threadstatus_wait (value);
/* Imports from the native-code runtime system */
#ifdef NATIVE_CODE
extern struct longjmp_buffer caml_termination_jmpbuf;
extern void (*caml_termination_hook)(void);
#endif
/* Hook for scanning the stacks of the other threads */
static void (*prev_scan_roots_hook) (scanning_action);
static void caml_thread_scan_roots(scanning_action action)
{
caml_thread_t th;
th = curr_thread;
do {
(*action)(th->descr, &th->descr);
(*action)(th->backtrace_last_exn, &th->backtrace_last_exn);
/* Don't rescan the stack of the current thread, it was done already */
if (th != curr_thread) {
#ifdef NATIVE_CODE
if (th->bottom_of_stack != NULL)
do_local_roots(action, th->bottom_of_stack, th->last_retaddr,
th->gc_regs, th->local_roots);
#else
do_local_roots(action, th->sp, th->stack_high, th->local_roots);
#endif
}
th = th->next;
} while (th != curr_thread);
/* Hook */
if (prev_scan_roots_hook != NULL) (*prev_scan_roots_hook)(action);
}
/* Hooks for enter_blocking_section and leave_blocking_section */
static void caml_thread_enter_blocking_section(void)
{
/* Save the stack-related global variables in the thread descriptor
of the current thread */
#ifdef NATIVE_CODE
curr_thread->bottom_of_stack = caml_bottom_of_stack;
curr_thread->last_retaddr = caml_last_return_address;
curr_thread->gc_regs = caml_gc_regs;
curr_thread->exception_pointer = caml_exception_pointer;
curr_thread->local_roots = local_roots;
#else
curr_thread->stack_low = stack_low;
curr_thread->stack_high = stack_high;
curr_thread->stack_threshold = stack_threshold;
curr_thread->sp = extern_sp;
curr_thread->trapsp = trapsp;
curr_thread->local_roots = local_roots;
curr_thread->external_raise = external_raise;
#endif
curr_thread->backtrace_pos = backtrace_pos;
curr_thread->backtrace_buffer = backtrace_buffer;
curr_thread->backtrace_last_exn = backtrace_last_exn;
/* Tell other threads that the runtime is free */
st_masterlock_release(&caml_master_lock);
}
static void caml_thread_leave_blocking_section(void)
{
/* Wait until the runtime is free */
st_masterlock_acquire(&caml_master_lock);
/* Update curr_thread to point to the thread descriptor corresponding
to the thread currently executing */
curr_thread = st_tls_get(thread_descriptor_key);
/* Restore the stack-related global variables */
#ifdef NATIVE_CODE
caml_bottom_of_stack= curr_thread->bottom_of_stack;
caml_last_return_address = curr_thread->last_retaddr;
caml_gc_regs = curr_thread->gc_regs;
caml_exception_pointer = curr_thread->exception_pointer;
local_roots = curr_thread->local_roots;
#else
stack_low = curr_thread->stack_low;
stack_high = curr_thread->stack_high;
stack_threshold = curr_thread->stack_threshold;
extern_sp = curr_thread->sp;
trapsp = curr_thread->trapsp;
local_roots = curr_thread->local_roots;
external_raise = curr_thread->external_raise;
#endif
backtrace_pos = curr_thread->backtrace_pos;
backtrace_buffer = curr_thread->backtrace_buffer;
backtrace_last_exn = curr_thread->backtrace_last_exn;
}
static int caml_thread_try_leave_blocking_section(void)
{
/* Disable immediate processing of signals (PR#3659).
try_leave_blocking_section always fails, forcing the signal to be
recorded and processed at the next leave_blocking_section or
polling. */
return 0;
}
/* Hooks for I/O locking */
static void caml_io_mutex_free(struct channel *chan)
{
st_mutex mutex = chan->mutex;
if (mutex != NULL) st_mutex_destroy(mutex);
}
static void caml_io_mutex_lock(struct channel *chan)
{
st_mutex mutex = chan->mutex;
if (mutex == NULL) {
st_mutex_create(&mutex);
chan->mutex = mutex;
}
/* PR#4351: first try to acquire mutex without releasing the master lock */
if (st_mutex_trylock(mutex) == PREVIOUSLY_UNLOCKED) {
st_tls_set(last_channel_locked_key, (void *) chan);
return;
}
/* If unsuccessful, block on mutex */
enter_blocking_section();
st_mutex_lock(mutex);
/* Problem: if a signal occurs at this point,
and the signal handler raises an exception, we will not
unlock the mutex. The alternative (doing the setspecific
before locking the mutex is also incorrect, since we could
then unlock a mutex that is unlocked or locked by someone else. */
st_tls_set(last_channel_locked_key, (void *) chan);
leave_blocking_section();
}
static void caml_io_mutex_unlock(struct channel *chan)
{
st_mutex_unlock(chan->mutex);
st_tls_set(last_channel_locked_key, NULL);
}
static void caml_io_mutex_unlock_exn(void)
{
struct channel * chan = st_tls_get(last_channel_locked_key);
if (chan != NULL) caml_io_mutex_unlock(chan);
}
/* Hook for estimating stack usage */
static uintnat (*prev_stack_usage_hook)(void);
static uintnat caml_thread_stack_usage(void)
{
uintnat sz;
caml_thread_t th;
/* Don't add stack for current thread, this is done elsewhere */
for (sz = 0, th = curr_thread->next;
th != curr_thread;
th = th->next) {
#ifdef NATIVE_CODE
sz += (value *) th->top_of_stack - (value *) th->bottom_of_stack;
#else
sz += th->stack_high - th->sp;
#endif
}
if (prev_stack_usage_hook != NULL)
sz += prev_stack_usage_hook();
return sz;
}
/* Create and setup a new thread info block.
This block has no associated thread descriptor and
is not inserted in the list of threads. */
static caml_thread_t caml_thread_new_info(void)
{
caml_thread_t th;
th = (caml_thread_t) malloc(sizeof(struct caml_thread_struct));
if (th == NULL) return NULL;
th->descr = Val_unit; /* filled later */
#ifdef NATIVE_CODE
th->bottom_of_stack = NULL;
th->top_of_stack = NULL;
th->last_retaddr = 1;
th->exception_pointer = NULL;
th->local_roots = NULL;
th->exit_buf = NULL;
#else
/* Allocate the stacks */
th->stack_low = (value *) stat_alloc(Thread_stack_size);
th->stack_high = th->stack_low + Thread_stack_size / sizeof(value);
th->stack_threshold = th->stack_low + Stack_threshold / sizeof(value);
th->sp = th->stack_high;
th->trapsp = th->stack_high;
th->local_roots = NULL;
th->external_raise = NULL;
#endif
th->backtrace_pos = 0;
th->backtrace_buffer = NULL;
th->backtrace_last_exn = Val_unit;
return th;
}
/* Allocate a thread descriptor block. */
static value caml_thread_new_descriptor(value clos)
{
value mu = Val_unit;
value descr;
Begin_roots2 (clos, mu)
/* Create and initialize the termination semaphore */
mu = caml_threadstatus_new();
/* Create a descriptor for the new thread */
descr = alloc_small(3, 0);
Ident(descr) = Val_long(thread_next_ident);
Start_closure(descr) = clos;
Terminated(descr) = mu;
thread_next_ident++;
End_roots();
return descr;
}
/* Remove a thread info block from the list of threads.
Free it and its stack resources. */
static void caml_thread_remove_info(caml_thread_t th)
{
if (th->next == th)
all_threads = NULL; /* last OCaml thread exiting */
else if (all_threads == th)
all_threads = th->next; /* PR#5295 */
th->next->prev = th->prev;
th->prev->next = th->next;
#ifndef NATIVE_CODE
stat_free(th->stack_low);
#endif
if (th->backtrace_buffer != NULL) free(th->backtrace_buffer);
stat_free(th);
}
/* Reinitialize the thread machinery after a fork() (PR#4577) */
static void caml_thread_reinitialize(void)
{
caml_thread_t thr, next;
struct channel * chan;
/* Remove all other threads (now nonexistent)
from the doubly-linked list of threads */
thr = curr_thread->next;
while (thr != curr_thread) {
next = thr->next;
stat_free(thr);
thr = next;
}
curr_thread->next = curr_thread;
curr_thread->prev = curr_thread;
all_threads = curr_thread;
/* Reinitialize the master lock machinery,
just in case the fork happened while other threads were doing
leave_blocking_section */
st_masterlock_init(&caml_master_lock);
/* Tick thread is not currently running in child process, will be
re-created at next Thread.create */
caml_tick_thread_running = 0;
/* Destroy all IO mutexes; will be reinitialized on demand */
for (chan = caml_all_opened_channels;
chan != NULL;
chan = chan->next) {
if (chan->mutex != NULL) {
st_mutex_destroy(chan->mutex);
chan->mutex = NULL;
}
}
}
/* Initialize the thread machinery */
CAMLprim value caml_thread_initialize(value unit) /* ML */
{
/* Protect against repeated initialization (PR#1325) */
if (curr_thread != NULL) return Val_unit;
/* OS-specific initialization */
st_initialize();
/* Initialize and acquire the master lock */
st_masterlock_init(&caml_master_lock);
/* Initialize the keys */
st_tls_newkey(&thread_descriptor_key);
st_tls_newkey(&last_channel_locked_key);
/* Set up a thread info block for the current thread */
curr_thread =
(caml_thread_t) stat_alloc(sizeof(struct caml_thread_struct));
curr_thread->descr = caml_thread_new_descriptor(Val_unit);
curr_thread->next = curr_thread;
curr_thread->prev = curr_thread;
all_threads = curr_thread;
curr_thread->backtrace_last_exn = Val_unit;
#ifdef NATIVE_CODE
curr_thread->exit_buf = &caml_termination_jmpbuf;
#endif
/* The stack-related fields will be filled in at the next
enter_blocking_section */
/* Associate the thread descriptor with the thread */
st_tls_set(thread_descriptor_key, (void *) curr_thread);
/* Set up the hooks */
prev_scan_roots_hook = scan_roots_hook;
scan_roots_hook = caml_thread_scan_roots;
enter_blocking_section_hook = caml_thread_enter_blocking_section;
leave_blocking_section_hook = caml_thread_leave_blocking_section;
try_leave_blocking_section_hook = caml_thread_try_leave_blocking_section;
#ifdef NATIVE_CODE
caml_termination_hook = st_thread_exit;
#endif
caml_channel_mutex_free = caml_io_mutex_free;
caml_channel_mutex_lock = caml_io_mutex_lock;
caml_channel_mutex_unlock = caml_io_mutex_unlock;
caml_channel_mutex_unlock_exn = caml_io_mutex_unlock_exn;
prev_stack_usage_hook = caml_stack_usage_hook;
caml_stack_usage_hook = caml_thread_stack_usage;
/* Set up fork() to reinitialize the thread machinery in the child
(PR#4577) */
st_atfork(caml_thread_reinitialize);
return Val_unit;
}
/* Cleanup the thread machinery on program exit or DLL unload. */
CAMLprim value caml_thread_cleanup(value unit) /* ML */
{
if (caml_tick_thread_running) st_thread_kill(caml_tick_thread_id);
return Val_unit;
}
/* Thread cleanup at termination */
static void caml_thread_stop(void)
{
#ifndef NATIVE_CODE
/* PR#5188: update curr_thread->stack_low because the stack may have
been reallocated since the last time we entered a blocking section */
curr_thread->stack_low = stack_low;
#endif
/* Signal that the thread has terminated */
caml_threadstatus_terminate(Terminated(curr_thread->descr));
/* Remove th from the doubly-linked list of threads and free its info block */
caml_thread_remove_info(curr_thread);
/* OS-specific cleanups */
st_thread_cleanup();
/* Release the runtime system */
st_masterlock_release(&caml_master_lock);
}
/* Create a thread */
static ST_THREAD_FUNCTION caml_thread_start(void * arg)
{
caml_thread_t th = (caml_thread_t) arg;
value clos;
#ifdef NATIVE_CODE
struct longjmp_buffer termination_buf;
char tos;
#endif
/* Associate the thread descriptor with the thread */
st_tls_set(thread_descriptor_key, (void *) th);
/* Acquire the global mutex */
leave_blocking_section();
#ifdef NATIVE_CODE
/* Record top of stack (approximative) */
th->top_of_stack = &tos;
/* Setup termination handler (for caml_thread_exit) */
if (sigsetjmp(termination_buf.buf, 0) == 0) {
th->exit_buf = &termination_buf;
#endif
/* Callback the closure */
clos = Start_closure(th->descr);
modify(&(Start_closure(th->descr)), Val_unit);
callback_exn(clos, Val_unit);
caml_thread_stop();
#ifdef NATIVE_CODE
}
#endif
/* The thread now stops running */
return 0;
}
CAMLprim value caml_thread_new(value clos) /* ML */
{
caml_thread_t th;
st_retcode err;
/* Create a thread info block */
th = caml_thread_new_info();
if (th == NULL) caml_raise_out_of_memory();
/* Equip it with a thread descriptor */
th->descr = caml_thread_new_descriptor(clos);
/* Add thread info block to the list of threads */
th->next = curr_thread->next;
th->prev = curr_thread;
curr_thread->next->prev = th;
curr_thread->next = th;
/* Create the new thread */
err = st_thread_create(NULL, caml_thread_start, (void *) th);
if (err != 0) {
/* Creation failed, remove thread info block from list of threads */
caml_thread_remove_info(th);
st_check_error(err, "Thread.create");
}
/* Create the tick thread if not already done.
Because of PR#4666, we start the tick thread late, only when we create
the first additional thread in the current process*/
if (! caml_tick_thread_running) {
err = st_thread_create(&caml_tick_thread_id, caml_thread_tick, NULL);
st_check_error(err, "Thread.create");
caml_tick_thread_running = 1;
}
return th->descr;
}
/* Register a thread already created from C */
CAMLexport int caml_c_thread_register(void)
{
caml_thread_t th;
st_retcode err;
/* Already registered? */
if (st_tls_get(thread_descriptor_key) != NULL) return 0;
/* Create a thread info block */
th = caml_thread_new_info();
if (th == NULL) return 0;
#ifdef NATIVE_CODE
th->top_of_stack = (char *) &err;
#endif
/* Take master lock to protect access to the chaining of threads */
st_masterlock_acquire(&caml_master_lock);
/* Add thread info block to the list of threads */
if (all_threads == NULL) {
th->next = th;
th->prev = th;
all_threads = th;
} else {
th->next = all_threads->next;
th->prev = all_threads;
all_threads->next->prev = th;
all_threads->next = th;
}
/* Associate the thread descriptor with the thread */
st_tls_set(thread_descriptor_key, (void *) th);
/* Release the master lock */
st_masterlock_release(&caml_master_lock);
/* Now we can re-enter the run-time system and heap-allocate the descriptor */
leave_blocking_section();
th->descr = caml_thread_new_descriptor(Val_unit); /* no closure */
/* Create the tick thread if not already done. */
if (! caml_tick_thread_running) {
err = st_thread_create(&caml_tick_thread_id, caml_thread_tick, NULL);
if (err == 0) caml_tick_thread_running = 1;
}
/* Exit the run-time system */
enter_blocking_section();
return 1;
}
/* Unregister a thread that was created from C and registered with
the function above */
CAMLexport int caml_c_thread_unregister(void)
{
caml_thread_t th = st_tls_get(thread_descriptor_key);
/* Not registered? */
if (th == NULL) return 0;
/* Wait until the runtime is available */
st_masterlock_acquire(&caml_master_lock);
/* Forget the thread descriptor */
st_tls_set(thread_descriptor_key, NULL);
/* Remove thread info block from list of threads, and free it */
caml_thread_remove_info(th);
/* Release the runtime */
st_masterlock_release(&caml_master_lock);
return 1;
}
/* Return the current thread */
CAMLprim value caml_thread_self(value unit) /* ML */
{
if (curr_thread == NULL) invalid_argument("Thread.self: not initialized");
return curr_thread->descr;
}
/* Return the identifier of a thread */
CAMLprim value caml_thread_id(value th) /* ML */
{
return Ident(th);
}
/* Print uncaught exception and backtrace */
CAMLprim value caml_thread_uncaught_exception(value exn) /* ML */
{
char * msg = format_caml_exception(exn);
fprintf(stderr, "Thread %d killed on uncaught exception %s\n",
Int_val(Ident(curr_thread->descr)), msg);
free(msg);
if (caml_backtrace_active) print_exception_backtrace();
fflush(stderr);
return Val_unit;
}
/* Terminate current thread */
CAMLprim value caml_thread_exit(value unit) /* ML */
{
struct longjmp_buffer * exit_buf = NULL;
if (curr_thread == NULL) invalid_argument("Thread.exit: not initialized");
/* In native code, we cannot call pthread_exit here because on some
systems this raises a C++ exception, and ocamlopt-generated stack
frames cannot be unwound. Instead, we longjmp to the thread
creation point (in caml_thread_start) or to the point in
caml_main where caml_termination_hook will be called.
Note that threads created in C then registered do not have
a creation point (exit_buf == NULL).
*/
#ifdef NATIVE_CODE
exit_buf = curr_thread->exit_buf;
#endif
caml_thread_stop();
if (exit_buf != NULL) {
/* Native-code and (main thread or thread created by OCaml) */
siglongjmp(exit_buf->buf, 1);
} else {
/* Bytecode, or thread created from C */
st_thread_exit();
}
return Val_unit; /* not reached */
}
/* Allow re-scheduling */
CAMLprim value caml_thread_yield(value unit) /* ML */
{
if (st_masterlock_waiters(&caml_master_lock) == 0) return Val_unit;
enter_blocking_section();
st_thread_yield();
leave_blocking_section();
return Val_unit;
}
/* Suspend the current thread until another thread terminates */
CAMLprim value caml_thread_join(value th) /* ML */
{
st_retcode rc = caml_threadstatus_wait(Terminated(th));
st_check_error(rc, "Thread.join");
return Val_unit;
}
/* Mutex operations */
#define Mutex_val(v) (* ((st_mutex *) Data_custom_val(v)))
#define Max_mutex_number 5000
static void caml_mutex_finalize(value wrapper)
{
st_mutex_destroy(Mutex_val(wrapper));
}
static int caml_mutex_compare(value wrapper1, value wrapper2)
{
st_mutex mut1 = Mutex_val(wrapper1);
st_mutex mut2 = Mutex_val(wrapper2);
return mut1 == mut2 ? 0 : mut1 < mut2 ? -1 : 1;
}
static intnat caml_mutex_hash(value wrapper)
{
return (intnat) (Mutex_val(wrapper));
}
static struct custom_operations caml_mutex_ops = {
"_mutex",
caml_mutex_finalize,
caml_mutex_compare,
caml_mutex_hash,
custom_serialize_default,
custom_deserialize_default
};
CAMLprim value caml_mutex_new(value unit) /* ML */
{
st_mutex mut = NULL; /* suppress warning */
value wrapper;
st_check_error(st_mutex_create(&mut), "Mutex.create");
wrapper = alloc_custom(&caml_mutex_ops, sizeof(st_mutex *),
1, Max_mutex_number);
Mutex_val(wrapper) = mut;
return wrapper;
}
CAMLprim value caml_mutex_lock(value wrapper) /* ML */
{
st_mutex mut = Mutex_val(wrapper);
st_retcode retcode;
/* PR#4351: first try to acquire mutex without releasing the master lock */
if (st_mutex_trylock(mut) == PREVIOUSLY_UNLOCKED) return Val_unit;
/* If unsuccessful, block on mutex */
Begin_root(wrapper) /* prevent the deallocation of mutex */
enter_blocking_section();
retcode = st_mutex_lock(mut);
leave_blocking_section();
End_roots();
st_check_error(retcode, "Mutex.lock");
return Val_unit;
}
CAMLprim value caml_mutex_unlock(value wrapper) /* ML */
{
st_mutex mut = Mutex_val(wrapper);
st_retcode retcode;
/* PR#4351: no need to release and reacquire master lock */
retcode = st_mutex_unlock(mut);
st_check_error(retcode, "Mutex.unlock");
return Val_unit;
}
CAMLprim value caml_mutex_try_lock(value wrapper) /* ML */
{
st_mutex mut = Mutex_val(wrapper);
st_retcode retcode;
retcode = st_mutex_trylock(mut);
if (retcode == ALREADY_LOCKED) return Val_false;
st_check_error(retcode, "Mutex.try_lock");
return Val_true;
}
/* Conditions operations */
#define Condition_val(v) (* (st_condvar *) Data_custom_val(v))
#define Max_condition_number 5000
static void caml_condition_finalize(value wrapper)
{
st_condvar_destroy(Condition_val(wrapper));
}
static int caml_condition_compare(value wrapper1, value wrapper2)
{
st_condvar cond1 = Condition_val(wrapper1);
st_condvar cond2 = Condition_val(wrapper2);
return cond1 == cond2 ? 0 : cond1 < cond2 ? -1 : 1;
}
static intnat caml_condition_hash(value wrapper)
{
return (intnat) (Condition_val(wrapper));
}
static struct custom_operations caml_condition_ops = {
"_condition",
caml_condition_finalize,
caml_condition_compare,
caml_condition_hash,
custom_serialize_default,
custom_deserialize_default,
custom_compare_ext_default
};
CAMLprim value caml_condition_new(value unit) /* ML */
{
st_condvar cond = NULL; /* suppress warning */
value wrapper;
st_check_error(st_condvar_create(&cond), "Condition.create");
wrapper = alloc_custom(&caml_condition_ops, sizeof(st_condvar *),
1, Max_condition_number);
Condition_val(wrapper) = cond;
return wrapper;
}
CAMLprim value caml_condition_wait(value wcond, value wmut) /* ML */
{
st_condvar cond = Condition_val(wcond);
st_mutex mut = Mutex_val(wmut);
st_retcode retcode;
Begin_roots2(wcond, wmut) /* prevent deallocation of cond and mutex */
enter_blocking_section();
retcode = st_condvar_wait(cond, mut);
leave_blocking_section();
End_roots();
st_check_error(retcode, "Condition.wait");
return Val_unit;
}
CAMLprim value caml_condition_signal(value wrapper) /* ML */
{
st_check_error(st_condvar_signal(Condition_val(wrapper)),
"Condition.signal");
return Val_unit;
}
CAMLprim value caml_condition_broadcast(value wrapper) /* ML */
{
st_check_error(st_condvar_broadcast(Condition_val(wrapper)),
"Condition.signal");
return Val_unit;
}
/* Thread status blocks */
#define Threadstatus_val(v) (* ((st_event *) Data_custom_val(v)))
#define Max_threadstatus_number 500
static void caml_threadstatus_finalize(value wrapper)
{
st_event_destroy(Threadstatus_val(wrapper));
}
static int caml_threadstatus_compare(value wrapper1, value wrapper2)
{
st_event ts1 = Threadstatus_val(wrapper1);
st_event ts2 = Threadstatus_val(wrapper2);
return ts1 == ts2 ? 0 : ts1 < ts2 ? -1 : 1;
}
static struct custom_operations caml_threadstatus_ops = {
"_threadstatus",
caml_threadstatus_finalize,
caml_threadstatus_compare,
custom_hash_default,
custom_serialize_default,
custom_deserialize_default,
custom_compare_ext_default
};
static value caml_threadstatus_new (void)
{
st_event ts = NULL; /* suppress warning */
value wrapper;
st_check_error(st_event_create(&ts), "Thread.create");
wrapper = alloc_custom(&caml_threadstatus_ops, sizeof(st_event *),
1, Max_threadstatus_number);
Threadstatus_val(wrapper) = ts;
return wrapper;
}
static void caml_threadstatus_terminate (value wrapper)
{
st_event_trigger(Threadstatus_val(wrapper));
}
static st_retcode caml_threadstatus_wait (value wrapper)
{
st_event ts = Threadstatus_val(wrapper);
st_retcode retcode;
Begin_roots1(wrapper) /* prevent deallocation of ts */
enter_blocking_section();
retcode = st_event_wait(ts);
leave_blocking_section();
End_roots();
return retcode;
}
|