OpenCores
URL https://opencores.org/ocsvn/openrisc/openrisc/trunk

Subversion Repositories openrisc

[/] [openrisc/] [trunk/] [gnu-dev/] [or1k-gcc/] [boehm-gc/] [finalize.c] - Blame information for rev 778

Go to most recent revision | Details | Compare with Previous | View Log

Line No. Rev Author Line
1 721 jeremybenn
/*
2
 * Copyright 1988, 1989 Hans-J. Boehm, Alan J. Demers
3
 * Copyright (c) 1991-1996 by Xerox Corporation.  All rights reserved.
4
 * Copyright (c) 1996-1999 by Silicon Graphics.  All rights reserved.
5
 * Copyright (C) 2007 Free Software Foundation, Inc
6
 
7
 * THIS MATERIAL IS PROVIDED AS IS, WITH ABSOLUTELY NO WARRANTY EXPRESSED
8
 * OR IMPLIED.  ANY USE IS AT YOUR OWN RISK.
9
 *
10
 * Permission is hereby granted to use or copy this program
11
 * for any purpose,  provided the above notices are retained on all copies.
12
 * Permission to modify the code and to distribute modified code is granted,
13
 * provided the above notices are retained, and a notice that the code was
14
 * modified is included with the above copyright notice.
15
 */
16
/* Boehm, February 1, 1996 1:19 pm PST */
17
# define I_HIDE_POINTERS
18
# include "private/gc_pmark.h"
19
 
20
# ifdef FINALIZE_ON_DEMAND
21
    int GC_finalize_on_demand = 1;
22
# else
23
    int GC_finalize_on_demand = 0;
24
# endif
25
 
26
# ifdef JAVA_FINALIZATION
27
    int GC_java_finalization = 1;
28
# else
29
    int GC_java_finalization = 0;
30
# endif
31
 
32
/* Type of mark procedure used for marking from finalizable object.     */
33
/* This procedure normally does not mark the object, only its           */
34
/* descendents.                                                         */
35
typedef void finalization_mark_proc(/* ptr_t finalizable_obj_ptr */);
36
 
37
# define HASH3(addr,size,log_size) \
38
    ((((word)(addr) >> 3) ^ ((word)(addr) >> (3+(log_size)))) \
39
    & ((size) - 1))
40
#define HASH2(addr,log_size) HASH3(addr, 1 << log_size, log_size)
41
 
42
struct hash_chain_entry {
43
    word hidden_key;
44
    struct hash_chain_entry * next;
45
};
46
 
47
unsigned GC_finalization_failures = 0;
48
        /* Number of finalization requests that failed for lack of memory. */
49
 
50
static struct disappearing_link {
51
    struct hash_chain_entry prolog;
52
#   define dl_hidden_link prolog.hidden_key
53
                                /* Field to be cleared.         */
54
#   define dl_next(x) (struct disappearing_link *)((x) -> prolog.next)
55
#   define dl_set_next(x,y) (x) -> prolog.next = (struct hash_chain_entry *)(y)
56
 
57
    word dl_hidden_obj;         /* Pointer to object base       */
58
} **dl_head = 0;
59
 
60
static signed_word log_dl_table_size = -1;
61
                        /* Binary log of                                */
62
                        /* current size of array pointed to by dl_head. */
63
                        /* -1 ==> size is 0.                            */
64
 
65
word GC_dl_entries = 0;  /* Number of entries currently in disappearing  */
66
                        /* link table.                                  */
67
 
68
static struct finalizable_object {
69
    struct hash_chain_entry prolog;
70
#   define fo_hidden_base prolog.hidden_key
71
                                /* Pointer to object base.      */
72
                                /* No longer hidden once object */
73
                                /* is on finalize_now queue.    */
74
#   define fo_next(x) (struct finalizable_object *)((x) -> prolog.next)
75
#   define fo_set_next(x,y) (x) -> prolog.next = (struct hash_chain_entry *)(y)
76
    GC_finalization_proc fo_fn; /* Finalizer.                   */
77
    ptr_t fo_client_data;
78
    word fo_object_size;        /* In bytes.                    */
79
    finalization_mark_proc * fo_mark_proc;      /* Mark-through procedure */
80
} **fo_head = 0;
81
 
82
struct finalizable_object * GC_finalize_now = 0;
83
        /* LIst of objects that should be finalized now.        */
84
 
85
static signed_word log_fo_table_size = -1;
86
 
87
word GC_fo_entries = 0;
88
 
89
void GC_push_finalizer_structures GC_PROTO((void))
90
{
91
    GC_push_all((ptr_t)(&dl_head), (ptr_t)(&dl_head) + sizeof(word));
92
    GC_push_all((ptr_t)(&fo_head), (ptr_t)(&fo_head) + sizeof(word));
93
    GC_push_all((ptr_t)(&GC_finalize_now),
94
                (ptr_t)(&GC_finalize_now) + sizeof(word));
95
}
96
 
97
/* Double the size of a hash table. *size_ptr is the log of its current */
98
/* size.  May be a noop.                                                */
99
/* *table is a pointer to an array of hash headers.  If we succeed, we  */
100
/* update both *table and *log_size_ptr.                                */
101
/* Lock is held.  Signals are disabled.                                 */
102
void GC_grow_table(table, log_size_ptr)
103
struct hash_chain_entry ***table;
104
signed_word * log_size_ptr;
105
{
106
    register word i;
107
    register struct hash_chain_entry *p;
108
    int log_old_size = *log_size_ptr;
109
    register int log_new_size = log_old_size + 1;
110
    word old_size = ((log_old_size == -1)? 0: (1 << log_old_size));
111
    register word new_size = 1 << log_new_size;
112
    struct hash_chain_entry **new_table = (struct hash_chain_entry **)
113
        GC_INTERNAL_MALLOC_IGNORE_OFF_PAGE(
114
                (size_t)new_size * sizeof(struct hash_chain_entry *), NORMAL);
115
 
116
    if (new_table == 0) {
117
        if (table == 0) {
118
            ABORT("Insufficient space for initial table allocation");
119
        } else {
120
            return;
121
        }
122
    }
123
    for (i = 0; i < old_size; i++) {
124
      p = (*table)[i];
125
      while (p != 0) {
126
        register ptr_t real_key = (ptr_t)REVEAL_POINTER(p -> hidden_key);
127
        register struct hash_chain_entry *next = p -> next;
128
        register int new_hash = HASH3(real_key, new_size, log_new_size);
129
 
130
        p -> next = new_table[new_hash];
131
        new_table[new_hash] = p;
132
        p = next;
133
      }
134
    }
135
    *log_size_ptr = log_new_size;
136
    *table = new_table;
137
}
138
 
139
# if defined(__STDC__) || defined(__cplusplus)
140
    int GC_register_disappearing_link(GC_PTR * link)
141
# else
142
    int GC_register_disappearing_link(link)
143
    GC_PTR * link;
144
# endif
145
{
146
    ptr_t base;
147
 
148
    base = (ptr_t)GC_base((GC_PTR)link);
149
    if (base == 0)
150
        ABORT("Bad arg to GC_register_disappearing_link");
151
    return(GC_general_register_disappearing_link(link, base));
152
}
153
 
154
# if defined(__STDC__) || defined(__cplusplus)
155
    int GC_general_register_disappearing_link(GC_PTR * link,
156
                                              GC_PTR obj)
157
# else
158
    int GC_general_register_disappearing_link(link, obj)
159
    GC_PTR * link;
160
    GC_PTR obj;
161
# endif
162
 
163
{
164
    struct disappearing_link *curr_dl;
165
    int index;
166
    struct disappearing_link * new_dl;
167
    DCL_LOCK_STATE;
168
 
169
    if ((word)link & (ALIGNMENT-1))
170
        ABORT("Bad arg to GC_general_register_disappearing_link");
171
#   ifdef THREADS
172
        DISABLE_SIGNALS();
173
        LOCK();
174
#   endif
175
    if (log_dl_table_size == -1
176
        || GC_dl_entries > ((word)1 << log_dl_table_size)) {
177
#       ifndef THREADS
178
            DISABLE_SIGNALS();
179
#       endif
180
        GC_grow_table((struct hash_chain_entry ***)(&dl_head),
181
                      &log_dl_table_size);
182
#       ifdef CONDPRINT
183
          if (GC_print_stats) {
184
            GC_printf1("Grew dl table to %lu entries\n",
185
                        (unsigned long)(1 << log_dl_table_size));
186
          }
187
#       endif
188
#       ifndef THREADS
189
            ENABLE_SIGNALS();
190
#       endif
191
    }
192
    index = HASH2(link, log_dl_table_size);
193
    curr_dl = dl_head[index];
194
    for (curr_dl = dl_head[index]; curr_dl != 0; curr_dl = dl_next(curr_dl)) {
195
        if (curr_dl -> dl_hidden_link == HIDE_POINTER(link)) {
196
            curr_dl -> dl_hidden_obj = HIDE_POINTER(obj);
197
#           ifdef THREADS
198
                UNLOCK();
199
                ENABLE_SIGNALS();
200
#           endif
201
            return(1);
202
        }
203
    }
204
    new_dl = (struct disappearing_link *)
205
        GC_INTERNAL_MALLOC(sizeof(struct disappearing_link),NORMAL);
206
    if (0 == new_dl) {
207
#     ifdef THREADS
208
        UNLOCK();
209
        ENABLE_SIGNALS();
210
#     endif
211
      new_dl = (struct disappearing_link *)
212
              GC_oom_fn(sizeof(struct disappearing_link));
213
      if (0 == new_dl) {
214
        GC_finalization_failures++;
215
        return(0);
216
      }
217
      /* It's not likely we'll make it here, but ... */
218
#     ifdef THREADS
219
        DISABLE_SIGNALS();
220
        LOCK();
221
#     endif
222
    }
223
    new_dl -> dl_hidden_obj = HIDE_POINTER(obj);
224
    new_dl -> dl_hidden_link = HIDE_POINTER(link);
225
    dl_set_next(new_dl, dl_head[index]);
226
    dl_head[index] = new_dl;
227
    GC_dl_entries++;
228
#   ifdef THREADS
229
        UNLOCK();
230
        ENABLE_SIGNALS();
231
#   endif
232
    return(0);
233
}
234
 
235
# if defined(__STDC__) || defined(__cplusplus)
236
    int GC_unregister_disappearing_link(GC_PTR * link)
237
# else
238
    int GC_unregister_disappearing_link(link)
239
    GC_PTR * link;
240
# endif
241
{
242
    struct disappearing_link *curr_dl, *prev_dl;
243
    int index;
244
    DCL_LOCK_STATE;
245
 
246
    DISABLE_SIGNALS();
247
    LOCK();
248
    index = HASH2(link, log_dl_table_size);
249
    if (((unsigned long)link & (ALIGNMENT-1))) goto out;
250
    prev_dl = 0; curr_dl = dl_head[index];
251
    while (curr_dl != 0) {
252
        if (curr_dl -> dl_hidden_link == HIDE_POINTER(link)) {
253
            if (prev_dl == 0) {
254
                dl_head[index] = dl_next(curr_dl);
255
            } else {
256
                dl_set_next(prev_dl, dl_next(curr_dl));
257
            }
258
            GC_dl_entries--;
259
            UNLOCK();
260
            ENABLE_SIGNALS();
261
#           ifdef DBG_HDRS_ALL
262
              dl_set_next(curr_dl, 0);
263
#           else
264
              GC_free((GC_PTR)curr_dl);
265
#           endif
266
            return(1);
267
        }
268
        prev_dl = curr_dl;
269
        curr_dl = dl_next(curr_dl);
270
    }
271
out:
272
    UNLOCK();
273
    ENABLE_SIGNALS();
274
    return(0);
275
}
276
 
277
/* Possible finalization_marker procedures.  Note that mark stack       */
278
/* overflow is handled by the caller, and is not a disaster.            */
279
GC_API void GC_normal_finalize_mark_proc(p)
280
ptr_t p;
281
{
282
    hdr * hhdr = HDR(p);
283
 
284
    PUSH_OBJ((word *)p, hhdr, GC_mark_stack_top,
285
             &(GC_mark_stack[GC_mark_stack_size]));
286
}
287
 
288
/* This only pays very partial attention to the mark descriptor.        */
289
/* It does the right thing for normal and atomic objects, and treats    */
290
/* most others as normal.                                               */
291
GC_API void GC_ignore_self_finalize_mark_proc(p)
292
ptr_t p;
293
{
294
    hdr * hhdr = HDR(p);
295
    word descr = hhdr -> hb_descr;
296
    ptr_t q, r;
297
    ptr_t scan_limit;
298
    ptr_t target_limit = p + WORDS_TO_BYTES(hhdr -> hb_sz) - 1;
299
 
300
    if ((descr & GC_DS_TAGS) == GC_DS_LENGTH) {
301
       scan_limit = p + descr - sizeof(word);
302
    } else {
303
       scan_limit = target_limit + 1 - sizeof(word);
304
    }
305
    for (q = p; q <= scan_limit; q += ALIGNMENT) {
306
        r = *(ptr_t *)q;
307
        if (r < p || r > target_limit) {
308
            GC_PUSH_ONE_HEAP((word)r, q);
309
        }
310
    }
311
}
312
 
313
/*ARGSUSED*/
314
GC_API void GC_null_finalize_mark_proc(p)
315
ptr_t p;
316
{
317
}
318
 
319
/* Possible finalization_marker procedures.  Note that mark stack       */
320
/* overflow is handled by the caller, and is not a disaster.            */
321
GC_API void GC_unreachable_finalize_mark_proc(p)
322
ptr_t p;
323
{
324
    return GC_normal_finalize_mark_proc(p);
325
}
326
 
327
 
328
 
329
/* Register a finalization function.  See gc.h for details.     */
330
/* in the nonthreads case, we try to avoid disabling signals,   */
331
/* since it can be expensive.  Threads packages typically       */
332
/* make it cheaper.                                             */
333
/* The last parameter is a procedure that determines            */
334
/* marking for finalization ordering.  Any objects marked       */
335
/* by that procedure will be guaranteed to not have been        */
336
/* finalized when this finalizer is invoked.                    */
337
GC_API void GC_register_finalizer_inner(obj, fn, cd, ofn, ocd, mp)
338
GC_PTR obj;
339
GC_finalization_proc fn;
340
GC_PTR cd;
341
GC_finalization_proc * ofn;
342
GC_PTR * ocd;
343
finalization_mark_proc * mp;
344
{
345
    ptr_t base;
346
    struct finalizable_object * curr_fo, * prev_fo;
347
    int index;
348
    struct finalizable_object *new_fo;
349
    hdr *hhdr;
350
    DCL_LOCK_STATE;
351
 
352
#   ifdef THREADS
353
        DISABLE_SIGNALS();
354
        LOCK();
355
#   endif
356
    if (log_fo_table_size == -1
357
        || GC_fo_entries > ((word)1 << log_fo_table_size)) {
358
#       ifndef THREADS
359
            DISABLE_SIGNALS();
360
#       endif
361
        GC_grow_table((struct hash_chain_entry ***)(&fo_head),
362
                      &log_fo_table_size);
363
#       ifdef CONDPRINT
364
          if (GC_print_stats) {
365
            GC_printf1("Grew fo table to %lu entries\n",
366
                        (unsigned long)(1 << log_fo_table_size));
367
          }
368
#       endif
369
#       ifndef THREADS
370
            ENABLE_SIGNALS();
371
#       endif
372
    }
373
    /* in the THREADS case signals are disabled and we hold allocation  */
374
    /* lock; otherwise neither is true.  Proceed carefully.             */
375
    base = (ptr_t)obj;
376
    index = HASH2(base, log_fo_table_size);
377
    prev_fo = 0; curr_fo = fo_head[index];
378
    while (curr_fo != 0) {
379
        if (curr_fo -> fo_hidden_base == HIDE_POINTER(base)) {
380
            /* Interruption by a signal in the middle of this   */
381
            /* should be safe.  The client may see only *ocd    */
382
            /* updated, but we'll declare that to be his        */
383
            /* problem.                                         */
384
            if (ocd) *ocd = (GC_PTR) curr_fo -> fo_client_data;
385
            if (ofn) *ofn = curr_fo -> fo_fn;
386
            /* Delete the structure for base. */
387
                if (prev_fo == 0) {
388
                  fo_head[index] = fo_next(curr_fo);
389
                } else {
390
                  fo_set_next(prev_fo, fo_next(curr_fo));
391
                }
392
            if (fn == 0) {
393
                GC_fo_entries--;
394
                  /* May not happen if we get a signal.  But a high     */
395
                  /* estimate will only make the table larger than      */
396
                  /* necessary.                                         */
397
#               if !defined(THREADS) && !defined(DBG_HDRS_ALL)
398
                  GC_free((GC_PTR)curr_fo);
399
#               endif
400
            } else {
401
                curr_fo -> fo_fn = fn;
402
                curr_fo -> fo_client_data = (ptr_t)cd;
403
                curr_fo -> fo_mark_proc = mp;
404
                /* Reinsert it.  We deleted it first to maintain        */
405
                /* consistency in the event of a signal.                */
406
                if (prev_fo == 0) {
407
                  fo_head[index] = curr_fo;
408
                } else {
409
                  fo_set_next(prev_fo, curr_fo);
410
                }
411
            }
412
#           ifdef THREADS
413
                UNLOCK();
414
                ENABLE_SIGNALS();
415
#           endif
416
            return;
417
        }
418
        prev_fo = curr_fo;
419
        curr_fo = fo_next(curr_fo);
420
    }
421
    if (ofn) *ofn = 0;
422
    if (ocd) *ocd = 0;
423
    if (fn == 0) {
424
#       ifdef THREADS
425
            UNLOCK();
426
            ENABLE_SIGNALS();
427
#       endif
428
        return;
429
    }
430
    GET_HDR(base, hhdr);
431
    if (0 == hhdr) {
432
      /* We won't collect it, hence finalizer wouldn't be run. */
433
#     ifdef THREADS
434
          UNLOCK();
435
          ENABLE_SIGNALS();
436
#     endif
437
      return;
438
    }
439
    new_fo = (struct finalizable_object *)
440
        GC_INTERNAL_MALLOC(sizeof(struct finalizable_object),NORMAL);
441
    if (0 == new_fo) {
442
#     ifdef THREADS
443
        UNLOCK();
444
        ENABLE_SIGNALS();
445
#     endif
446
      new_fo = (struct finalizable_object *)
447
              GC_oom_fn(sizeof(struct finalizable_object));
448
      if (0 == new_fo) {
449
        GC_finalization_failures++;
450
        return;
451
      }
452
      /* It's not likely we'll make it here, but ... */
453
#     ifdef THREADS
454
        DISABLE_SIGNALS();
455
        LOCK();
456
#     endif
457
    }
458
    new_fo -> fo_hidden_base = (word)HIDE_POINTER(base);
459
    new_fo -> fo_fn = fn;
460
    new_fo -> fo_client_data = (ptr_t)cd;
461
    new_fo -> fo_object_size = hhdr -> hb_sz;
462
    new_fo -> fo_mark_proc = mp;
463
    fo_set_next(new_fo, fo_head[index]);
464
    GC_fo_entries++;
465
    fo_head[index] = new_fo;
466
#   ifdef THREADS
467
        UNLOCK();
468
        ENABLE_SIGNALS();
469
#   endif
470
}
471
 
472
# if defined(__STDC__)
473
    void GC_register_finalizer(void * obj,
474
                               GC_finalization_proc fn, void * cd,
475
                               GC_finalization_proc *ofn, void ** ocd)
476
# else
477
    void GC_register_finalizer(obj, fn, cd, ofn, ocd)
478
    GC_PTR obj;
479
    GC_finalization_proc fn;
480
    GC_PTR cd;
481
    GC_finalization_proc * ofn;
482
    GC_PTR * ocd;
483
# endif
484
{
485
    GC_register_finalizer_inner(obj, fn, cd, ofn,
486
                                ocd, GC_normal_finalize_mark_proc);
487
}
488
 
489
# if defined(__STDC__)
490
    void GC_register_finalizer_ignore_self(void * obj,
491
                               GC_finalization_proc fn, void * cd,
492
                               GC_finalization_proc *ofn, void ** ocd)
493
# else
494
    void GC_register_finalizer_ignore_self(obj, fn, cd, ofn, ocd)
495
    GC_PTR obj;
496
    GC_finalization_proc fn;
497
    GC_PTR cd;
498
    GC_finalization_proc * ofn;
499
    GC_PTR * ocd;
500
# endif
501
{
502
    GC_register_finalizer_inner(obj, fn, cd, ofn,
503
                                ocd, GC_ignore_self_finalize_mark_proc);
504
}
505
 
506
# if defined(__STDC__)
507
    void GC_register_finalizer_no_order(void * obj,
508
                               GC_finalization_proc fn, void * cd,
509
                               GC_finalization_proc *ofn, void ** ocd)
510
# else
511
    void GC_register_finalizer_no_order(obj, fn, cd, ofn, ocd)
512
    GC_PTR obj;
513
    GC_finalization_proc fn;
514
    GC_PTR cd;
515
    GC_finalization_proc * ofn;
516
    GC_PTR * ocd;
517
# endif
518
{
519
    GC_register_finalizer_inner(obj, fn, cd, ofn,
520
                                ocd, GC_null_finalize_mark_proc);
521
}
522
 
523
# if defined(__STDC__)
524
    void GC_register_finalizer_unreachable(void * obj,
525
                               GC_finalization_proc fn, void * cd,
526
                               GC_finalization_proc *ofn, void ** ocd)
527
# else
528
    void GC_register_finalizer_unreachable(obj, fn, cd, ofn, ocd)
529
    GC_PTR obj;
530
    GC_finalization_proc fn;
531
    GC_PTR cd;
532
    GC_finalization_proc * ofn;
533
    GC_PTR * ocd;
534
# endif
535
{
536
    GC_register_finalizer_inner(obj, fn, cd, ofn,
537
                                ocd, GC_unreachable_finalize_mark_proc);
538
}
539
 
540
#ifndef NO_DEBUGGING
541
void GC_dump_finalization()
542
{
543
    struct disappearing_link * curr_dl;
544
    struct finalizable_object * curr_fo;
545
    ptr_t real_ptr, real_link;
546
    int dl_size = (log_dl_table_size == -1 ) ? 0 : (1 << log_dl_table_size);
547
    int fo_size = (log_fo_table_size == -1 ) ? 0 : (1 << log_fo_table_size);
548
    int i;
549
 
550
    GC_printf0("Disappearing links:\n");
551
    for (i = 0; i < dl_size; i++) {
552
      for (curr_dl = dl_head[i]; curr_dl != 0; curr_dl = dl_next(curr_dl)) {
553
        real_ptr = (ptr_t)REVEAL_POINTER(curr_dl -> dl_hidden_obj);
554
        real_link = (ptr_t)REVEAL_POINTER(curr_dl -> dl_hidden_link);
555
        GC_printf2("Object: 0x%lx, Link:0x%lx\n", real_ptr, real_link);
556
      }
557
    }
558
    GC_printf0("Finalizers:\n");
559
    for (i = 0; i < fo_size; i++) {
560
      for (curr_fo = fo_head[i]; curr_fo != 0; curr_fo = fo_next(curr_fo)) {
561
        real_ptr = (ptr_t)REVEAL_POINTER(curr_fo -> fo_hidden_base);
562
        GC_printf1("Finalizable object: 0x%lx\n", real_ptr);
563
      }
564
    }
565
}
566
#endif
567
 
568
/* Called with world stopped.  Cause disappearing links to disappear,   */
569
/* and invoke finalizers.                                               */
570
void GC_finalize()
571
{
572
    struct disappearing_link * curr_dl, * prev_dl, * next_dl;
573
    struct finalizable_object * curr_fo, * prev_fo, * next_fo;
574
    ptr_t real_ptr, real_link;
575
    register int i;
576
    int dl_size = (log_dl_table_size == -1 ) ? 0 : (1 << log_dl_table_size);
577
    int fo_size = (log_fo_table_size == -1 ) ? 0 : (1 << log_fo_table_size);
578
 
579
  /* Make disappearing links disappear */
580
    for (i = 0; i < dl_size; i++) {
581
      curr_dl = dl_head[i];
582
      prev_dl = 0;
583
      while (curr_dl != 0) {
584
        real_ptr = (ptr_t)REVEAL_POINTER(curr_dl -> dl_hidden_obj);
585
        real_link = (ptr_t)REVEAL_POINTER(curr_dl -> dl_hidden_link);
586
        if (!GC_is_marked(real_ptr)) {
587
            *(word *)real_link = 0;
588
            next_dl = dl_next(curr_dl);
589
            if (prev_dl == 0) {
590
                dl_head[i] = next_dl;
591
            } else {
592
                dl_set_next(prev_dl, next_dl);
593
            }
594
            GC_clear_mark_bit((ptr_t)curr_dl);
595
            GC_dl_entries--;
596
            curr_dl = next_dl;
597
        } else {
598
            prev_dl = curr_dl;
599
            curr_dl = dl_next(curr_dl);
600
        }
601
      }
602
    }
603
  /* Mark all objects reachable via chains of 1 or more pointers        */
604
  /* from finalizable objects.                                          */
605
    GC_ASSERT(GC_mark_state == MS_NONE);
606
    for (i = 0; i < fo_size; i++) {
607
      for (curr_fo = fo_head[i]; curr_fo != 0; curr_fo = fo_next(curr_fo)) {
608
        real_ptr = (ptr_t)REVEAL_POINTER(curr_fo -> fo_hidden_base);
609
        if (!GC_is_marked(real_ptr)) {
610
            GC_MARKED_FOR_FINALIZATION(real_ptr);
611
            GC_MARK_FO(real_ptr, curr_fo -> fo_mark_proc);
612
            if (GC_is_marked(real_ptr)) {
613
                WARN("Finalization cycle involving %lx\n", real_ptr);
614
            }
615
        }
616
      }
617
    }
618
  /* Enqueue for finalization all objects that are still                */
619
  /* unreachable.                                                       */
620
    GC_words_finalized = 0;
621
    for (i = 0; i < fo_size; i++) {
622
      curr_fo = fo_head[i];
623
      prev_fo = 0;
624
      while (curr_fo != 0) {
625
        real_ptr = (ptr_t)REVEAL_POINTER(curr_fo -> fo_hidden_base);
626
        if (!GC_is_marked(real_ptr)) {
627
            if (!GC_java_finalization) {
628
              GC_set_mark_bit(real_ptr);
629
            }
630
            /* Delete from hash table */
631
              next_fo = fo_next(curr_fo);
632
              if (prev_fo == 0) {
633
                fo_head[i] = next_fo;
634
              } else {
635
                fo_set_next(prev_fo, next_fo);
636
              }
637
              GC_fo_entries--;
638
            /* Add to list of objects awaiting finalization.    */
639
              fo_set_next(curr_fo, GC_finalize_now);
640
              GC_finalize_now = curr_fo;
641
              /* unhide object pointer so any future collections will   */
642
              /* see it.                                                */
643
              curr_fo -> fo_hidden_base =
644
                        (word) REVEAL_POINTER(curr_fo -> fo_hidden_base);
645
              GC_words_finalized +=
646
                        ALIGNED_WORDS(curr_fo -> fo_object_size)
647
                        + ALIGNED_WORDS(sizeof(struct finalizable_object));
648
            GC_ASSERT(GC_is_marked(GC_base((ptr_t)curr_fo)));
649
            curr_fo = next_fo;
650
        } else {
651
            prev_fo = curr_fo;
652
            curr_fo = fo_next(curr_fo);
653
        }
654
      }
655
    }
656
 
657
  if (GC_java_finalization) {
658
    /* make sure we mark everything reachable from objects finalized
659
       using the no_order mark_proc */
660
      for (curr_fo = GC_finalize_now;
661
         curr_fo != NULL; curr_fo = fo_next(curr_fo)) {
662
        real_ptr = (ptr_t)curr_fo -> fo_hidden_base;
663
        if (!GC_is_marked(real_ptr)) {
664
            if (curr_fo -> fo_mark_proc == GC_null_finalize_mark_proc) {
665
                GC_MARK_FO(real_ptr, GC_normal_finalize_mark_proc);
666
            }
667
            if (curr_fo -> fo_mark_proc != GC_unreachable_finalize_mark_proc) {
668
                GC_set_mark_bit(real_ptr);
669
            }
670
        }
671
      }
672
 
673
      /* now revive finalize-when-unreachable objects reachable from
674
         other finalizable objects */
675
      curr_fo = GC_finalize_now;
676
      prev_fo = 0;
677
      while (curr_fo != 0) {
678
        next_fo = fo_next(curr_fo);
679
        if (curr_fo -> fo_mark_proc == GC_unreachable_finalize_mark_proc) {
680
          real_ptr = (ptr_t)curr_fo -> fo_hidden_base;
681
          if (!GC_is_marked(real_ptr)) {
682
              GC_set_mark_bit(real_ptr);
683
          } else {
684
              if (prev_fo == 0)
685
                GC_finalize_now = next_fo;
686
              else
687
                fo_set_next(prev_fo, next_fo);
688
 
689
              curr_fo -> fo_hidden_base =
690
                        (word) HIDE_POINTER(curr_fo -> fo_hidden_base);
691
              GC_words_finalized -=
692
                        ALIGNED_WORDS(curr_fo -> fo_object_size)
693
                        + ALIGNED_WORDS(sizeof(struct finalizable_object));
694
 
695
              i = HASH2(real_ptr, log_fo_table_size);
696
              fo_set_next (curr_fo, fo_head[i]);
697
              GC_fo_entries++;
698
              fo_head[i] = curr_fo;
699
              curr_fo = prev_fo;
700
          }
701
        }
702
        prev_fo = curr_fo;
703
        curr_fo = next_fo;
704
      }
705
  }
706
 
707
  /* Remove dangling disappearing links. */
708
    for (i = 0; i < dl_size; i++) {
709
      curr_dl = dl_head[i];
710
      prev_dl = 0;
711
      while (curr_dl != 0) {
712
        real_link = GC_base((ptr_t)REVEAL_POINTER(curr_dl -> dl_hidden_link));
713
        if (real_link != 0 && !GC_is_marked(real_link)) {
714
            next_dl = dl_next(curr_dl);
715
            if (prev_dl == 0) {
716
                dl_head[i] = next_dl;
717
            } else {
718
                dl_set_next(prev_dl, next_dl);
719
            }
720
            GC_clear_mark_bit((ptr_t)curr_dl);
721
            GC_dl_entries--;
722
            curr_dl = next_dl;
723
        } else {
724
            prev_dl = curr_dl;
725
            curr_dl = dl_next(curr_dl);
726
        }
727
      }
728
    }
729
}
730
 
731
#ifndef JAVA_FINALIZATION_NOT_NEEDED
732
 
733
/* Enqueue all remaining finalizers to be run - Assumes lock is
734
 * held, and signals are disabled */
735
void GC_enqueue_all_finalizers()
736
{
737
    struct finalizable_object * curr_fo, * prev_fo, * next_fo;
738
    ptr_t real_ptr;
739
    register int i;
740
    int fo_size;
741
 
742
    fo_size = (log_fo_table_size == -1 ) ? 0 : (1 << log_fo_table_size);
743
    GC_words_finalized = 0;
744
    for (i = 0; i < fo_size; i++) {
745
        curr_fo = fo_head[i];
746
        prev_fo = 0;
747
      while (curr_fo != 0) {
748
          real_ptr = (ptr_t)REVEAL_POINTER(curr_fo -> fo_hidden_base);
749
          GC_MARK_FO(real_ptr, GC_normal_finalize_mark_proc);
750
          GC_set_mark_bit(real_ptr);
751
 
752
          /* Delete from hash table */
753
          next_fo = fo_next(curr_fo);
754
          if (prev_fo == 0) {
755
              fo_head[i] = next_fo;
756
          } else {
757
              fo_set_next(prev_fo, next_fo);
758
          }
759
          GC_fo_entries--;
760
 
761
          /* Add to list of objects awaiting finalization.      */
762
          fo_set_next(curr_fo, GC_finalize_now);
763
          GC_finalize_now = curr_fo;
764
 
765
          /* unhide object pointer so any future collections will       */
766
          /* see it.                                            */
767
          curr_fo -> fo_hidden_base =
768
                        (word) REVEAL_POINTER(curr_fo -> fo_hidden_base);
769
 
770
          GC_words_finalized +=
771
                ALIGNED_WORDS(curr_fo -> fo_object_size)
772
                        + ALIGNED_WORDS(sizeof(struct finalizable_object));
773
          curr_fo = next_fo;
774
        }
775
    }
776
 
777
    return;
778
}
779
 
780
/* Invoke all remaining finalizers that haven't yet been run.
781
 * This is needed for strict compliance with the Java standard,
782
 * which can make the runtime guarantee that all finalizers are run.
783
 * Unfortunately, the Java standard implies we have to keep running
784
 * finalizers until there are no more left, a potential infinite loop.
785
 * YUCK.
786
 * Note that this is even more dangerous than the usual Java
787
 * finalizers, in that objects reachable from static variables
788
 * may have been finalized when these finalizers are run.
789
 * Finalizers run at this point must be prepared to deal with a
790
 * mostly broken world.
791
 * This routine is externally callable, so is called without
792
 * the allocation lock.
793
 */
794
GC_API void GC_finalize_all()
795
{
796
    DCL_LOCK_STATE;
797
 
798
    DISABLE_SIGNALS();
799
    LOCK();
800
    while (GC_fo_entries > 0) {
801
      GC_enqueue_all_finalizers();
802
      UNLOCK();
803
      ENABLE_SIGNALS();
804
      GC_INVOKE_FINALIZERS();
805
      DISABLE_SIGNALS();
806
      LOCK();
807
    }
808
    UNLOCK();
809
    ENABLE_SIGNALS();
810
}
811
#endif
812
 
813
/* Returns true if it is worth calling GC_invoke_finalizers. (Useful if */
814
/* finalizers can only be called from some kind of `safe state' and     */
815
/* getting into that safe state is expensive.)                          */
816
int GC_should_invoke_finalizers GC_PROTO((void))
817
{
818
    return GC_finalize_now != 0;
819
}
820
 
821
/* Invoke finalizers for all objects that are ready to be finalized.    */
822
/* Should be called without allocation lock.                            */
823
int GC_invoke_finalizers()
824
{
825
    struct finalizable_object * curr_fo;
826
    int count = 0;
827
    word mem_freed_before;
828
    DCL_LOCK_STATE;
829
 
830
    while (GC_finalize_now != 0) {
831
#       ifdef THREADS
832
            DISABLE_SIGNALS();
833
            LOCK();
834
#       endif
835
        if (count == 0) {
836
            mem_freed_before = GC_mem_freed;
837
        }
838
        curr_fo = GC_finalize_now;
839
#       ifdef THREADS
840
            if (curr_fo != 0) GC_finalize_now = fo_next(curr_fo);
841
            UNLOCK();
842
            ENABLE_SIGNALS();
843
            if (curr_fo == 0) break;
844
#       else
845
            GC_finalize_now = fo_next(curr_fo);
846
#       endif
847
        fo_set_next(curr_fo, 0);
848
        (*(curr_fo -> fo_fn))((ptr_t)(curr_fo -> fo_hidden_base),
849
                              curr_fo -> fo_client_data);
850
        curr_fo -> fo_client_data = 0;
851
        ++count;
852
#       ifdef UNDEFINED
853
            /* This is probably a bad idea.  It throws off accounting if */
854
            /* nearly all objects are finalizable.  O.w. it shouldn't    */
855
            /* matter.                                                   */
856
            GC_free((GC_PTR)curr_fo);
857
#       endif
858
    }
859
    if (count != 0 && mem_freed_before != GC_mem_freed) {
860
        LOCK();
861
        GC_finalizer_mem_freed += (GC_mem_freed - mem_freed_before);
862
        UNLOCK();
863
    }
864
    return count;
865
}
866
 
867
void (* GC_finalizer_notifier)() = (void (*) GC_PROTO((void)))0;
868
 
869
static GC_word last_finalizer_notification = 0;
870
 
871
void GC_notify_or_invoke_finalizers GC_PROTO((void))
872
{
873
    /* This is a convenient place to generate backtraces if appropriate, */
874
    /* since that code is not callable with the allocation lock.         */
875
#   if defined(KEEP_BACK_PTRS) || defined(MAKE_BACK_GRAPH)
876
      static word last_back_trace_gc_no = 1;    /* Skip first one. */
877
 
878
      if (GC_gc_no > last_back_trace_gc_no) {
879
        word i;
880
 
881
#       ifdef KEEP_BACK_PTRS
882
          LOCK();
883
          /* Stops when GC_gc_no wraps; that's OK.      */
884
          last_back_trace_gc_no = (word)(-1);  /* disable others. */
885
          for (i = 0; i < GC_backtraces; ++i) {
886
              /* FIXME: This tolerates concurrent heap mutation,        */
887
              /* which may cause occasional mysterious results.         */
888
              /* We need to release the GC lock, since GC_print_callers */
889
              /* acquires it.  It probably shouldn't.                   */
890
              UNLOCK();
891
              GC_generate_random_backtrace_no_gc();
892
              LOCK();
893
          }
894
          last_back_trace_gc_no = GC_gc_no;
895
          UNLOCK();
896
#       endif
897
#       ifdef MAKE_BACK_GRAPH
898
          if (GC_print_back_height)
899
            GC_print_back_graph_stats();
900
#       endif
901
      }
902
#   endif
903
    if (GC_finalize_now == 0) return;
904
    if (!GC_finalize_on_demand) {
905
        (void) GC_invoke_finalizers();
906
#       ifndef THREADS
907
          GC_ASSERT(GC_finalize_now == 0);
908
#       endif   /* Otherwise GC can run concurrently and add more */
909
        return;
910
    }
911
    if (GC_finalizer_notifier != (void (*) GC_PROTO((void)))0
912
        && last_finalizer_notification != GC_gc_no) {
913
        last_finalizer_notification = GC_gc_no;
914
        GC_finalizer_notifier();
915
    }
916
}
917
 
918
# ifdef __STDC__
919
    GC_PTR GC_call_with_alloc_lock(GC_fn_type fn,
920
                                         GC_PTR client_data)
921
# else
922
    GC_PTR GC_call_with_alloc_lock(fn, client_data)
923
    GC_fn_type fn;
924
    GC_PTR client_data;
925
# endif
926
{
927
    GC_PTR result;
928
    DCL_LOCK_STATE;
929
 
930
#   ifdef THREADS
931
      DISABLE_SIGNALS();
932
      LOCK();
933
      SET_LOCK_HOLDER();
934
#   endif
935
    result = (*fn)(client_data);
936
#   ifdef THREADS
937
#     ifndef GC_ASSERTIONS
938
        UNSET_LOCK_HOLDER();
939
#     endif /* o.w. UNLOCK() does it implicitly */
940
      UNLOCK();
941
      ENABLE_SIGNALS();
942
#   endif
943
    return(result);
944
}
945
 
946
#if !defined(NO_DEBUGGING)
947
 
948
void GC_print_finalization_stats()
949
{
950
    struct finalizable_object *fo = GC_finalize_now;
951
    size_t ready = 0;
952
 
953
    GC_printf2("%lu finalization table entries; %lu disappearing links\n",
954
               GC_fo_entries, GC_dl_entries);
955
    for (; 0 != fo; fo = fo_next(fo)) ++ready;
956
    GC_printf1("%lu objects are eligible for immediate finalization\n", ready);
957
}
958
 
959
#endif /* NO_DEBUGGING */

powered by: WebSVN 2.1.0

© copyright 1999-2024 OpenCores.org, equivalent to Oliscience, all rights reserved. OpenCores®, registered trademark.