racket/c/vfasl.c
Matthew Flatt f78dc5724e add pb (portable bytecode) backend
This commit does four things:

 * Adds "pb.ss" and "pb.c", which implement a portable bytecode
   backend and interpreter that is intended for bootstrapping. A
   single set of pb bootfiles can support bootstrapping on all
   platforms --- as long as the C compiler supports a 64-bit integer
   type. The pb machine supports foreign calls for only a small set of
   recognized prototypes, and it does not support foriegn callables.
   Use `./configure --pb` to build the pb variant.

 * Changes the kernel's casts between `ptr` and `void*` types. In a pb
   build, the `ptr` type can be a 64-bit integer type while `void*` is
   a 32-bit pointer type, so casts must go through an intermediate
   integer type.

 * Adjusts the compiler to accomodate run-time-determined endianness.
   Making the compiler agnostic to word size is not practical, but
   only a few pieces depend on the target machine's endianness, and
   those can generally be deferred to a run-time choice of byte-based
   operations. The one exception is that ftype bit fields are not
   allowed unless accompanied by an explicit endianness declaration.

 * Start reducing duplication among platform-specific makefiles. For
   example, `Mf-ta6osx` chains to `Mf-a6osx` to avoid repeating most
   of it. A lot more can be done here.

original commit: 97533fa9d8b8400b0dc1a890768c7d30c91257e0
2020-07-24 13:13:46 -06:00

1475 lines
42 KiB
C

/* vfasl.c
* Copyright 1984-2017 Cisco Systems, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include "system.h"
#include "popcount.h"
/*
vfasl ("very fast load") format, where "data" corresponds to an
image to load into memory, "table" is metadata to relocate that
data, and the fixed-size header determines the overall size. The
data can be loaded directly to the static generation on boot, since
it's organized into pieces that should reside in a particular
space.
[vfasl_header]
_
/ [symbol] ... -> space_symbol
/ [rtd] ... -> space_pure
d | [closure] ... -> space_pure
a | [impure] ... -> space_impure
t | [pure_typed] ... -> space_pure_typed
a | [impure_record] ... -> space_impure_record
| [code] ... -> space_code
\ [data] ... -> space_data
\_ [reloc] ... -> (not kept for direct-to-static)
_
t / [symbol reference offset] ...
a / [rtd reference offset] ...
b | [singleton reference offset] ...
l \
e \_ [bitmap of pointers to relocate]
The bitmap at the end has one bit for each pointer-sized word in
the data, but it's shorter than the "data" size (divided by the
pointer size then divided by 8 bits per byte) because the trailing
zeros for data are omitted. The bitmap doesn't include some fixups
that are handled more directly, such as for code and its
relocations.
*/
typedef uptr vfoff;
/* Similar to allocation spaces, but not all allocation spaces are
represented, and these spaces are more fine-grained in some
cases: */
enum {
vspace_symbol,
vspace_rtd,
vspace_closure,
vspace_impure,
vspace_pure_typed,
vspace_impure_record,
/* rest rest are at then end to make the pointer bitmap
end with zeros (that can be dropped): */
vspace_code,
vspace_data,
vspace_reloc, /* can be dropped after direct to static generation */
vspaces_count
};
/* Needs to match order above, maps vfasl spaces to allocation
spaces: */
static ISPC vspace_spaces[] = {
space_symbol,
space_pure, /* rtd */
space_pure, /* closure */
space_impure,
space_pure_typed_object,
space_impure_record,
space_code,
space_data,
space_data /* reloc --- but not really, since relocs are never in static */
};
typedef struct vfasl_header {
vfoff data_size;
vfoff table_size;
vfoff result_offset;
/* first starting offset is 0, so skip it in this array: */
vfoff vspace_rel_offsets[vspaces_count-1];
vfoff symref_count;
vfoff rtdref_count;
vfoff singletonref_count;
} vfasl_header;
/************************************************************/
/* Encode-time data structures */
/* During encoding, we use many chunks per vspace on first pass, one
per vspace on second pass: */
typedef struct vfasl_chunk {
ptr bytes;
uptr length;
uptr used;
uptr swept;
struct vfasl_chunk *next, *prev;
} vfasl_chunk;
/* One per vspace: */
struct vfasl_count_and_chunk {
uptr total_bytes;
vfasl_chunk *first;
};
typedef struct vfasl_info {
ptr base_addr; /* address to make relocations relative to */
uptr sym_count;
vfoff symref_count;
vfoff *symrefs;
ptr base_rtd; /* track replacement base_rtd to recognize other rtds */
vfoff rtdref_count;
vfoff *rtdrefs;
vfoff singletonref_count;
vfoff *singletonrefs;
struct vfasl_count_and_chunk spaces[vspaces_count];
octet *ptr_bitmap;
struct vfasl_hash_table *graph;
IBOOL installs_library_entry; /* to determine whether vfasls can be combined */
} vfasl_info;
#define ptr_add(p, n) ((ptr)((uptr)(p) + (n)))
#define ptr_subtract(p, n) ((ptr)((uptr)(p) - (n)))
#define ptr_diff(p, q) ((uptr)(p) - (uptr)(q))
#define byte_bits 8
#define log2_byte_bits 3
#define segment_align(size) (((size)+bytes_per_segment-1) & ~(bytes_per_segment-1))
static uptr symbol_pos_to_offset(uptr sym_pos) {
uptr syms_per_segment = bytes_per_segment / size_symbol;
uptr segs = sym_pos / syms_per_segment;
uptr syms = sym_pos - (segs * syms_per_segment);
return (segs * bytes_per_segment) + (syms * size_symbol);
}
static ptr vfasl_copy_all(vfasl_info *vfi, ptr v);
static ptr copy(vfasl_info *vfi, ptr pp, seginfo *si);
static uptr sweep(vfasl_info *vfi, ptr p);
static int is_rtd(ptr tf, vfasl_info *vfi);
static IFASLCODE abs_reloc_variant(IFASLCODE type);
static ptr vfasl_encode_relocation(vfasl_info *vfi, ptr obj);
static void relink_code(ptr co, ptr sym_base, ptr *vspaces, uptr *vspace_offsets, IBOOL to_static);
static ptr find_pointer_from_offset(uptr p_off, ptr *vspaces, uptr *vspace_offsets);
static void vfasl_relocate(vfasl_info *vfi, ptr *ppp);
static ptr vfasl_relocate_help(vfasl_info *vfi, ptr pp);
static ptr vfasl_relocate_code(vfasl_info *vfi, ptr code);
static ptr vfasl_find_room(vfasl_info *vfi, int s, ITYPE t, iptr n);
static void vfasl_register_rtd_reference(vfasl_info *vfi, ptr pp);
static void vfasl_register_symbol_reference(vfasl_info *vfi, ptr *pp, ptr p);
static void vfasl_register_singleton_reference(vfasl_info *vfi, ptr *pp, int which);
static void vfasl_register_forward(vfasl_info *vfi, ptr pp, ptr p);
static ptr vfasl_lookup_forward(vfasl_info *vfi, ptr p);
static iptr vfasl_symbol_to_index(vfasl_info *vfi, ptr pp);
static void fasl_init_entry_tables();
static void vfasl_check_install_library_entry(vfasl_info *vfi, ptr name);
static int detect_singleton(ptr p);
static ptr lookup_singleton(int which);
typedef struct vfasl_hash_table vfasl_hash_table;
static vfasl_hash_table *make_vfasl_hash_table(IBOOL permanent);
static void vfasl_hash_table_set(vfasl_hash_table *ht, ptr key, ptr value);
static ptr vfasl_hash_table_ref(vfasl_hash_table *ht, ptr key);
static void *vfasl_malloc(uptr sz);
static void *vfasl_calloc(uptr sz, uptr n);
static void sort_offsets(vfoff *p, vfoff len);
#define vfasl_fail(vfi, what) S_error("vfasl", "cannot encode " what)
/************************************************************/
/* Loading */
ptr S_vfasl(ptr bv, void *stream, iptr offset, iptr input_len)
{
ptr vspaces[vspaces_count];
uptr vspace_offsets[vspaces_count+1];
# define VSPACE_LENGTH(s) (vspace_offsets[(s)+1] - vspace_offsets[(s)])
# define VSPACE_END(s) ptr_add(vspaces[(s)], VSPACE_LENGTH(s))
ptr tc = get_thread_context();
vfasl_header header;
ptr data, table;
vfoff *symrefs, *rtdrefs, *singletonrefs;
octet *bm, *bm_end;
iptr used_len;
int s;
IBOOL to_static = 0;
used_len = sizeof(header);
if (used_len > input_len)
S_error("fasl-read", "input length mismatch");
if (bv)
memcpy(&header, &BVIT(bv, offset), sizeof(vfasl_header));
else {
if (S_fasl_stream_read(stream, (octet*)&header, sizeof(header)) < 0)
S_error("fasl-read", "input truncated");
}
used_len += header.data_size + header.table_size;
if (used_len > input_len)
S_error("fasl-read", "input length mismatch");
vspace_offsets[0] = 0;
for (s = 1; s < vspaces_count; s++) {
vspace_offsets[s] = header.vspace_rel_offsets[s-1];
}
vspace_offsets[vspaces_count] = header.data_size;
if (bv) {
void *base_addr = &BVIT(bv, sizeof(vfasl_header) + offset);
thread_find_room(tc, typemod, header.data_size, data);
memcpy(TO_VOIDP(data), base_addr, header.data_size);
table = ptr_add(TO_PTR(base_addr), header.data_size);
} else {
if (S_vfasl_boot_mode > 0) {
for (s = 0; s < vspaces_count; s++) {
uptr sz = vspace_offsets[s+1] - vspace_offsets[s];
if (sz > 0) {
if ((s == vspace_reloc) && !S_G.retain_static_relocation) {
thread_find_room(tc, typemod, sz, vspaces[s])
} else {
find_room(vspace_spaces[s], static_generation, typemod, sz, vspaces[s])
}
if (S_fasl_stream_read(stream, TO_VOIDP(vspaces[s]), sz) < 0)
S_error("fasl-read", "input truncated");
} else
vspaces[s] = (ptr)0;
}
for (s = vspaces_count - 1; s--; ) {
if (!vspaces[s])
vspaces[s] = vspaces[s+1];
}
data = (ptr)0; /* => initialize below */
to_static = 1;
} else {
thread_find_room(tc, typemod, header.data_size, data)
if (S_fasl_stream_read(stream, TO_VOIDP(data), header.data_size) < 0)
S_error("fasl-read", "input truncated");
}
thread_find_room(tc, typemod, ptr_align(header.table_size), table)
if (S_fasl_stream_read(stream, TO_VOIDP(table), header.table_size) < 0)
S_error("fasl-read", "input truncated");
}
if (data) {
for (s = 0; s < vspaces_count; s++)
vspaces[s] = ptr_add(data, vspace_offsets[s]);
} else
data = vspaces[0];
symrefs = TO_VOIDP(table);
rtdrefs = TO_VOIDP(ptr_add(TO_PTR(symrefs), header.symref_count * sizeof(vfoff)));
singletonrefs = TO_VOIDP(ptr_add(TO_PTR(rtdrefs), header.rtdref_count * sizeof(vfoff)));
bm = TO_VOIDP(ptr_add(TO_PTR(singletonrefs), header.singletonref_count * sizeof(vfoff)));
bm_end = TO_VOIDP(ptr_add(TO_PTR(table), header.table_size));
#if 0
printf("\n"
"hdr %ld\n"
"syms %ld\n"
"rtds %ld\n"
"clos %ld\n"
"code %ld\n"
"rloc %ld\n"
"data %ld\n"
"othr %ld\n"
"tabl %ld symref %ld rtdref %ld sglref %ld\n",
sizeof(vfasl_header),
VSPACE_LENGTH(vspace_symbol),
VSPACE_LENGTH(vspace_rtd),
VSPACE_LENGTH(vspace_closure),
VSPACE_LENGTH(vspace_code),
VSPACE_LENGTH(vspace_reloc),
VSPACE_LENGTH(vspace_data),
(VSPACE_LENGTH(vspace_impure)
+ VSPACE_LENGTH(vspace_pure_typed)
+ VSPACE_LENGTH(vspace_impure_record)),
header.table_size,
header.symref_count * sizeof(vfoff),
header.rtdref_count * sizeof(vfoff),
header.singletonref_count * sizeof(vfoff));
#endif
/* We have to convert an offset relative to the start of data in the
vfasl format to an offset relative to an individual space, at
least for target generations other than 0. Rely on the fact that
the spaces and the references are both sorted. */
#define SPACE_OFFSET_DECLS \
int s2 = 0; \
uptr offset2 = vspace_offsets[s2]; \
uptr next_offset2 = vspace_offsets[s2+1]
#define INC_SPACE_OFFSET(off) \
do { \
while ((off) >= next_offset2) { \
s2++; \
offset2 = next_offset2; \
next_offset2 = vspace_offsets[s2+1]; \
} \
} while (0)
#define SPACE_PTR(off) TO_VOIDP(ptr_add(vspaces[s2], (off) - offset2))
/* Fix up pointers. The initial content has all pointers relative to
the start of the data. Since the spaces of referenced pointers
may be discontiguous, use `find_pointer_from_offset` to get each
new pointer. */
{
SPACE_OFFSET_DECLS;
uptr p_off = 0;
while (bm != bm_end) {
octet m = *bm;
# define MAYBE_FIXUP(i) \
if (m & (1 << i)) { \
ptr *p3; \
INC_SPACE_OFFSET(p_off); \
p3 = SPACE_PTR(p_off); \
*p3 = find_pointer_from_offset((uptr)*p3, vspaces, vspace_offsets); \
} \
p_off += sizeof(uptr);
MAYBE_FIXUP(0);
MAYBE_FIXUP(1);
MAYBE_FIXUP(2);
MAYBE_FIXUP(3);
MAYBE_FIXUP(4);
MAYBE_FIXUP(5);
MAYBE_FIXUP(6);
MAYBE_FIXUP(7);
# undef MAYBE_FIXUP
bm++;
}
}
/* Replace references to singletons like "" and #vu8().
This needs to be before interning symbols, in case ""
is a symbol name. */
{
SPACE_OFFSET_DECLS;
vfoff i;
for (i = 0; i < header.singletonref_count; i++) {
uptr r_off;
ptr *ref;
r_off = singletonrefs[i];
INC_SPACE_OFFSET(r_off);
ref = SPACE_PTR(r_off);
*ref = lookup_singleton(UNFIX(*ref));
}
}
/* Intern symbols */
{
uptr in_seg_off = 0;
ptr sym = TYPE(vspaces[vspace_symbol], type_symbol);
ptr end_syms = TYPE(VSPACE_END(vspace_symbol), type_symbol);
if (sym != end_syms) {
tc_mutex_acquire()
while (sym < end_syms) {
ptr isym;
/* Make sure we don't try to claim a symbol that crosses
a segment boundary */
if ((in_seg_off + size_symbol) > bytes_per_segment) {
if (in_seg_off == bytes_per_segment) {
in_seg_off = 0;
} else {
/* Back up, then round up to next segment */
sym = ptr_add(ptr_subtract(sym, size_symbol),
(bytes_per_segment - (in_seg_off - size_symbol)));
in_seg_off = 0;
}
}
INITSYMVAL(sym) = sunbound;
INITSYMCODE(sym,S_G.nonprocedure_code);
isym = S_intern4(sym);
if (isym != sym) {
/* The symbol was already interned, so point to the existing one */
INITSYMVAL(sym) = isym;
if (S_vfasl_boot_mode > 0) {
IGEN gen = SegInfo(ptr_get_segment(isym))->generation;
if (gen < static_generation) {
printf("WARNING: vfasl symbol already interned, but at generation %d: %p ", gen, TO_VOIDP(isym));
S_prin1(isym);
printf("\n");
}
}
} else {
if (INITSYMPLIST(sym) != Snil) printf("oops\n");
if (INITSYMSPLIST(sym) != Snil) printf("oops\n");
}
sym = ptr_add(sym, size_symbol);
in_seg_off += size_symbol;
}
tc_mutex_release()
}
}
/* Replace symbol references with interned references */
{
SPACE_OFFSET_DECLS;
ptr syms = vspaces[vspace_symbol];
vfoff i;
for (i = 0; i < header.symref_count; i++) {
uptr p2_off, sym_pos;
ptr *p2, sym, val;
p2_off = symrefs[i];
INC_SPACE_OFFSET(p2_off);
p2 = SPACE_PTR(p2_off);
sym_pos = UNFIX(*p2);
sym = TYPE(ptr_add(syms, symbol_pos_to_offset(sym_pos)), type_symbol);
if ((val = SYMVAL(sym)) != sunbound)
sym = val;
*p2 = sym;
}
}
/* Intern rtds */
if (VSPACE_LENGTH(vspace_rtd) > 0) {
ptr rtd = TYPE(vspaces[vspace_rtd], type_typed_object);
ptr rtd_end = TYPE(VSPACE_END(vspace_rtd), type_typed_object);
/* first one corresponds to base_rtd */
RECORDINSTTYPE(rtd) = S_G.base_rtd;
RECORDDESCUID(rtd) = S_G.base_rtd;
while (1) {
ptr new_rtd, meta_rtd, parent_rtd;
rtd = ptr_add(rtd, size_record_inst(UNFIX(RECORDDESCSIZE(RECORDINSTTYPE(rtd)))));
if (rtd == rtd_end)
break;
/* fixup type of rtd (where the type is usually base_rtd) */
meta_rtd = RECORDINSTTYPE(rtd);
if (!Ssymbolp(RECORDDESCUID(meta_rtd)))
RECORDINSTTYPE(rtd) = RECORDDESCUID(meta_rtd);
/* fixup parent before continuing, relying on parents being earlier in `rtd`s */
parent_rtd = RECORDDESCPARENT(rtd);
if (parent_rtd != Sfalse) {
ptr parent_uid = RECORDDESCUID(parent_rtd);
if (!Ssymbolp(parent_uid))
RECORDDESCPARENT(rtd) = parent_uid;
}
new_rtd = rtd;
if (S_fasl_intern_rtd(&new_rtd)) {
if (new_rtd == rtd) {
S_error1("vfasl", "incompatible record type ~s", RECORDDESCNAME(rtd));
} else {
/* Use the UID field to record already-interned replacement: */
RECORDDESCUID(rtd) = new_rtd;
}
}
}
}
/* Replace rtd references to interned references */
{
SPACE_OFFSET_DECLS;
vfoff i;
for (i = 0; i < header.rtdref_count; i++) {
uptr r_off;
ptr *ref, rtd, uid;
r_off = rtdrefs[i];
INC_SPACE_OFFSET(r_off);
ref = SPACE_PTR(r_off);
rtd = *ref;
uid = RECORDDESCUID(rtd);
if (!Ssymbolp(uid)) {
/* uid is replacement interned rtd */
*ref = uid;
}
}
}
/* Fix code pointers on closures */
{
ptr cl = TYPE(vspaces[vspace_closure], type_closure);
ptr end_closures = TYPE(VSPACE_END(vspace_closure), type_closure);
uptr code_delta = (uptr)ptr_subtract(vspaces[vspace_code], vspace_offsets[vspace_code]);
while (cl != end_closures) {
ptr code = CLOSCODE(cl);
code = ptr_add(code, code_delta);
#if 0
printf("%p ", code);
S_prin1(CODENAME(code));
printf("\n");
#endif
SETCLOSCODE(cl,code);
cl = ptr_add(cl, size_closure(CLOSLEN(cl)));
}
}
/* Fix code via relocations */
{
ptr sym_base = vspaces[vspace_symbol];
ptr code = TYPE(vspaces[vspace_code], type_typed_object);
ptr code_end = TYPE(VSPACE_END(vspace_code), type_typed_object);
S_record_code_mod(tc, (uptr)vspaces[vspace_code], (uptr)code_end - (uptr)code);
while (code != code_end) {
relink_code(code, sym_base, vspaces, vspace_offsets, to_static);
code = ptr_add(code, size_code(CODELEN(code)));
}
}
/* Turn result offset into a value, unboxing if it's a box (which
supports a symbol result, for example). */
{
ptr v;
ITYPE t;
v = find_pointer_from_offset(header.result_offset, vspaces, vspace_offsets);
if (((t = TYPEBITS(v)) == type_typed_object)
&& TYPEP(TYPEFIELD(v), mask_box, type_box))
v = Sunbox(v);
return v;
}
}
ptr S_vfasl_to(ptr bv)
{
return S_vfasl(bv, NULL, 0, Sbytevector_length(bv));
}
/************************************************************/
/* Saving */
static void vfasl_init(vfasl_info *vfi) {
int s;
vfi->base_addr = (ptr)0;
vfi->sym_count = 0;
vfi->symref_count = 0;
vfi->symrefs = NULL;
vfi->base_rtd = S_G.base_rtd;
vfi->rtdref_count = 0;
vfi->rtdrefs = NULL;
vfi->singletonref_count = 0;
vfi->singletonrefs = NULL;
vfi->graph = make_vfasl_hash_table(0);
vfi->ptr_bitmap = NULL;
vfi->installs_library_entry = 0;
for (s = 0; s < vspaces_count; s++) {
vfasl_chunk *c;
c = vfasl_malloc(sizeof(vfasl_chunk));
c->bytes = (ptr)0;
c->length = 0;
c->used = 0;
c->swept = 0;
c->next = NULL;
c->prev = NULL;
vfi->spaces[s].first = c;
vfi->spaces[s].total_bytes = 0;
}
}
ptr S_to_vfasl(ptr v)
{
vfasl_info *vfi;
vfasl_header header;
ITYPE t;
int s;
uptr size, data_size, bitmap_size;
ptr bv, p;
fasl_init_entry_tables();
/* Box certain kinds of values where the vfasl process needs a
pointer into data */
if (IMMEDIATE(v)
|| detect_singleton(v)
|| ((t = TYPEBITS(v)) == type_symbol)
|| ((t == type_typed_object)
&& TYPEP(TYPEFIELD(v), mask_record, type_record)
&& (TYPEFIELD(v) == v))
|| ((t == type_typed_object)
&& TYPEP(TYPEFIELD(v), mask_box, type_box))) {
v = Sbox(v);
}
vfi = vfasl_malloc(sizeof(vfasl_info));
vfasl_init(vfi);
/* First pass: determine sizes */
(void)vfasl_copy_all(vfi, v);
/* Setup for second pass: allocate to contiguous bytes */
size = sizeof(vfasl_header);
data_size = vfi->spaces[0].total_bytes;
for (s = 1; s < vspaces_count; s++) {
header.vspace_rel_offsets[s-1] = data_size;
data_size += vfi->spaces[s].total_bytes;
}
header.data_size = data_size;
size += data_size;
size += vfi->symref_count * sizeof(vfoff);
size += vfi->rtdref_count * sizeof(vfoff);
size += vfi->singletonref_count * sizeof(vfoff);
header.symref_count = vfi->symref_count;
header.rtdref_count = vfi->rtdref_count;
header.singletonref_count = vfi->singletonref_count;
header.table_size = size - data_size - sizeof(header); /* doesn't yet include the bitmap */
bitmap_size = (data_size + (byte_bits-1)) >> log2_byte_bits;
size += bitmap_size;
bv = S_bytevector(size);
memset(&BVIT(bv, 0), 0, size);
p = TO_PTR(&BVIT(bv, 0));
/* Skip header for now */
p = ptr_add(p, sizeof(vfasl_header));
vfi->base_addr = p;
/* Set pointers to vspaces based on sizes from first pass */
for (s = 0; s < vspaces_count; s++) {
vfasl_chunk *c;
c = vfasl_malloc(sizeof(vfasl_chunk));
c->bytes = p;
c->length = vfi->spaces[s].total_bytes;
c->used = 0;
c->swept = 0;
c->next = NULL;
c->prev = NULL;
vfi->spaces[s].first = c;
p = ptr_add(p, vfi->spaces[s].total_bytes);
vfi->spaces[s].total_bytes = 0;
}
vfi->symrefs = TO_VOIDP(p);
p = ptr_add(p, sizeof(vfoff) * vfi->symref_count);
vfi->base_rtd = S_G.base_rtd;
vfi->rtdrefs = TO_VOIDP(p);
p = ptr_add(p, sizeof(vfoff) * vfi->rtdref_count);
vfi->singletonrefs = TO_VOIDP(p);
p = ptr_add(p, sizeof(vfoff) * vfi->singletonref_count);
vfi->sym_count = 0;
vfi->symref_count = 0;
vfi->rtdref_count = 0;
vfi->singletonref_count = 0;
vfi->graph = make_vfasl_hash_table(0);
vfi->ptr_bitmap = TO_VOIDP(p);
/* Write data */
v = vfasl_copy_all(vfi, v);
header.result_offset = ptr_diff(v, vfi->base_addr);
/* Make all pointers relative to the start of the data area */
{
ptr *p2 = TO_VOIDP(vfi->base_addr);
uptr base_addr = (uptr)vfi->base_addr;
octet *bm = vfi->ptr_bitmap;
octet *bm_end = bm + bitmap_size;
uptr zeros = 0;
for (; bm != bm_end; bm++, p2 += byte_bits) {
octet m = *bm;
if (m == 0) {
zeros++;
} else {
# define MAYBE_FIXUP(i) if (m & (1 << i)) ((uptr *)p2)[i] -= base_addr;
MAYBE_FIXUP(0);
MAYBE_FIXUP(1);
MAYBE_FIXUP(2);
MAYBE_FIXUP(3);
MAYBE_FIXUP(4);
MAYBE_FIXUP(5);
MAYBE_FIXUP(6);
MAYBE_FIXUP(7);
# undef MAYBE_FIXUP
zeros = 0;
}
}
/* We can ignore trailing zeros */
header.table_size += (bitmap_size - zeros);
}
/* Truncate bytevector to match end of bitmaps */
{
uptr sz = sizeof(vfasl_header) + header.data_size + header.table_size;
BYTEVECTOR_TYPE(bv) = (sz << bytevector_length_offset) | type_bytevector;
}
memcpy(&BVIT(bv, 0), &header, sizeof(vfasl_header));
sort_offsets(vfi->symrefs, vfi->symref_count);
sort_offsets(vfi->rtdrefs, vfi->rtdref_count);
sort_offsets(vfi->singletonrefs, vfi->singletonref_count);
return bv;
}
/* If compiled code uses `$install-library-entry`, then it can't be
combined into a single vfasled object, because the installation
needs to be evaluated for laster vfasls. Recognize a non-combinable
value as anything that references the C entry or even mentions the
symbol `$install-library-entry` (as defined in "library.ss"). If
non-boot code mentions the symbol `$install-library-entry`, it just
isn't as optimal.
This is an expensive test, since we perform half of a vfasl
encoding to look for `$install-library-entry`. */
IBOOL S_vfasl_can_combinep(ptr v)
{
IBOOL installs;
vfasl_info *vfi;
if (IMMEDIATE(v))
return 1;
fasl_init_entry_tables();
/* Run a "first pass" */
vfi = vfasl_malloc(sizeof(vfasl_info));
vfasl_init(vfi);
(void)vfasl_copy_all(vfi, v);
installs = vfi->installs_library_entry;
return !installs;
}
/************************************************************/
/* Traversals for saving */
static ptr vfasl_copy_all(vfasl_info *vfi, ptr v) {
seginfo *si;
int s;
int changed = 1;
si = MaybeSegInfo(ptr_get_segment(v));
v = copy(vfi, v, si);
while (changed) {
changed = 0;
for (s = 0; s < vspaces_count; s++) {
vfasl_chunk *c = vfi->spaces[s].first;
/* consistent order of sweeping by older chunks first: */
if (c) {
while ((c->swept < c->used) && c->next)
c = c->next;
if (c->swept >= c->used)
c = c->prev;
}
while (c) {
ptr pp, pp_end;
pp = ptr_add(c->bytes, c->swept);
pp_end = ptr_add(c->bytes, c->used);
c->swept = c->used;
switch(s) {
case vspace_symbol:
while (pp < pp_end) {
pp = ptr_add(pp, sweep(vfi, TYPE((ptr)pp, type_symbol)));
}
break;
case vspace_closure:
while (pp < pp_end) {
pp = ptr_add(pp, sweep(vfi, TYPE((ptr)pp, type_closure)));
}
break;
case vspace_impure:
while (pp < pp_end) {
vfasl_relocate(vfi, TO_VOIDP(pp));
pp = ptr_add(pp, sizeof(ptr));
}
break;
case vspace_rtd:
case vspace_code:
case vspace_pure_typed:
case vspace_impure_record:
while (pp < pp_end) {
pp = ptr_add(pp, sweep(vfi, TYPE((ptr)pp, type_typed_object)));
}
break;
case vspace_data:
case vspace_reloc:
break;
default:
S_error_abort("vfasl: unrecognized space");
break;
}
if (c->swept >= c->used)
c = c->prev;
changed = 1;
}
}
}
return v;
}
static void vfasl_register_pointer(vfasl_info *vfi, ptr *pp) {
if (vfi->ptr_bitmap) {
uptr delta = ptr_diff(TO_PTR(pp), vfi->base_addr) >> log2_ptr_bytes;
uptr i = delta >> log2_byte_bits;
uptr bit = (((uptr)1) << (delta & (byte_bits - 1)));
vfi->ptr_bitmap[i] |= bit;
}
}
static uptr ptr_base_diff(vfasl_info *vfi, ptr p) {
if ((uptr)vfi->base_addr > (uptr)UNTYPE(p, TYPEBITS(p)))
S_error_abort("vfasl: pointer not in region");
return ptr_diff(p, vfi->base_addr);
}
static void vfasl_register_symbol_reference(vfasl_info *vfi, ptr *pp, ptr p) {
if (vfi->symrefs)
vfi->symrefs[vfi->symref_count] = ptr_base_diff(vfi, TO_PTR(pp));
vfi->symref_count++;
*pp = SYMVAL(p); /* replace symbol reference with index of symbol */
}
static void vfasl_register_rtd_reference(vfasl_info *vfi, ptr pp) {
if (vfi->rtdrefs)
vfi->rtdrefs[vfi->rtdref_count] = ptr_base_diff(vfi, pp);
vfi->rtdref_count++;
}
static void vfasl_register_singleton_reference(vfasl_info *vfi, ptr *pp, int which) {
if (vfi->singletonrefs)
vfi->singletonrefs[vfi->singletonref_count] = ptr_base_diff(vfi, TO_PTR(pp));
vfi->singletonref_count++;
*pp = FIX(which);
}
static void vfasl_register_forward(vfasl_info *vfi, ptr pp, ptr p) {
vfasl_hash_table_set(vfi->graph, pp, p);
}
static ptr vfasl_lookup_forward(vfasl_info *vfi, ptr p) {
return vfasl_hash_table_ref(vfi->graph, p);
}
static void vfasl_relocate_parents(vfasl_info *vfi, ptr p) {
ptr ancestors = Snil;
while ((p != Sfalse) && !vfasl_lookup_forward(vfi, p)) {
ancestors = Scons(p, ancestors);
p = RECORDDESCPARENT(p);
}
while (ancestors != Snil) {
(void)vfasl_relocate_help(vfi, Scar(ancestors));
ancestors = Scdr(ancestors);
}
}
static ptr vfasl_find_room(vfasl_info *vfi, int s, ITYPE t, iptr n) {
ptr p;
uptr sz = vfi->spaces[s].total_bytes;
switch (s) {
case vspace_symbol:
case vspace_impure_record:
/* For these spaces, in case they will be loaded into the static
generation, objects must satisfy an extra constraint: an object
must not span segments unless it's at the start of a
segment. */
if (sz & (bytes_per_segment-1)) {
/* Since we're not at the start of a segment, don't let an
object span a segment */
if ((segment_align(sz) != segment_align(sz+n))
&& ((sz+n) != segment_align(sz+n))) {
/* Fill in to next segment, instead. */
uptr delta = segment_align(sz) - sz;
vfasl_chunk *c, *new_c;
vfi->spaces[s].total_bytes += delta;
/* Mark the end of the old segment */
c = vfi->spaces[s].first;
p = ptr_add(c->bytes, c->used);
FWDMARKER(p) = forward_marker;
/* Create a new chunk so the old one tracks the current
swept-to-used region, and the new chunk starts a new
segment. If the old chunk doesn't have leftover bytes
(because we're in the first pass), then we'll need to
clean out this useless chunk below. */
new_c = vfasl_malloc(sizeof(vfasl_chunk));
new_c->bytes = ptr_add(c->bytes, c->used + delta);
new_c->length = c->length - (c->used + delta);
new_c->used = 0;
new_c->swept = 0;
new_c->prev = NULL;
new_c->next = c;
c->prev = new_c;
vfi->spaces[s].first = new_c;
}
}
break;
default:
break;
}
vfi->spaces[s].total_bytes += n;
if (vfi->spaces[s].first->used + n > vfi->spaces[s].first->length) {
vfasl_chunk *c, *old_c;
iptr newlen = segment_align(n);
c = vfasl_malloc(sizeof(vfasl_chunk));
c->bytes = TO_PTR(vfasl_malloc(newlen));
c->length = newlen;
c->used = 0;
c->swept = 0;
old_c = vfi->spaces[s].first;
if (old_c->next && !old_c->length)
old_c = old_c->next; /* drop useless chunk created above */
c->prev = NULL;
c->next = old_c;
old_c->prev = c;
vfi->spaces[s].first = c;
}
p = ptr_add(vfi->spaces[s].first->bytes, vfi->spaces[s].first->used);
vfi->spaces[s].first->used += n;
return TYPE(p, t);
}
#define FIND_ROOM(vfi, s, t, n, p) p = vfasl_find_room(vfi, s, t, n)
#include "vfasl.inc"
static ptr vfasl_relocate_help(vfasl_info *vfi, ptr pp) {
ptr fpp;
seginfo *si;
si = MaybeSegInfo(ptr_get_segment(pp));
if (!si)
vfasl_fail(vfi, "unknown");
fpp = vfasl_lookup_forward(vfi, pp);
if (fpp)
return fpp;
else
return copy(vfi, pp, si);
}
/* Use vfasl_relocate only on addresses that are in the vfasl target area */
static void vfasl_relocate(vfasl_info *vfi, ptr *ppp) {
ptr pp = *ppp, tf;
if (!IMMEDIATE(pp)) {
int which_singleton;
if ((which_singleton = detect_singleton(pp)))
vfasl_register_singleton_reference(vfi, ppp, which_singleton);
else {
pp = vfasl_relocate_help(vfi, pp);
*ppp = pp;
if (!IMMEDIATE(pp)) {
if (TYPEBITS(pp) == type_symbol)
vfasl_register_symbol_reference(vfi, ppp, pp);
else {
if ((TYPEBITS(pp) == type_typed_object)
&& TYPEP((tf = TYPEFIELD(pp)), mask_record, type_record)
&& is_rtd(tf, vfi))
vfasl_register_rtd_reference(vfi, TO_PTR(ppp));
vfasl_register_pointer(vfi, ppp);
}
}
}
}
}
static ptr vfasl_relocate_code(vfasl_info *vfi, ptr code) {
/* We don't want to register `code` as a pointer, since it is
treated more directly */
return vfasl_relocate_help(vfi, code);
}
static int is_rtd(ptr tf, vfasl_info *vfi)
{
while (1) {
if (tf == vfi->base_rtd)
return 1;
if (tf == S_G.base_rtd)
return 1;
tf = RECORDDESCPARENT(tf);
if (tf == Sfalse)
return 0;
}
}
/*************************************************************/
/* Code and relocation handling for save and load */
#define VFASL_RELOC_TAG_BITS 3
#define VFASL_RELOC_C_ENTRY_TAG 1
#define VFASL_RELOC_LIBRARY_ENTRY_TAG 2
#define VFASL_RELOC_LIBRARY_ENTRY_CODE_TAG 3
#define VFASL_RELOC_SYMBOL_TAG 4
#define VFASL_RELOC_SINGLETON_TAG 5
/* FXIME: rtds? */
#define VFASL_RELOC_C_ENTRY(p) (((uptr)(p) << VFASL_RELOC_TAG_BITS) | VFASL_RELOC_C_ENTRY_TAG)
#define VFASL_RELOC_LIBRARY_ENTRY(p) (((uptr)(p) << VFASL_RELOC_TAG_BITS) | VFASL_RELOC_LIBRARY_ENTRY_TAG)
#define VFASL_RELOC_LIBRARY_ENTRY_CODE(p) (((uptr)(p) << VFASL_RELOC_TAG_BITS) | VFASL_RELOC_LIBRARY_ENTRY_CODE_TAG)
#define VFASL_RELOC_SYMBOL(p) (((uptr)(p) << VFASL_RELOC_TAG_BITS) | VFASL_RELOC_SYMBOL_TAG)
#define VFASL_RELOC_SINGLETON(p) (((uptr)(p) << VFASL_RELOC_TAG_BITS) | VFASL_RELOC_SINGLETON_TAG)
#define VFASL_RELOC_TAG(p) (UNFIX(p) & ((1 << VFASL_RELOC_TAG_BITS) - 1))
#define VFASL_RELOC_POS(p) (UNFIX(p) >> VFASL_RELOC_TAG_BITS)
/* Picks a relocation variant that fits into the actual relocation's
shape, but holds an absolue value */
static IFASLCODE abs_reloc_variant(IFASLCODE type) {
if (type == reloc_abs)
return reloc_abs;
#if defined(I386) || defined(X86_64)
return reloc_abs;
#elif defined(ARMV6)
return reloc_arm32_abs;
#elif defined(AARCH64)
return reloc_arm64_abs;
#elif defined(PPC32)
if (type == reloc_ppc32_abs)
return reloc_ppc32_abs;
else
return reloc_abs;
#elif defined(PORTABLE_BYTECODE)
return reloc_pb_abs;
#else
>> need to fill in for this platform <<
#endif
}
static ptr vfasl_encode_relocation(vfasl_info *vfi, ptr obj) {
ptr pos;
int which_singleton;
if ((which_singleton = detect_singleton(obj))) {
obj = FIX(VFASL_RELOC_SINGLETON(which_singleton));
} else if ((pos = vfasl_hash_table_ref(S_G.c_entries, obj))) {
pos = (ptr)((uptr)pos - 1);
if ((uptr)pos == CENTRY_install_library_entry)
vfi->installs_library_entry = 1;
obj = FIX(VFASL_RELOC_C_ENTRY(pos));
} else if ((pos = vfasl_hash_table_ref(S_G.library_entries, obj))) {
pos = (ptr)((uptr)pos - 1);
obj = FIX(VFASL_RELOC_LIBRARY_ENTRY(pos));
} else if ((pos = vfasl_hash_table_ref(S_G.library_entry_codes, obj))) {
pos = (ptr)((uptr)pos - 1);
obj = FIX(VFASL_RELOC_LIBRARY_ENTRY_CODE(pos));
} else if (Ssymbolp(obj)) {
obj = vfasl_relocate_help(vfi, obj);
obj = FIX(VFASL_RELOC_SYMBOL(UNFIX(SYMVAL(obj))));
} else if (IMMEDIATE(obj)) {
/* as-is */
if (Sfixnump(obj))
if (obj != FIX(0)) /* allow 0 for fcallable cookie */
S_error("vfasl", "unexpected fixnum in relocation");
} else {
obj = vfasl_relocate_help(vfi, obj);
obj = (ptr)ptr_diff(obj, vfi->base_addr);
}
return obj;
}
static void relink_code(ptr co, ptr sym_base, ptr *vspaces, uptr *vspace_offsets, IBOOL to_static) {
ptr t; iptr a, m, n;
t = CODERELOC(co);
t = ptr_add(vspaces[vspace_reloc], (uptr)t - vspace_offsets[vspace_reloc]);
if (to_static && !S_G.retain_static_relocation
&& ((CODETYPE(co) & (code_flag_template << code_flags_offset)) == 0))
CODERELOC(co) = (ptr)0;
else {
CODERELOC(co) = t;
RELOCCODE(t) = co;
}
m = RELOCSIZE(t);
a = 0;
n = 0;
while (n < m) {
uptr entry, item_off, code_off; ptr obj;
entry = RELOCIT(t, n); n += 1;
if (RELOC_EXTENDED_FORMAT(entry)) {
item_off = RELOCIT(t, n); n += 1;
code_off = RELOCIT(t, n); n += 1;
} else {
item_off = RELOC_ITEM_OFFSET(entry);
code_off = RELOC_CODE_OFFSET(entry);
}
a += code_off;
obj = S_get_code_obj(abs_reloc_variant(RELOC_TYPE(entry)), co, a, item_off);
if (IMMEDIATE(obj)) {
if (Sfixnump(obj)) {
int tag = VFASL_RELOC_TAG(obj);
int pos = VFASL_RELOC_POS(obj);
if (tag == VFASL_RELOC_SINGLETON_TAG)
obj = lookup_singleton(pos);
else if (tag == VFASL_RELOC_C_ENTRY_TAG)
obj = S_lookup_c_entry(pos);
else if ((tag == VFASL_RELOC_LIBRARY_ENTRY_TAG)
|| (tag == VFASL_RELOC_LIBRARY_ENTRY_CODE_TAG)) {
obj = S_lookup_library_entry(pos, 1);
if (tag == VFASL_RELOC_LIBRARY_ENTRY_CODE_TAG)
obj = CLOSCODE(obj);
} else if (tag == VFASL_RELOC_SYMBOL_TAG) {
ptr val;
obj = TYPE(ptr_add(sym_base, symbol_pos_to_offset(pos)), type_symbol);
if ((val = SYMVAL(obj)) != sunbound)
obj = val;
} else if (obj == FIX(0)) {
/* leave as-is */
} else {
S_error_abort("vfasl: bad relocation tag");
}
} else {
/* some other immediate, such as black-hole; leave as-is */
}
} else {
uptr offset = (uptr)obj;
obj = find_pointer_from_offset(offset, vspaces, vspace_offsets);
if (TYPEBITS(obj) == type_typed_object) {
ptr tf = TYPEFIELD(obj);
if (TYPEP(tf, mask_record, type_record)) {
while (1) {
if (tf == S_G.base_rtd) {
/* Similar to symbols: potentially replace with interned */
ptr uid = RECORDDESCUID(obj);
if (!Ssymbolp(uid)) {
/* "uid" is actually the interned rtd to use instead */
obj = uid;
}
break;
}
tf = RECORDDESCPARENT(tf);
if (tf == Sfalse)
break;
}
}
}
}
S_set_code_obj("vfasl", RELOC_TYPE(entry), co, a, obj, item_off);
}
}
static ptr find_pointer_from_offset(uptr p_off, ptr *vspaces, uptr *vspace_offsets)
{
int s = 0;
ITYPE t = TYPEBITS(p_off);
p_off = (uptr)UNTYPE(p_off, t);
while (p_off >= vspace_offsets[s+1])
s++;
return TYPE(ptr_add(vspaces[s], p_off - vspace_offsets[s]), t);
}
/*************************************************************/
/* Symbol names */
static iptr vfasl_symbol_to_index(vfasl_info *vfi, ptr pp)
{
uptr pos = vfi->sym_count++;
ptr name = SYMNAME(pp);
if (Sstringp(name))
vfasl_check_install_library_entry(vfi, name);
else if (!Spairp(name) || (Scar(name) == Sfalse))
vfasl_fail(vfi, "gensym without unique name");
return pos;
}
/*************************************************************/
/* C and library entries */
static void fasl_init_entry_tables()
{
tc_mutex_acquire()
if (!S_G.c_entries) {
iptr i;
S_G.c_entries = make_vfasl_hash_table(1);
S_G.library_entries = make_vfasl_hash_table(1);
S_G.library_entry_codes = make_vfasl_hash_table(1);
for (i = Svector_length(S_G.c_entry_vector); i--; ) {
ptr entry = Svector_ref(S_G.c_entry_vector, i);
vfasl_hash_table_set(S_G.c_entries, entry, (ptr)(i+1));
}
for (i = Svector_length(S_G.library_entry_vector); i--; ) {
ptr entry = Svector_ref(S_G.library_entry_vector, i);
if (entry != Sfalse) {
vfasl_hash_table_set(S_G.library_entries, entry, (ptr)(i+1));
if (Sprocedurep(entry))
vfasl_hash_table_set(S_G.library_entry_codes, CLOSCODE(entry), (ptr)(i+1));
}
}
}
tc_mutex_release()
}
static void vfasl_check_install_library_entry(vfasl_info *vfi, ptr name)
{
const char *ile = "$install-library-entry";
iptr len = Sstring_length(name), i;
for (i = 0; i < len; i++) {
if (Sstring_ref(name, i) != (unsigned)ile[i])
return;
}
if (!ile[i])
vfi->installs_library_entry = 1;
}
/*************************************************************/
/* Singletons, such as "" */
static ptr *singleton_refs[] = { &S_G.null_string,
&S_G.null_vector,
&S_G.null_fxvector,
&S_G.null_bytevector,
&S_G.null_immutable_string,
&S_G.null_immutable_vector,
&S_G.null_immutable_fxvector,
&S_G.null_immutable_bytevector,
&S_G.eqp,
&S_G.eqvp,
&S_G.equalp,
&S_G.symboleqp };
static int detect_singleton(ptr p) {
unsigned i;
for (i = 0; i < sizeof(singleton_refs) / sizeof(ptr*); i++) {
if (p == *(singleton_refs[i]))
return i+1;
}
return 0;
}
static ptr lookup_singleton(int which) {
return *(singleton_refs[which-1]);
}
/*************************************************************/
/* `eq?`-based hash table during saving as critical section */
typedef struct hash_entry {
ptr key, value;
} hash_entry;
struct vfasl_hash_table {
IBOOL permanent;
uptr count;
uptr size;
hash_entry *entries;
};
#define HASH_CODE(p) ((uptr)(p) >> log2_ptr_bytes)
#define HASH_CODE2(p) (((uptr)(p) >> (log2_ptr_bytes + log2_ptr_bytes)) | 1)
static vfasl_hash_table *make_vfasl_hash_table(IBOOL permanent) {
vfasl_hash_table *ht;
if (permanent)
ht = malloc(sizeof(vfasl_hash_table));
else
ht = vfasl_malloc(sizeof(vfasl_hash_table));
ht->permanent = permanent;
ht->count = 0;
ht->size = 16;
if (permanent)
ht->entries = calloc(sizeof(hash_entry), ht->size);
else
ht->entries = vfasl_calloc(sizeof(hash_entry), ht->size);
return ht;
}
static void vfasl_hash_table_set(vfasl_hash_table *ht, ptr key, ptr value) {
uptr hc = HASH_CODE(key);
uptr hc2 = HASH_CODE2(key);
uptr size = ht->size;
if (ht->count > ht->size >> 1) {
/* rehash */
uptr i;
hash_entry *old_entries = ht->entries;
ht->count = 0;
ht->size *= 2;
if (ht->permanent)
ht->entries = calloc(sizeof(hash_entry), ht->size);
else
ht->entries = vfasl_calloc(sizeof(hash_entry), ht->size);
for (i = 0; i < size; i++) {
if (old_entries[i].key)
vfasl_hash_table_set(ht, old_entries[i].key, old_entries[i].value);
}
if (ht->permanent)
free(old_entries);
size = ht->size;
}
hc = hc & (size - 1);
while (ht->entries[hc].key) {
hc = (hc + hc2) & (size - 1);
}
ht->entries[hc].key = key;
ht->entries[hc].value = value;
ht->count++;
}
static ptr vfasl_hash_table_ref(vfasl_hash_table *ht, ptr key) {
uptr hc = HASH_CODE(key);
uptr hc2 = HASH_CODE2(key);
uptr size = ht->size;
ptr old_key;
hc = hc & (size - 1);
while ((old_key = ht->entries[hc].key) != key) {
if (!old_key)
return (ptr)0;
hc = (hc + hc2) & (size - 1);
}
return ht->entries[hc].value;
}
/*************************************************************/
static void *vfasl_malloc(uptr sz) {
ptr tc = get_thread_context();
void *p;
thread_find_room_voidp(tc, ptr_align(sz), p)
return p;
}
static void *vfasl_calloc(uptr sz, uptr n) {
void *p;
sz *= n;
p = vfasl_malloc(sz);
memset(p, 0, sz);
return p;
}
/*************************************************************/
static void sort_offsets(vfoff *p, vfoff len)
{
while (1) {
if (len > 1) {
vfoff i, pivot = 0;
{
vfoff mid = len >> 2;
vfoff tmp = p[mid];
p[mid] = p[0];
p[0] = tmp;
}
for (i = 1; i < len; i++) {
if (p[i] < p[pivot]) {
vfoff tmp = p[pivot];
p[pivot] = p[i];
pivot++;
p[i] = p[pivot];
p[pivot] = tmp;
}
}
if (pivot > (len >> 1)) {
sort_offsets(p+pivot+1, len-pivot-1);
len = pivot;
} else {
sort_offsets(p, pivot);
p = p+pivot+1;
len = len-pivot-1;
}
} else
return;
}
}