c-capnproto/lib/capn.c
2023-08-07 17:26:44 -07:00

1147 lines
24 KiB
C

/* vim: set sw=8 ts=8 sts=8 noet: */
/* capn.c
*
* Copyright (C) 2013 James McKaskill
*
* This software may be modified and distributed under the terms
* of the MIT license. See the LICENSE file for details.
*/
#ifdef __GNUC__
#pragma GCC diagnostic ignored "-Wmissing-field-initializers"
#endif
#include "capnp_c.h"
#include <stdlib.h>
#include <string.h>
#ifndef _MSC_VER
#include <sys/param.h>
#endif
#define STRUCT_PTR 0
#define LIST_PTR 1
#define FAR_PTR 2
#define DOUBLE_PTR 6
#define VOID_LIST 0
#define BIT_1_LIST 1
#define BYTE_1_LIST 2
#define BYTE_2_LIST 3
#define BYTE_4_LIST 4
#define BYTE_8_LIST 5
#define PTR_LIST 6
#define COMPOSITE_LIST 7
#define U64(val) ((uint64_t) (val))
#define I64(val) ((int64_t) (val))
#define U32(val) ((uint32_t) (val))
#define I32(val) ((int32_t) (val))
#define U16(val) ((uint16_t) (val))
#define I16(val) ((int16_t) (val))
#ifndef min
static int min(int a, int b) { return (a < b) ? a : b; }
#endif
#ifdef BYTE_ORDER
#define CAPN_LITTLE (BYTE_ORDER == LITTLE_ENDIAN)
#elif defined(__BYTE_ORDER)
#define CAPN_LITTLE (__BYTE_ORDER == __LITTLE_ENDIAN)
#else
#define CAPN_LITTLE 0
#endif
struct capn_tree *capn_tree_insert(struct capn_tree *root, struct capn_tree *n) {
n->red = 1;
n->link[0] = n->link[1] = NULL;
for (;;) {
/* parent, uncle, grandparent, great grandparent link */
struct capn_tree *p, *u, *g, **gglink;
int dir;
/* Case 1: N is root */
p = n->parent;
if (!p) {
n->red = 0;
root = n;
break;
}
/* Case 2: p is black */
if (!p->red) {
break;
}
g = p->parent;
dir = (p == g->link[1]);
/* Case 3: P and U are red, switch g to red, but must
* loop as G could be root or have a red parent
* g to G
* / \ / \
* P U p u
* / /
* N N
*/
u = g->link[!dir];
if (u != NULL && u->red) {
p->red = 0;
u->red = 0;
g->red = 1;
n = g;
continue;
}
if (!g->parent) {
gglink = &root;
} else if (g->parent->link[1] == g) {
gglink = &g->parent->link[1];
} else {
gglink = &g->parent->link[0];
}
if (dir != (n == p->link[1])) {
/* Case 4: rotate on P, then on g
* here dir is /
* g to g to n
* / \ / \ / \
* P u N u P G
* / \ / \ /| / \
* 1 N P 3 1 2 3 u
* / \ / \
* 2 3 1 2
*/
struct capn_tree *two = n->link[dir];
struct capn_tree *three = n->link[!dir];
p->link[!dir] = two;
g->link[dir] = three;
n->link[dir] = p;
n->link[!dir] = g;
*gglink = n;
n->parent = g->parent;
p->parent = n;
g->parent = n;
if (two)
two->parent = p;
if (three)
three->parent = g;
n->red = 0;
g->red = 1;
} else {
/* Case 5: rotate on g
* here dir is /
* g to p
* / \ / \
* P u N G
* / \ /| / \
* N 3 1 2 3 u
* / \
* 1 2
*/
struct capn_tree *three = p->link[!dir];
g->link[dir] = three;
p->link[!dir] = g;
*gglink = p;
p->parent = g->parent;
g->parent = p;
if (three)
three->parent = g;
g->red = 1;
p->red = 0;
}
break;
}
return root;
}
void capn_append_segment(struct capn *c, struct capn_segment *s) {
s->id = c->segnum++;
s->capn = c;
s->next = NULL;
if (c->lastseg) {
c->lastseg->next = s;
c->lastseg->hdr.link[1] = &s->hdr;
s->hdr.parent = &c->lastseg->hdr;
} else {
c->seglist = s;
s->hdr.parent = NULL;
}
c->lastseg = s;
c->segtree = capn_tree_insert(c->segtree, &s->hdr);
}
static char *new_data(struct capn *c, int sz, struct capn_segment **ps) {
struct capn_segment *s;
/* find a segment with sufficient data */
for (s = c->seglist; s != NULL; s = s->next) {
if (s->len + sz <= s->cap) {
goto end;
}
}
s = c->create ? c->create(c->user, c->segnum, sz) : NULL;
if (!s) {
*ps = NULL;
return NULL;
}
capn_append_segment(c, s);
end:
*ps = s;
s->len += sz;
return s->data + s->len - sz;
}
static struct capn_segment *lookup_segment(struct capn* c, struct capn_segment *s, uint32_t id) {
struct capn_tree **x;
struct capn_segment *y = NULL;
if (s && s->id == id)
return s;
if (!c)
return NULL;
if (id < c->segnum) {
x = &c->segtree;
while (*x) {
y = (struct capn_segment*) *x;
if (id == y->id) {
return y;
} else if (id < y->id) {
x = &y->hdr.link[0];
} else {
x = &y->hdr.link[1];
}
}
} else {
/* Otherwise `x` may be uninitialized */
return NULL;
}
s = c->lookup ? c->lookup(c->user, id) : NULL;
if (!s)
return NULL;
if (id < c->segnum) {
s->id = id;
s->capn = c;
s->next = c->seglist;
c->seglist = s;
s->hdr.parent = &y->hdr;
*x = &s->hdr;
c->segtree = capn_tree_insert(c->segtree, &s->hdr);
} else {
c->segnum = id;
capn_append_segment(c, s);
}
return s;
}
static uint64_t lookup_double(struct capn_segment **s, char **d, uint64_t val) {
uint64_t far, tag;
size_t off = (U32(val) >> 3) * 8;
char *p;
if ((*s = lookup_segment((*s)->capn, *s, U32(val >> 32))) == NULL) {
return 0;
}
p = (*s)->data + off;
if (off + 16 > (*s)->len) {
return 0;
}
far = capn_flip64(*(uint64_t*) p);
tag = capn_flip64(*(uint64_t*) (p+8));
/* the far tag should not be another double, and the tag
* should be struct/list and have no offset */
if ((far&7) != FAR_PTR || U32(tag) > LIST_PTR) {
return 0;
}
if ((*s = lookup_segment((*s)->capn, *s, U32(far >> 32))) == NULL) {
return 0;
}
/* -8 because far pointers reference from the start of
* the segment, but offsets reference the end of the
* pointer data. Here *d points to where an equivalent
* ptr would be.
*/
*d = (*s)->data - 8;
return U64(U32(far) >> 3 << 2) | tag;
}
static uint64_t lookup_far(struct capn_segment **s, char **d, uint64_t val) {
size_t off = (U32(val) >> 3) * 8;
if ((*s = lookup_segment((*s)->capn, *s, U32(val >> 32))) == NULL) {
return 0;
}
if (off + 8 > (*s)->len) {
return 0;
}
*d = (*s)->data + off;
return capn_flip64(*(uint64_t*)*d);
}
static char *struct_ptr(struct capn_segment *s, char *d, int minsz) {
uint64_t val = capn_flip64(*(uint64_t*)d);
uint16_t datasz;
switch (val&7) {
case FAR_PTR:
val = lookup_far(&s, &d, val);
break;
case DOUBLE_PTR:
val = lookup_double(&s, &d, val);
break;
}
datasz = U16(val >> 32);
d += (I32(U32(val)) << 1) + 8;
if (val != 0 && (val&3) != STRUCT_PTR && datasz >= minsz && s->data <= d && d < s->data + s->len) {
return d;
}
return NULL;
}
static capn_ptr read_ptr(struct capn_segment *s, char *d) {
capn_ptr ret = {CAPN_NULL};
uint64_t val;
char *e = 0;
val = capn_flip64(*(uint64_t*) d);
switch (val&7) {
case FAR_PTR:
val = lookup_far(&s, &d, val);
ret.has_ptr_tag = (U32(val) >> 2) == 0;
break;
case DOUBLE_PTR:
val = lookup_double(&s, &d, val);
break;
}
d += (I32(U32(val)) >> 2) * 8 + 8;
if (d < s->data) {
goto err;
}
switch (val & 3) {
case STRUCT_PTR:
ret.type = val ? CAPN_STRUCT : CAPN_NULL;
goto struct_common;
struct_common:
ret.datasz = U32(U16(val >> 32)) * 8;
ret.ptrs = U32(U16(val >> 48));
e = d + ret.datasz + 8 * ret.ptrs;
break;
case LIST_PTR:
ret.type = CAPN_LIST;
ret.len = val >> 35;
switch ((val >> 32) & 7) {
case VOID_LIST:
e = d;
break;
case BIT_1_LIST:
ret.type = CAPN_BIT_LIST;
ret.datasz = (ret.len+7)/8;
e = d + ret.datasz;
break;
case BYTE_1_LIST:
ret.datasz = 1;
e = d + ret.len;
break;
case BYTE_2_LIST:
ret.datasz = 2;
e = d + ret.len * 2;
break;
case BYTE_4_LIST:
ret.datasz = 4;
e = d + ret.len * 4;
break;
case BYTE_8_LIST:
ret.datasz = 8;
e = d + ret.len * 8;
break;
case PTR_LIST:
ret.type = CAPN_PTR_LIST;
e = d + ret.len * 8;
break;
case COMPOSITE_LIST:
if ((size_t)((d+8) - s->data) > s->len) {
goto err;
}
val = capn_flip64(*(uint64_t*) d);
d += 8;
e = d + ret.len * 8;
ret.datasz = U32(U16(val >> 32)) * 8;
ret.ptrs = U32(U16(val >> 48));
ret.len = U32(val) >> 2;
ret.is_composite_list = 1;
if ((ret.datasz + 8*ret.ptrs) * ret.len != e - d) {
goto err;
}
break;
}
break;
default:
goto err;
}
if ((size_t)(e - s->data) > s->len)
goto err;
ret.data = d;
ret.seg = s;
return ret;
err:
memset(&ret, 0, sizeof(ret));
return ret;
}
void capn_resolve(capn_ptr *p) {
if (p->type == CAPN_FAR_POINTER) {
*p = read_ptr(p->seg, p->data);
}
}
/* TODO: should this handle CAPN_BIT_LIST? */
capn_ptr capn_getp(capn_ptr p, int off, int resolve) {
capn_ptr ret = {CAPN_FAR_POINTER};
ret.seg = p.seg;
capn_resolve(&p);
switch (p.type) {
case CAPN_LIST:
/* Return an inner pointer */
if (off < p.len) {
capn_ptr ret = {CAPN_STRUCT};
ret.is_list_member = 1;
ret.data = p.data + off * (p.datasz + 8*p.ptrs);
ret.seg = p.seg;
ret.datasz = p.datasz;
ret.ptrs = p.ptrs;
return ret;
} else {
goto err;
}
case CAPN_STRUCT:
if (off >= p.ptrs) {
goto err;
}
ret.data = p.data + p.datasz + 8*off;
break;
case CAPN_PTR_LIST:
if (off >= p.len) {
goto err;
}
ret.data = p.data + 8*off;
break;
default:
goto err;
}
if (resolve) {
ret = read_ptr(ret.seg, ret.data);
}
return ret;
err:
memset(&p, 0, sizeof(p));
return p;
}
static void write_ptr_tag(char *d, capn_ptr p, int off) {
/*
lsb struct pointer msb
+-+-----------------------------+---------------+---------------+
|A| B | C | D |
+-+-----------------------------+---------------+---------------+
A (2 bits) = 0, to indicate that this is a struct pointer.
B (30 bits) = Offset, in words, from the end of the pointer to the
start of the struct's data section. Signed.
C (16 bits) = Size of the struct's data section, in words.
D (16 bits) = Size of the struct's pointer section, in words.
For B we can't simply left-shift by 2 bits since C11 6.5.7.4
https://www.open-std.org/jtc1/sc22/wg14/www/docs/n1570.pdf
says we can get undefined behavior when the left-shift exceeds
the signed integer (ie. values run into the sign bit). The
ASAN detector will rightly complain. So we do two's complement
manually, and check bounds, to stay within unsigned arithmetic.
*/
const int off_words = off / 8;
uint64_t val;
if (off_words < 0) {
if (off_words < -(2147483647 >> 2) - 1) {
goto err;
}
uint32_t twos = 1 + ~(U32(-off_words) << 2);
val = U64(twos);
} else {
val = U64(U32(off_words) << 2);
}
switch (p.type) {
case CAPN_STRUCT:
val |= STRUCT_PTR | (U64(p.datasz/8) << 32) | (U64(p.ptrs) << 48);
break;
case CAPN_LIST:
if (p.is_composite_list) {
val |= LIST_PTR | (U64(COMPOSITE_LIST) << 32) | (U64(p.len * (p.datasz/8 + p.ptrs)) << 35);
} else {
val |= LIST_PTR | (U64(p.len) << 35);
switch (p.datasz) {
case 8:
val |= (U64(BYTE_8_LIST) << 32);
break;
case 4:
val |= (U64(BYTE_4_LIST) << 32);
break;
case 2:
val |= (U64(BYTE_2_LIST) << 32);
break;
case 1:
val |= (U64(BYTE_1_LIST) << 32);
break;
case 0:
val |= (U64(VOID_LIST) << 32);
break;
}
}
break;
case CAPN_BIT_LIST:
val |= LIST_PTR | (U64(BIT_1_LIST) << 32) | (U64(p.len) << 35);
break;
case CAPN_PTR_LIST:
val |= LIST_PTR | (U64(PTR_LIST) << 32) | (U64(p.len) << 35);
break;
default:
val = 0;
break;
}
*(uint64_t*) d = capn_flip64(val);
err:
memset(&p, 0, sizeof(p));
}
static void write_far_ptr(char *d, struct capn_segment *s, char *tgt) {
*(uint64_t*) d = capn_flip64(FAR_PTR | U64(tgt - s->data) | (U64(s->id) << 32));
}
static void write_double_far(char *d, struct capn_segment *s, char *tgt) {
*(uint64_t*) d = capn_flip64(DOUBLE_PTR | U64(tgt - s->data) | (U64(s->id) << 32));
}
#define NEED_TO_COPY 1
static int write_ptr(struct capn_segment *s, char *d, capn_ptr p) {
/* note p.seg can be NULL if its a ptr to static data */
char *pdata = p.data - 8*p.is_composite_list;
if (p.type == CAPN_NULL || (p.type == CAPN_STRUCT && p.datasz == 0 && p.ptrs == 0)) {
write_ptr_tag(d, p, 0);
return 0;
} else if (!p.seg || p.seg->capn != s->capn || p.is_list_member) {
return NEED_TO_COPY;
} else if (p.seg == s) {
write_ptr_tag(d, p, pdata - d - 8);
return 0;
} else if (p.has_ptr_tag) {
/* By lucky chance, the data has a tag in front
* of it. This happens when new_object had to move
* the data to a new segment. */
write_far_ptr(d, p.seg, pdata-8);
return 0;
} else if (p.seg->len + 8 <= p.seg->cap) {
/* The target segment has enough room for tag */
char *t = p.seg->data + p.seg->len;
write_ptr_tag(t, p, pdata - t - 8);
write_far_ptr(d, p.seg, t);
p.seg->len += 8;
return 0;
} else {
/* have to allocate room for a double far
* pointer */
char *t;
if (s->len + 16 <= s->cap) {
/* Try and allocate in the src segment
* first. This should improve lookup on
* read. */
t = s->data + s->len;
s->len += 16;
} else {
t = new_data(s->capn, 16, &s);
if (!t) return -1;
}
write_far_ptr(t, p.seg, pdata);
write_ptr_tag(t+8, p, 0);
write_double_far(d, s, t);
return 0;
}
}
struct copy {
struct capn_tree hdr;
struct capn_ptr to, from;
char *fbegin, *fend;
};
static capn_ptr new_clone(struct capn_segment *s, capn_ptr p) {
switch (p.type) {
case CAPN_STRUCT:
return capn_new_struct(s, p.datasz, p.ptrs);
case CAPN_PTR_LIST:
return capn_new_ptr_list(s, p.len);
case CAPN_BIT_LIST:
return capn_new_list1(s, p.len).p;
case CAPN_LIST:
return capn_new_list(s, p.len, p.datasz, p.ptrs);
default:
return p;
}
}
static int is_ptr_equal(const struct capn_ptr *a, const struct capn_ptr *b) {
return a->data == b->data
&& a->type == b->type
&& a->len == b->len
&& a->datasz == b->datasz
&& a->ptrs == b->ptrs;
}
static int data_size(struct capn_ptr p) {
switch (p.type) {
case CAPN_BIT_LIST:
return p.datasz;
case CAPN_PTR_LIST:
return p.len*8;
case CAPN_STRUCT:
return p.datasz + 8*p.ptrs;
case CAPN_LIST:
return p.len * (p.datasz + 8*p.ptrs) + 8*p.is_composite_list;
default:
return 0;
}
}
static int copy_ptr(struct capn_segment *seg, char *data, struct capn_ptr *t, struct capn_ptr *f, int *dep) {
struct capn *c = seg->capn;
struct copy *cp = NULL;
struct capn_tree **xcp;
char *fbegin = f->data - 8*f->is_composite_list;
char *fend = fbegin + data_size(*f);
int zero_sized = (fend == fbegin);
/* We always copy list members as it would otherwise be an
* overlapped pointer (the data is owned by the enclosing list).
* We do not bother with the overlapped lookup for zero sized
* structures/lists as they never overlap. Nor do we add them to
* the copy list as there is no data to be shared by multiple
* pointers.
*/
xcp = &c->copy;
while (*xcp && !zero_sized) {
cp = (struct copy*) *xcp;
if (fend <= cp->fbegin) {
xcp = &cp->hdr.link[0];
} else if (cp->fend <= fbegin) {
xcp = &cp->hdr.link[1];
} else if (is_ptr_equal(f, &cp->from)) {
/* we already have a copy so just point to that */
return write_ptr(seg, data, cp->to);
} else {
/* pointer to overlapped data */
return -1;
}
}
/* no copy found - have to create a new copy */
*t = new_clone(seg, *f);
if (write_ptr(seg, data, *t))
return -1;
/* add the copy to the copy tree so we can look for overlapping
* source pointers and handle recursive structures */
if (!zero_sized) {
struct copy *n;
struct capn_segment *cs = c->copylist;
/* need to allocate a struct copy */
if (!cs || cs->len + (int)sizeof(*n) > cs->cap) {
cs = c->create_local ? c->create_local(c->user, sizeof(*n)) : NULL;
if (!cs) {
/* can't allocate a copy structure */
return -1;
}
cs->next = c->copylist;
c->copylist = cs;
}
n = (struct copy*) (cs->data + cs->len);
cs->len += sizeof(*n);
n->from = *f;
n->to = *t;
n->fbegin = fbegin;
n->fend = fend;
*xcp = &n->hdr;
n->hdr.parent = &cp->hdr;
c->copy = capn_tree_insert(c->copy, &n->hdr);
}
/* minimize the number of types the main copy routine has to
* deal with to just CAPN_LIST and CAPN_PTR_LIST. ptr list only
* needs t->type, t->len, t->data, t->seg, f->data, f->seg to
* be valid */
switch (t->type) {
case CAPN_STRUCT:
if (t->datasz) {
memcpy(t->data, f->data, t->datasz);
t->data += t->datasz;
f->data += t->datasz;
}
if (t->ptrs) {
t->type = CAPN_PTR_LIST;
t->len = t->ptrs;
(*dep)++;
}
return 0;
case CAPN_BIT_LIST:
memcpy(t->data, f->data, t->datasz);
return 0;
case CAPN_LIST:
if (!t->len) {
/* empty list - nothing to copy */
} else if (t->ptrs && t->datasz) {
(*dep)++;
} else if (t->datasz) {
memcpy(t->data, f->data, t->len * t->datasz);
} else if (t->ptrs) {
t->type = CAPN_PTR_LIST;
t->len *= t->ptrs;
(*dep)++;
}
return 0;
case CAPN_PTR_LIST:
if (t->len) {
(*dep)++;
}
return 0;
default:
return -1;
}
}
static void copy_list_member(capn_ptr* t, capn_ptr *f, int *dep) {
/* copy struct data */
int sz = min(t->datasz, f->datasz);
memcpy(t->data, f->data, sz);
memset(t->data + sz, 0, t->datasz - sz);
t->data += t->datasz;
f->data += f->datasz;
/* reset excess pointers */
sz = min(t->ptrs, f->ptrs);
memset(t->data + sz, 0, 8*(t->ptrs - sz));
/* create a pointer list for the main loop to copy */
if (t->ptrs) {
t->type = CAPN_PTR_LIST;
t->len = t->ptrs;
(*dep)++;
}
}
#define MAX_COPY_DEPTH 32
/* TODO: handle CAPN_BIT_LIST and setting from an inner bit list member */
int capn_setp(capn_ptr p, int off, capn_ptr tgt) {
struct capn_ptr to[MAX_COPY_DEPTH], from[MAX_COPY_DEPTH];
char *data;
int err, dep = 0;
capn_resolve(&p);
if (tgt.type == CAPN_FAR_POINTER && tgt.seg->capn == p.seg->capn) {
uint64_t val = capn_flip64(*(uint64_t*) tgt.data);
if ((val & 3) == FAR_PTR) {
*(uint64_t*) p.data = *(uint64_t*) tgt.data;
return 0;
}
}
capn_resolve(&tgt);
switch (p.type) {
case CAPN_LIST:
if (off >= p.len || tgt.type != CAPN_STRUCT)
return -1;
to[0] = p;
to[0].data += off * (p.datasz + 8*p.ptrs);
from[0] = tgt;
copy_list_member(to, from, &dep);
break;
case CAPN_PTR_LIST:
if (off >= p.len)
return -1;
data = p.data + 8*off;
goto copy_ptr;
case CAPN_STRUCT:
if (off >= p.ptrs)
return -1;
data = p.data + p.datasz + 8*off;
goto copy_ptr;
copy_ptr:
err = write_ptr(p.seg, data, tgt);
if (err != NEED_TO_COPY)
return err;
/* Depth first copy the source whilst using a pointer stack to
* maintain the ptr to set and size left to copy at each level.
* We also maintain a rbtree (capn->copy) of the copies indexed
* by the source data. This way we can detect overlapped
* pointers in the source (and bail) and recursive structures
* (and point to the previous copy).
*/
from[0] = tgt;
if (copy_ptr(p.seg, data, to, from, &dep))
return -1;
break;
default:
return -1;
}
while (dep) {
struct capn_ptr *tc = &to[dep-1], *tn = &to[dep];
struct capn_ptr *fc = &from[dep-1], *fn = &from[dep];
if (dep+1 == MAX_COPY_DEPTH) {
return -1;
}
if (!tc->len) {
dep--;
continue;
}
if (tc->type == CAPN_LIST) {
*fn = capn_getp(*fc, 0, 1);
*tn = capn_getp(*tc, 0, 1);
copy_list_member(tn, fn, &dep);
fc->data += fc->datasz + 8*fc->ptrs;
tc->data += tc->datasz + 8*tc->ptrs;
tc->len--;
} else { /* CAPN_PTR_LIST */
*fn = read_ptr(fc->seg, fc->data);
if (fn->type && copy_ptr(tc->seg, tc->data, tn, fn, &dep))
return -1;
fc->data += 8;
tc->data += 8;
tc->len--;
}
}
return 0;
}
/* TODO: handle CAPN_LIST, CAPN_PTR_LIST for bit lists */
int capn_get1(capn_list1 l, int off) {
return l.p.type == CAPN_BIT_LIST
&& off < l.p.len
&& (l.p.data[off/8] & (1 << (off%8))) != 0;
}
int capn_set1(capn_list1 l, int off, int val) {
if (l.p.type != CAPN_BIT_LIST || off >= l.p.len)
return -1;
if (val) {
l.p.data[off/8] |= 1 << (off%8);
} else {
l.p.data[off/8] &= ~(1 << (off%8));
}
return 0;
}
int capn_getv1(capn_list1 l, int off, uint8_t *data, int sz) {
/* Note we only support aligned reads */
int bsz;
capn_ptr p;
capn_resolve(&l.p);
p = l.p;
if (p.type != CAPN_BIT_LIST || (off & 7) != 0)
return -1;
bsz = (sz + 7) / 8;
off /= 8;
if (off + sz > p.datasz) {
memcpy(data, p.data + off, p.datasz - off);
return p.len - off*8;
} else {
memcpy(data, p.data + off, bsz);
return sz;
}
}
int capn_setv1(capn_list1 l, int off, const uint8_t *data, int sz) {
/* Note we only support aligned writes */
int bsz;
capn_ptr p = l.p;
if (p.type != CAPN_BIT_LIST || (off & 7) != 0)
return -1;
bsz = (sz + 7) / 8;
off /= 8;
if (off + sz > p.datasz) {
memcpy(p.data + off, data, p.datasz - off);
return p.len - off*8;
} else {
memcpy(p.data + off, data, bsz);
return sz;
}
}
/* pull out whether we add a tag or not as a define so the unit test can
* test double far pointers by not creating tags */
#ifndef ADD_TAG
#define ADD_TAG 1
#endif
static void new_object(capn_ptr *p, int bytes) {
struct capn_segment *s = p->seg;
if (!s) {
memset(p, 0, sizeof(*p));
return;
}
/* pointer needs to be initialised to get a valid offset on write */
if (!bytes) {
p->data = s->data + s->len;
return;
}
/* all allocations are 8 byte aligned */
bytes = (bytes + 7) & ~7;
if (s->len + bytes <= s->cap) {
p->data = s->data + s->len;
s->len += bytes;
return;
}
/* add a tag whenever we switch segments so that write_ptr can
* use it */
p->data = new_data(s->capn, bytes + ADD_TAG*8, &p->seg);
if (!p->data) {
memset(p, 0, sizeof(*p));
return;
}
if (ADD_TAG) {
write_ptr_tag(p->data, *p, 0);
p->data += 8;
p->has_ptr_tag = 1;
}
}
capn_ptr capn_root(struct capn *c) {
capn_ptr r = {CAPN_PTR_LIST};
r.seg = lookup_segment(c, NULL, 0);
r.data = r.seg ? r.seg->data : new_data(c, 8, &r.seg);
r.len = 1;
if (!r.seg || r.seg->cap < 8) {
memset(&r, 0, sizeof(r));
} else if (r.seg->len < 8) {
r.seg->len = 8;
}
return r;
}
capn_ptr capn_new_struct(struct capn_segment *seg, int datasz, int ptrs) {
capn_ptr p = {CAPN_STRUCT};
p.seg = seg;
p.datasz = (datasz + 7) & ~7;
p.ptrs = ptrs;
new_object(&p, p.datasz + 8*p.ptrs);
return p;
}
capn_ptr capn_new_list(struct capn_segment *seg, int sz, int datasz, int ptrs) {
capn_ptr p = {CAPN_LIST};
p.seg = seg;
p.len = sz;
if (ptrs || datasz > 8) {
p.is_composite_list = 1;
p.datasz = (datasz + 7) & ~7;
p.ptrs = ptrs;
new_object(&p, p.len * (p.datasz + 8*p.ptrs) + 8);
if (p.data) {
uint64_t hdr = STRUCT_PTR | (U64(p.len) << 2) | (U64(p.datasz/8) << 32) | (U64(ptrs) << 48);
*(uint64_t*) p.data = capn_flip64(hdr);
p.data += 8;
}
} else if (datasz > 4) {
p.datasz = 8;
new_object(&p, p.len * 8);
} else if (datasz > 2) {
p.datasz = 4;
new_object(&p, p.len * 4);
} else {
p.datasz = datasz;
new_object(&p, p.len * datasz);
}
return p;
}
capn_list1 capn_new_list1(struct capn_segment *seg, int sz) {
capn_list1 l = {{CAPN_BIT_LIST}};
l.p.seg = seg;
l.p.datasz = (sz+7)/8;
l.p.len = sz;
new_object(&l.p, l.p.datasz);
return l;
}
capn_ptr capn_new_ptr_list(struct capn_segment *seg, int sz) {
capn_ptr p = {CAPN_PTR_LIST};
p.seg = seg;
p.len = sz;
p.ptrs = 0;
p.datasz = 0;
new_object(&p, sz*8);
return p;
}
capn_ptr capn_new_string(struct capn_segment *seg, const char *str, ssize_t sz) {
capn_ptr p = {CAPN_LIST};
p.seg = seg;
p.len = ((sz >= 0) ? (size_t)sz : strlen(str)) + 1;
p.datasz = 1;
new_object(&p, p.len);
if (p.data) {
memcpy(p.data, str, p.len - 1);
p.data[p.len - 1] = '\0';
}
return p;
}
capn_text capn_get_text(capn_ptr p, int off, capn_text def) {
capn_ptr m = capn_getp(p, off, 1);
capn_text ret = def;
if (m.type == CAPN_LIST && m.datasz == 1 && m.len && m.data[m.len - 1] == 0) {
ret.seg = m.seg;
ret.str = m.data;
ret.len = m.len - 1;
}
return ret;
}
int capn_set_text(capn_ptr p, int off, capn_text tgt) {
capn_ptr m = {CAPN_NULL};
if (tgt.seg) {
m.type = CAPN_LIST;
m.seg = tgt.seg;
m.data = (char*)tgt.str;
m.len = tgt.len + 1;
m.datasz = 1;
} else if (tgt.str) {
m = capn_new_string(p.seg, tgt.str, tgt.len);
}
return capn_setp(p, off, m);
}
capn_data capn_get_data(capn_ptr p, int off) {
capn_data ret;
ret.p = capn_getp(p, off, 1);
if (ret.p.type != CAPN_LIST || ret.p.datasz != 1) {
memset(&ret, 0, sizeof(ret));
}
return ret;
}
#define SZ 8
#include "capn-list.inc"
#undef SZ
#define SZ 16
#include "capn-list.inc"
#undef SZ
#define SZ 32
#include "capn-list.inc"
#undef SZ
#define SZ 64
#include "capn-list.inc"
#undef SZ