blob: 0d8efb09b78ddb3ae92e3ecc7734baeb2ae4d5b4 [file] [log] [blame]
/*
* Duktape.Buffer, Node.js Buffer, and Khronos/ES6 TypedArray built-ins
*/
#include "duk_internal.h"
/*
* Misc helpers
*/
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
/* Map DUK_HBUFFEROBJECT_ELEM_xxx to duk_hobject class number.
* Sync with duk_hbufferobject.h and duk_hobject.h.
*/
static const duk_uint8_t duk__buffer_class_from_elemtype[9] = {
DUK_HOBJECT_CLASS_UINT8ARRAY,
DUK_HOBJECT_CLASS_UINT8CLAMPEDARRAY,
DUK_HOBJECT_CLASS_INT8ARRAY,
DUK_HOBJECT_CLASS_UINT16ARRAY,
DUK_HOBJECT_CLASS_INT16ARRAY,
DUK_HOBJECT_CLASS_UINT32ARRAY,
DUK_HOBJECT_CLASS_INT32ARRAY,
DUK_HOBJECT_CLASS_FLOAT32ARRAY,
DUK_HOBJECT_CLASS_FLOAT64ARRAY
};
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
/* Map DUK_HBUFFEROBJECT_ELEM_xxx to prototype object built-in index.
* Sync with duk_hbufferobject.h.
*/
static const duk_uint8_t duk__buffer_proto_from_elemtype[9] = {
DUK_BIDX_UINT8ARRAY_PROTOTYPE,
DUK_BIDX_UINT8CLAMPEDARRAY_PROTOTYPE,
DUK_BIDX_INT8ARRAY_PROTOTYPE,
DUK_BIDX_UINT16ARRAY_PROTOTYPE,
DUK_BIDX_INT16ARRAY_PROTOTYPE,
DUK_BIDX_UINT32ARRAY_PROTOTYPE,
DUK_BIDX_INT32ARRAY_PROTOTYPE,
DUK_BIDX_FLOAT32ARRAY_PROTOTYPE,
DUK_BIDX_FLOAT64ARRAY_PROTOTYPE
};
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
/* Map DUK__FLX_xxx to byte size.
*/
static const duk_uint8_t duk__buffer_nbytes_from_fldtype[6] = {
1, /* DUK__FLD_8BIT */
2, /* DUK__FLD_16BIT */
4, /* DUK__FLD_32BIT */
4, /* DUK__FLD_FLOAT */
8, /* DUK__FLD_DOUBLE */
0 /* DUK__FLD_VARINT; not relevant here */
};
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
/* Bitfield for each DUK_HBUFFEROBJECT_ELEM_xxx indicating which element types
* are compatible with a blind byte copy for the TypedArray set() method (also
* used for TypedArray constructor). Array index is target buffer elem type,
* bitfield indicates compatible source types. The types must have same byte
* size and they must be coercion compatible.
*/
static duk_uint16_t duk__buffer_elemtype_copy_compatible[9] = {
/* xxx -> DUK_HBUFFEROBJECT_ELEM_UINT8 */
(1U << DUK_HBUFFEROBJECT_ELEM_UINT8) |
(1U << DUK_HBUFFEROBJECT_ELEM_UINT8CLAMPED) |
(1U << DUK_HBUFFEROBJECT_ELEM_INT8),
/* xxx -> DUK_HBUFFEROBJECT_ELEM_UINT8CLAMPED
* Note: INT8 is -not- copy compatible, e.g. -1 would coerce to 0x00.
*/
(1U << DUK_HBUFFEROBJECT_ELEM_UINT8) |
(1U << DUK_HBUFFEROBJECT_ELEM_UINT8CLAMPED),
/* xxx -> DUK_HBUFFEROBJECT_ELEM_INT8 */
(1U << DUK_HBUFFEROBJECT_ELEM_UINT8) |
(1U << DUK_HBUFFEROBJECT_ELEM_UINT8CLAMPED) |
(1U << DUK_HBUFFEROBJECT_ELEM_INT8),
/* xxx -> DUK_HBUFFEROBJECT_ELEM_UINT16 */
(1U << DUK_HBUFFEROBJECT_ELEM_UINT16) |
(1U << DUK_HBUFFEROBJECT_ELEM_INT16),
/* xxx -> DUK_HBUFFEROBJECT_ELEM_INT16 */
(1U << DUK_HBUFFEROBJECT_ELEM_UINT16) |
(1U << DUK_HBUFFEROBJECT_ELEM_INT16),
/* xxx -> DUK_HBUFFEROBJECT_ELEM_UINT32 */
(1U << DUK_HBUFFEROBJECT_ELEM_UINT32) |
(1U << DUK_HBUFFEROBJECT_ELEM_INT32),
/* xxx -> DUK_HBUFFEROBJECT_ELEM_INT32 */
(1U << DUK_HBUFFEROBJECT_ELEM_UINT32) |
(1U << DUK_HBUFFEROBJECT_ELEM_INT32),
/* xxx -> DUK_HBUFFEROBJECT_ELEM_FLOAT32 */
(1U << DUK_HBUFFEROBJECT_ELEM_FLOAT32),
/* xxx -> DUK_HBUFFEROBJECT_ELEM_FLOAT64 */
(1U << DUK_HBUFFEROBJECT_ELEM_FLOAT64)
};
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
/* Shared helper. */
DUK_LOCAL duk_hbufferobject *duk__getrequire_bufobj_this(duk_context *ctx, duk_bool_t throw_flag) {
duk_hthread *thr;
duk_tval *tv;
duk_hbufferobject *h_this;
DUK_ASSERT(ctx != NULL);
thr = (duk_hthread *) ctx;
tv = duk_get_borrowed_this_tval(ctx);
DUK_ASSERT(tv != NULL);
if (DUK_TVAL_IS_OBJECT(tv)) {
h_this = (duk_hbufferobject *) DUK_TVAL_GET_OBJECT(tv);
DUK_ASSERT(h_this != NULL);
if (DUK_HOBJECT_IS_BUFFEROBJECT((duk_hobject *) h_this)) {
DUK_ASSERT_HBUFFEROBJECT_VALID(h_this);
return h_this;
}
}
if (throw_flag) {
DUK_ERROR_TYPE(thr, DUK_STR_NOT_BUFFER);
}
return NULL;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
/* Check that 'this' is a duk_hbufferobject and return a pointer to it. */
DUK_LOCAL duk_hbufferobject *duk__get_bufobj_this(duk_context *ctx) {
return duk__getrequire_bufobj_this(ctx, 0);
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
/* Check that 'this' is a duk_hbufferobject and return a pointer to it
* (NULL if not).
*/
DUK_LOCAL duk_hbufferobject *duk__require_bufobj_this(duk_context *ctx) {
return duk__getrequire_bufobj_this(ctx, 1);
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
/* Check that value is a duk_hbufferobject and return a pointer to it. */
DUK_LOCAL duk_hbufferobject *duk__require_bufobj_value(duk_context *ctx, duk_idx_t index) {
duk_hthread *thr;
duk_tval *tv;
duk_hbufferobject *h_obj;
thr = (duk_hthread *) ctx;
/* Don't accept relative indices now. */
DUK_ASSERT(index >= 0);
tv = duk_require_tval(ctx, index);
DUK_ASSERT(tv != NULL);
if (DUK_TVAL_IS_OBJECT(tv)) {
h_obj = (duk_hbufferobject *) DUK_TVAL_GET_OBJECT(tv);
DUK_ASSERT(h_obj != NULL);
if (DUK_HOBJECT_IS_BUFFEROBJECT((duk_hobject *) h_obj)) {
DUK_ASSERT_HBUFFEROBJECT_VALID(h_obj);
return h_obj;
}
}
DUK_ERROR_TYPE(thr, DUK_STR_NOT_BUFFER);
return NULL; /* not reachable */
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
DUK_LOCAL void duk__set_bufobj_buffer(duk_context *ctx, duk_hbufferobject *h_bufobj, duk_hbuffer *h_val) {
duk_hthread *thr;
thr = (duk_hthread *) ctx;
DUK_UNREF(thr);
DUK_ASSERT(ctx != NULL);
DUK_ASSERT(h_bufobj != NULL);
DUK_ASSERT(h_bufobj->buf == NULL); /* no need to decref */
DUK_ASSERT(h_val != NULL);
DUK_ASSERT_HBUFFEROBJECT_VALID(h_bufobj);
h_bufobj->buf = h_val;
DUK_HBUFFER_INCREF(thr, h_val);
h_bufobj->length = (duk_uint_t) DUK_HBUFFER_GET_SIZE(h_val);
DUK_ASSERT(h_bufobj->shift == 0);
DUK_ASSERT(h_bufobj->elem_type == DUK_HBUFFEROBJECT_ELEM_UINT8);
DUK_ASSERT_HBUFFEROBJECT_VALID(h_bufobj);
}
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
DUK_LOCAL duk_hbufferobject *duk__push_arraybuffer_with_length(duk_context *ctx, duk_uint_t len) {
duk_hbuffer *h_val;
duk_hbufferobject *h_bufobj;
(void) duk_push_fixed_buffer(ctx, (duk_size_t) len);
h_val = (duk_hbuffer *) duk_get_hbuffer(ctx, -1);
DUK_ASSERT(h_val != NULL);
h_bufobj = duk_push_bufferobject_raw(ctx,
DUK_HOBJECT_FLAG_EXTENSIBLE |
DUK_HOBJECT_FLAG_BUFFEROBJECT |
DUK_HOBJECT_CLASS_AS_FLAGS(DUK_HOBJECT_CLASS_ARRAYBUFFER),
DUK_BIDX_ARRAYBUFFER_PROTOTYPE);
DUK_ASSERT(h_bufobj != NULL);
duk__set_bufobj_buffer(ctx, h_bufobj, h_val);
DUK_ASSERT_HBUFFEROBJECT_VALID(h_bufobj);
return h_bufobj;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
/* Shared offset/length coercion helper. */
DUK_LOCAL void duk__resolve_offset_opt_length(duk_context *ctx,
duk_hbufferobject *h_bufarg,
duk_idx_t idx_offset,
duk_idx_t idx_length,
duk_uint_t *out_offset,
duk_uint_t *out_length,
duk_bool_t throw_flag) {
duk_hthread *thr;
duk_int_t offset_signed;
duk_int_t length_signed;
duk_uint_t offset;
duk_uint_t length;
thr = (duk_hthread *) ctx;
DUK_UNREF(thr);
offset_signed = duk_to_int(ctx, idx_offset);
if (offset_signed < 0) {
goto fail_range;
}
offset = (duk_uint_t) offset_signed;
if (offset > h_bufarg->length) {
goto fail_range;
}
DUK_ASSERT_DISABLE(offset >= 0); /* unsigned */
DUK_ASSERT(offset <= h_bufarg->length);
if (duk_is_undefined(ctx, idx_length)) {
DUK_ASSERT(h_bufarg->length >= offset);
length = h_bufarg->length - offset; /* >= 0 */
} else {
length_signed = duk_to_int(ctx, idx_length);
if (length_signed < 0) {
goto fail_range;
}
length = (duk_uint_t) length_signed;
DUK_ASSERT(h_bufarg->length >= offset);
if (length > h_bufarg->length - offset) {
/* Unlike for negative arguments, some call sites
* want length to be clamped if it's positive.
*/
if (throw_flag) {
goto fail_range;
} else {
length = h_bufarg->length - offset;
}
}
}
DUK_ASSERT_DISABLE(length >= 0); /* unsigned */
DUK_ASSERT(offset + length <= h_bufarg->length);
*out_offset = offset;
*out_length = length;
return;
fail_range:
DUK_ERROR_RANGE(thr, DUK_STR_INVALID_CALL_ARGS);
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
/* Shared lenient buffer length clamping helper. No negative indices, no
* element/byte shifting.
*/
DUK_LOCAL void duk__clamp_startend_nonegidx_noshift(duk_context *ctx,
duk_hbufferobject *h_bufobj,
duk_idx_t idx_start,
duk_idx_t idx_end,
duk_int_t *out_start_offset,
duk_int_t *out_end_offset) {
duk_int_t buffer_length;
duk_int_t start_offset;
duk_int_t end_offset;
DUK_ASSERT(out_start_offset != NULL);
DUK_ASSERT(out_end_offset != NULL);
buffer_length = (duk_int_t) h_bufobj->length;
/* undefined coerces to zero which is correct */
start_offset = duk_to_int_clamped(ctx, idx_start, 0, buffer_length);
if (duk_is_undefined(ctx, idx_end)) {
end_offset = buffer_length;
} else {
end_offset = duk_to_int_clamped(ctx, idx_end, start_offset, buffer_length);
}
DUK_ASSERT(start_offset >= 0);
DUK_ASSERT(start_offset <= buffer_length);
DUK_ASSERT(end_offset >= 0);
DUK_ASSERT(end_offset <= buffer_length);
DUK_ASSERT(start_offset <= end_offset);
*out_start_offset = start_offset;
*out_end_offset = end_offset;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
/* Shared lenient buffer length clamping helper. Indices are treated as
* element indices (though output values are byte offsets) which only
* really matters for TypedArray views as other buffer object have a zero
* shift. Negative indices are counted from end of input slice; crossed
* indices are clamped to zero length; and final indices are clamped
* against input slice. Used for e.g. ArrayBuffer slice().
*/
DUK_LOCAL void duk__clamp_startend_negidx_shifted(duk_context *ctx,
duk_hbufferobject *h_bufobj,
duk_idx_t idx_start,
duk_idx_t idx_end,
duk_int_t *out_start_offset,
duk_int_t *out_end_offset) {
duk_int_t buffer_length;
duk_int_t start_offset;
duk_int_t end_offset;
DUK_ASSERT(out_start_offset != NULL);
DUK_ASSERT(out_end_offset != NULL);
buffer_length = (duk_int_t) h_bufobj->length;
buffer_length >>= h_bufobj->shift; /* as elements */
/* Resolve start/end offset as element indices first; arguments
* at idx_start/idx_end are element offsets. Working with element
* indices first also avoids potential for wrapping.
*/
start_offset = duk_to_int(ctx, idx_start);
if (start_offset < 0) {
start_offset = buffer_length + start_offset;
}
if (duk_is_undefined(ctx, idx_end)) {
end_offset = buffer_length;
} else {
end_offset = duk_to_int(ctx, idx_end);
if (end_offset < 0) {
end_offset = buffer_length + end_offset;
}
}
/* Note: start_offset/end_offset can still be < 0 here. */
if (start_offset < 0) {
start_offset = 0;
} else if (start_offset > buffer_length) {
start_offset = buffer_length;
}
if (end_offset < start_offset) {
end_offset = start_offset;
} else if (end_offset > buffer_length) {
end_offset = buffer_length;
}
DUK_ASSERT(start_offset >= 0);
DUK_ASSERT(start_offset <= buffer_length);
DUK_ASSERT(end_offset >= 0);
DUK_ASSERT(end_offset <= buffer_length);
DUK_ASSERT(start_offset <= end_offset);
/* Convert indices to byte offsets. */
start_offset <<= h_bufobj->shift;
end_offset <<= h_bufobj->shift;
*out_start_offset = start_offset;
*out_end_offset = end_offset;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
/*
* Indexed read/write helpers (also used from outside this file)
*/
DUK_INTERNAL void duk_hbufferobject_push_validated_read(duk_context *ctx, duk_hbufferobject *h_bufobj, duk_uint8_t *p, duk_small_uint_t elem_size) {
duk_double_union du;
DUK_MEMCPY((void *) du.uc, (const void *) p, (size_t) elem_size);
switch (h_bufobj->elem_type) {
case DUK_HBUFFEROBJECT_ELEM_UINT8:
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
case DUK_HBUFFEROBJECT_ELEM_UINT8CLAMPED:
#endif
duk_push_uint(ctx, (duk_uint_t) du.uc[0]);
break;
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
/* These are not needed when only Duktape.Buffer is supported. */
case DUK_HBUFFEROBJECT_ELEM_INT8:
duk_push_int(ctx, (duk_int_t) (duk_int8_t) du.uc[0]);
break;
case DUK_HBUFFEROBJECT_ELEM_UINT16:
duk_push_uint(ctx, (duk_uint_t) du.us[0]);
break;
case DUK_HBUFFEROBJECT_ELEM_INT16:
duk_push_int(ctx, (duk_int_t) (duk_int16_t) du.us[0]);
break;
case DUK_HBUFFEROBJECT_ELEM_UINT32:
duk_push_uint(ctx, (duk_uint_t) du.ui[0]);
break;
case DUK_HBUFFEROBJECT_ELEM_INT32:
duk_push_int(ctx, (duk_int_t) (duk_int32_t) du.ui[0]);
break;
case DUK_HBUFFEROBJECT_ELEM_FLOAT32:
duk_push_number(ctx, (duk_double_t) du.f[0]);
break;
case DUK_HBUFFEROBJECT_ELEM_FLOAT64:
duk_push_number(ctx, (duk_double_t) du.d);
break;
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
default:
DUK_UNREACHABLE();
}
}
DUK_INTERNAL void duk_hbufferobject_validated_write(duk_context *ctx, duk_hbufferobject *h_bufobj, duk_uint8_t *p, duk_small_uint_t elem_size) {
duk_double_union du;
/* NOTE! Caller must ensure that any side effects from the
* coercions below are safe. If that cannot be guaranteed
* (which is normally the case), caller must coerce the
* argument using duk_to_number() before any pointer
* validations; the result of duk_to_number() always coerces
* without side effects here.
*/
switch (h_bufobj->elem_type) {
case DUK_HBUFFEROBJECT_ELEM_UINT8:
du.uc[0] = (duk_uint8_t) duk_to_uint32(ctx, -1);
break;
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
/* These are not needed when only Duktape.Buffer is supported. */
case DUK_HBUFFEROBJECT_ELEM_UINT8CLAMPED:
du.uc[0] = (duk_uint8_t) duk_to_uint8clamped(ctx, -1);
break;
case DUK_HBUFFEROBJECT_ELEM_INT8:
du.uc[0] = (duk_uint8_t) duk_to_int32(ctx, -1);
break;
case DUK_HBUFFEROBJECT_ELEM_UINT16:
du.us[0] = (duk_uint16_t) duk_to_uint32(ctx, -1);
break;
case DUK_HBUFFEROBJECT_ELEM_INT16:
du.us[0] = (duk_uint16_t) duk_to_int32(ctx, -1);
break;
case DUK_HBUFFEROBJECT_ELEM_UINT32:
du.ui[0] = (duk_uint32_t) duk_to_uint32(ctx, -1);
break;
case DUK_HBUFFEROBJECT_ELEM_INT32:
du.ui[0] = (duk_uint32_t) duk_to_int32(ctx, -1);
break;
case DUK_HBUFFEROBJECT_ELEM_FLOAT32:
du.f[0] = (duk_float_t) duk_to_number(ctx, -1);
break;
case DUK_HBUFFEROBJECT_ELEM_FLOAT64:
du.d = (duk_double_t) duk_to_number(ctx, -1);
break;
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
default:
DUK_UNREACHABLE();
}
DUK_MEMCPY((void *) p, (const void *) du.uc, (size_t) elem_size);
}
/*
* Duktape.Buffer: constructor
*/
DUK_INTERNAL duk_ret_t duk_bi_buffer_constructor(duk_context *ctx) {
duk_hthread *thr;
duk_size_t buf_size;
duk_small_int_t buf_dynamic;
duk_uint8_t *buf_data;
const duk_uint8_t *src_data;
thr = (duk_hthread *) ctx;
DUK_UNREF(thr);
/*
* Constructor arguments are currently somewhat compatible with
* (keep it that way if possible):
*
* http://nodejs.org/api/buffer.html
*
* Note that the ToBuffer() coercion (duk_to_buffer()) does NOT match
* the constructor behavior.
*/
buf_dynamic = duk_get_boolean(ctx, 1); /* default to false */
switch (duk_get_type(ctx, 0)) {
case DUK_TYPE_NUMBER: {
/* new buffer of specified size */
buf_size = (duk_size_t) duk_to_int(ctx, 0);
(void) duk_push_buffer(ctx, buf_size, buf_dynamic);
break;
}
case DUK_TYPE_BUFFER: {
/* return input buffer, converted to a Duktape.Buffer object
* if called as a constructor (no change if called as a
* function).
*/
duk_set_top(ctx, 1);
break;
}
case DUK_TYPE_STRING: {
/* new buffer with string contents */
src_data = (const duk_uint8_t *) duk_get_lstring(ctx, 0, &buf_size);
DUK_ASSERT(src_data != NULL); /* even for zero-length string */
buf_data = (duk_uint8_t *) duk_push_buffer(ctx, buf_size, buf_dynamic);
DUK_MEMCPY((void *) buf_data, (const void *) src_data, (size_t) buf_size);
break;
}
case DUK_TYPE_OBJECT: {
/* For all duk_hbufferobjects, get the plain buffer inside
* without making a copy. This is compatible with Duktape 1.2
* but means that a slice/view information is ignored and the
* full underlying buffer is returned.
*
* If called as a constructor, a new Duktape.Buffer object
* pointing to the same plain buffer is created below.
*/
duk_hbufferobject *h_bufobj;
h_bufobj = (duk_hbufferobject *) duk_get_hobject(ctx, 0);
DUK_ASSERT(h_bufobj != NULL);
if (!DUK_HOBJECT_IS_BUFFEROBJECT((duk_hobject *) h_bufobj)) {
return DUK_RET_TYPE_ERROR;
}
if (h_bufobj->buf == NULL) {
return DUK_RET_TYPE_ERROR;
}
duk_push_hbuffer(ctx, h_bufobj->buf);
break;
}
case DUK_TYPE_NONE:
default: {
return DUK_RET_TYPE_ERROR;
}
}
DUK_ASSERT(duk_is_buffer(ctx, -1));
/* stack is unbalanced, but: [ <something> buf ] */
if (duk_is_constructor_call(ctx)) {
duk_hbufferobject *h_bufobj;
duk_hbuffer *h_val;
h_val = duk_get_hbuffer(ctx, -1);
DUK_ASSERT(h_val != NULL);
h_bufobj = duk_push_bufferobject_raw(ctx,
DUK_HOBJECT_FLAG_EXTENSIBLE |
DUK_HOBJECT_FLAG_BUFFEROBJECT |
DUK_HOBJECT_CLASS_AS_FLAGS(DUK_HOBJECT_CLASS_BUFFER),
DUK_BIDX_BUFFER_PROTOTYPE);
DUK_ASSERT(h_bufobj != NULL);
duk__set_bufobj_buffer(ctx, h_bufobj, h_val);
DUK_ASSERT_HBUFFEROBJECT_VALID(h_bufobj);
}
/* Note: unbalanced stack on purpose */
return 1;
}
/*
* Node.js Buffer: constructor
*/
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
DUK_INTERNAL duk_ret_t duk_bi_nodejs_buffer_constructor(duk_context *ctx) {
/* Internal class is Object: Object.prototype.toString.call(new Buffer(0))
* prints "[object Object]".
*/
duk_int_t len;
duk_int_t i;
duk_hbuffer *h_buf;
duk_hbufferobject *h_bufobj;
duk_size_t buf_size;
switch (duk_get_type(ctx, 0)) {
case DUK_TYPE_BUFFER: {
/* Custom behavior: plain buffer is used as internal buffer
* without making a copy (matches Duktape.Buffer).
*/
duk_set_top(ctx, 1); /* -> [ buffer ] */
break;
}
case DUK_TYPE_NUMBER: {
len = duk_to_int_clamped(ctx, 0, 0, DUK_INT_MAX);
(void) duk_push_fixed_buffer(ctx, (duk_size_t) len);
break;
}
case DUK_TYPE_OBJECT: {
duk_uint8_t *buf;
(void) duk_get_prop_string(ctx, 0, "length");
len = duk_to_int_clamped(ctx, -1, 0, DUK_INT_MAX);
duk_pop(ctx);
buf = (duk_uint8_t *) duk_push_fixed_buffer(ctx, (duk_size_t) len);
for (i = 0; i < len; i++) {
/* XXX: fast path for array arguments? */
duk_get_prop_index(ctx, 0, (duk_uarridx_t) i);
buf[i] = (duk_uint8_t) (duk_to_uint32(ctx, -1) & 0xffU);
duk_pop(ctx);
}
break;
}
case DUK_TYPE_STRING: {
/* ignore encoding for now */
duk_dup(ctx, 0);
(void) duk_to_buffer(ctx, -1, &buf_size);
break;
}
default:
return DUK_RET_TYPE_ERROR;
}
DUK_ASSERT(duk_is_buffer(ctx, -1));
h_buf = duk_get_hbuffer(ctx, -1);
DUK_ASSERT(h_buf != NULL);
h_bufobj = duk_push_bufferobject_raw(ctx,
DUK_HOBJECT_FLAG_EXTENSIBLE |
DUK_HOBJECT_FLAG_BUFFEROBJECT |
DUK_HOBJECT_CLASS_AS_FLAGS(DUK_HOBJECT_CLASS_BUFFER),
DUK_BIDX_NODEJS_BUFFER_PROTOTYPE);
DUK_ASSERT(h_bufobj != NULL);
h_bufobj->buf = h_buf;
DUK_HBUFFER_INCREF(thr, h_buf);
DUK_ASSERT(h_bufobj->offset == 0);
h_bufobj->length = (duk_int_t) DUK_HBUFFER_GET_SIZE(h_buf);
DUK_ASSERT(h_bufobj->elem_type == DUK_HBUFFEROBJECT_ELEM_UINT8);
DUK_ASSERT_HBUFFEROBJECT_VALID(h_bufobj);
return 1;
}
#else /* DUK_USE_BUFFEROBJECT_SUPPORT */
DUK_INTERNAL duk_ret_t duk_bi_nodejs_buffer_constructor(duk_context *ctx) {
DUK_UNREF(ctx);
return DUK_RET_UNSUPPORTED_ERROR;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
/*
* ArrayBuffer, DataView, and TypedArray constructors
*/
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
DUK_INTERNAL duk_ret_t duk_bi_arraybuffer_constructor(duk_context *ctx) {
duk_hthread *thr;
duk_hbufferobject *h_bufobj;
duk_hbuffer *h_val;
DUK_ASSERT_CTX_VALID(ctx);
thr = (duk_hthread *) ctx;
DUK_UNREF(thr);
/* XXX: function flag to make this automatic? */
if (!duk_is_constructor_call(ctx)) {
return DUK_RET_TYPE_ERROR;
}
if (duk_is_buffer(ctx, 0)) {
/* Custom behavior: plain buffer is used as internal buffer
* without making a copy (matches Duktape.Buffer).
*/
h_val = duk_get_hbuffer(ctx, 0);
DUK_ASSERT(h_val != NULL);
/* XXX: accept any duk_hbufferobject type as an input also? */
} else {
duk_int_t len;
len = duk_to_int(ctx, 0);
if (len < 0) {
goto fail_length;
}
(void) duk_push_fixed_buffer(ctx, (duk_size_t) len);
h_val = (duk_hbuffer *) duk_get_hbuffer(ctx, -1);
DUK_ASSERT(h_val != NULL);
#if !defined(DUK_USE_ZERO_BUFFER_DATA)
/* Khronos/ES6 requires zeroing even when DUK_USE_ZERO_BUFFER_DATA
* is not set.
*/
DUK_ASSERT(!DUK_HBUFFER_HAS_DYNAMIC((duk_hbuffer *) h_val));
DUK_MEMZERO((void *) DUK_HBUFFER_FIXED_GET_DATA_PTR(thr->heap, h_val), (duk_size_t) len);
#endif
}
h_bufobj = duk_push_bufferobject_raw(ctx,
DUK_HOBJECT_FLAG_EXTENSIBLE |
DUK_HOBJECT_FLAG_BUFFEROBJECT |
DUK_HOBJECT_CLASS_AS_FLAGS(DUK_HOBJECT_CLASS_ARRAYBUFFER),
DUK_BIDX_ARRAYBUFFER_PROTOTYPE);
DUK_ASSERT(h_bufobj != NULL);
duk__set_bufobj_buffer(ctx, h_bufobj, h_val);
DUK_ASSERT_HBUFFEROBJECT_VALID(h_bufobj);
return 1;
fail_length:
return DUK_RET_RANGE_ERROR;
}
#else /* DUK_USE_BUFFEROBJECT_SUPPORT */
DUK_INTERNAL duk_ret_t duk_bi_arraybuffer_constructor(duk_context *ctx) {
DUK_UNREF(ctx);
return DUK_RET_UNSUPPORTED_ERROR;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
/* Format of magic, bits:
* 0...1: elem size shift (0-3)
* 2...5: elem type (DUK_HBUFFEROBJECT_ELEM_xxx)
*/
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
DUK_INTERNAL duk_ret_t duk_bi_typedarray_constructor(duk_context *ctx) {
duk_hthread *thr;
duk_tval *tv;
duk_hobject *h_obj;
duk_hbufferobject *h_bufobj = NULL;
duk_hbufferobject *h_bufarr = NULL;
duk_hbufferobject *h_bufarg = NULL;
duk_hbuffer *h_val;
duk_small_uint_t magic;
duk_small_uint_t shift;
duk_small_uint_t elem_type;
duk_small_uint_t elem_size;
duk_small_uint_t class_num;
duk_small_uint_t proto_bidx;
duk_uint_t align_mask;
duk_uint_t elem_length;
duk_int_t elem_length_signed;
duk_uint_t byte_length;
duk_small_uint_t copy_mode;
thr = (duk_hthread *) ctx;
DUK_UNREF(thr);
/* XXX: function flag to make this automatic? */
if (!duk_is_constructor_call(ctx)) {
return DUK_RET_TYPE_ERROR;
}
/* We could fit built-in index into magic but that'd make the magic
* number dependent on built-in numbering (genbuiltins.py doesn't
* handle that yet). So map both class and prototype from the
* element type.
*/
magic = duk_get_current_magic(ctx);
shift = magic & 0x03; /* bits 0...1: shift */
elem_type = (magic >> 2) & 0x0f; /* bits 2...5: type */
elem_size = 1 << shift;
align_mask = elem_size - 1;
DUK_ASSERT(elem_type < sizeof(duk__buffer_proto_from_elemtype) / sizeof(duk_uint8_t));
proto_bidx = duk__buffer_proto_from_elemtype[elem_type];
DUK_ASSERT(proto_bidx < DUK_NUM_BUILTINS);
DUK_ASSERT(elem_type < sizeof(duk__buffer_class_from_elemtype) / sizeof(duk_uint8_t));
class_num = duk__buffer_class_from_elemtype[elem_type];
DUK_DD(DUK_DDPRINT("typedarray constructor, magic=%d, shift=%d, elem_type=%d, "
"elem_size=%d, proto_bidx=%d, class_num=%d",
(int) magic, (int) shift, (int) elem_type, (int) elem_size,
(int) proto_bidx, (int) class_num));
/* Argument variants. When the argument is an ArrayBuffer a view to
* the same buffer is created; otherwise a new ArrayBuffer is always
* created.
*/
tv = duk_get_tval(ctx, 0);
DUK_ASSERT(tv != NULL); /* arg count */
if (DUK_TVAL_IS_OBJECT(tv)) {
h_obj = DUK_TVAL_GET_OBJECT(tv);
DUK_ASSERT(h_obj != NULL);
if (DUK_HOBJECT_GET_CLASS_NUMBER(h_obj) == DUK_HOBJECT_CLASS_ARRAYBUFFER) {
/* ArrayBuffer: unlike any other argument variant, create
* a view into the existing buffer.
*/
duk_int_t byte_offset_signed;
duk_uint_t byte_offset;
h_bufarg = (duk_hbufferobject *) h_obj;
byte_offset_signed = duk_to_int(ctx, 1);
if (byte_offset_signed < 0) {
goto fail_arguments;
}
byte_offset = (duk_uint_t) byte_offset_signed;
if (byte_offset > h_bufarg->length ||
(byte_offset & align_mask) != 0) {
/* Must be >= 0 and multiple of element size. */
goto fail_arguments;
}
if (duk_is_undefined(ctx, 2)) {
DUK_ASSERT(h_bufarg->length >= byte_offset);
byte_length = h_bufarg->length - byte_offset;
if ((byte_length & align_mask) != 0) {
/* Must be element size multiple from
* start offset to end of buffer.
*/
goto fail_arguments;
}
elem_length = (byte_length >> shift);
} else {
elem_length_signed = duk_to_int(ctx, 2);
if (elem_length_signed < 0) {
goto fail_arguments;
}
elem_length = (duk_uint_t) elem_length_signed;
byte_length = elem_length << shift;
if ((byte_length >> shift) != elem_length) {
/* Byte length would overflow. */
/* XXX: easier check with less code? */
goto fail_arguments;
}
DUK_ASSERT(h_bufarg->length >= byte_offset);
if (byte_length > h_bufarg->length - byte_offset) {
/* Not enough data. */
goto fail_arguments;
}
}
DUK_UNREF(elem_length);
DUK_ASSERT_DISABLE(byte_offset >= 0);
DUK_ASSERT(byte_offset <= h_bufarg->length);
DUK_ASSERT_DISABLE(byte_length >= 0);
DUK_ASSERT(byte_offset + byte_length <= h_bufarg->length);
DUK_ASSERT((elem_length << shift) == byte_length);
h_bufobj = duk_push_bufferobject_raw(ctx,
DUK_HOBJECT_FLAG_EXTENSIBLE |
DUK_HOBJECT_FLAG_BUFFEROBJECT |
DUK_HOBJECT_CLASS_AS_FLAGS(class_num),
proto_bidx);
h_val = h_bufarg->buf;
if (h_val == NULL) {
return DUK_RET_TYPE_ERROR;
}
h_bufobj->buf = h_val;
DUK_HBUFFER_INCREF(thr, h_val);
h_bufobj->offset = h_bufarg->offset + byte_offset;
h_bufobj->length = byte_length;
h_bufobj->shift = (duk_uint8_t) shift;
h_bufobj->elem_type = (duk_uint8_t) elem_type;
h_bufobj->is_view = 1;
DUK_ASSERT_HBUFFEROBJECT_VALID(h_bufobj);
/* Set .buffer to the argument ArrayBuffer. */
duk_dup(ctx, 0);
duk_xdef_prop_stridx(ctx, -2, DUK_STRIDX_LC_BUFFER, DUK_PROPDESC_FLAGS_NONE);
duk_compact(ctx, -1);
return 1;
} else if (DUK_HOBJECT_IS_BUFFEROBJECT(h_obj)) {
/* TypedArray (or other non-ArrayBuffer duk_hbufferobject).
* Conceptually same behavior as for an Array-like argument,
* with a few fast paths.
*/
h_bufarg = (duk_hbufferobject *) h_obj;
DUK_ASSERT_HBUFFEROBJECT_VALID(h_bufarg);
elem_length_signed = (duk_int_t) (h_bufarg->length >> h_bufarg->shift);
if (h_bufarg->buf == NULL) {
return DUK_RET_TYPE_ERROR;
}
/* Select copy mode. Must take into account element
* compatibility and validity of the underlying source
* buffer.
*/
DUK_DDD(DUK_DDDPRINT("selecting copy mode for bufobj arg, "
"src byte_length=%ld, src shift=%d, "
"src/dst elem_length=%ld; "
"dst shift=%d -> dst byte_length=%ld",
(long) h_bufarg->length, (int) h_bufarg->shift,
(long) elem_length_signed, (int) shift,
(long) (elem_length_signed << shift)));
copy_mode = 2; /* default is explicit index read/write copy */
DUK_ASSERT(elem_type < sizeof(duk__buffer_elemtype_copy_compatible) / sizeof(duk_uint16_t));
if (DUK_HBUFFEROBJECT_VALID_SLICE(h_bufarg)) {
if ((duk__buffer_elemtype_copy_compatible[elem_type] & (1 << h_bufarg->elem_type)) != 0) {
DUK_DDD(DUK_DDDPRINT("source/target are copy compatible, memcpy"));
DUK_ASSERT(shift == h_bufarg->shift); /* byte sizes will match */
copy_mode = 0;
} else {
DUK_DDD(DUK_DDDPRINT("source/target not copy compatible but valid, fast copy"));
copy_mode = 1;
}
}
} else {
/* Array or Array-like */
elem_length_signed = (duk_int_t) duk_get_length(ctx, 0);
copy_mode = 2;
}
} else if (DUK_TVAL_IS_BUFFER(tv)) {
/* Accept plain buffer values like array initializers
* (new in Duktape 1.4.0).
*/
duk_hbuffer *h_srcbuf;
h_srcbuf = DUK_TVAL_GET_BUFFER(tv);
elem_length_signed = (duk_int_t) DUK_HBUFFER_GET_SIZE(h_srcbuf);
copy_mode = 2; /* XXX: could add fast path for u8 compatible views */
} else {
/* Non-object argument is simply int coerced, matches
* V8 behavior (except for "null", which we coerce to
* 0 but V8 TypeErrors).
*/
elem_length_signed = duk_to_int(ctx, 0);
copy_mode = 3;
}
if (elem_length_signed < 0) {
goto fail_arguments;
}
elem_length = (duk_uint_t) elem_length_signed;
byte_length = (duk_uint_t) (elem_length << shift);
if ((byte_length >> shift) != elem_length) {
/* Byte length would overflow. */
/* XXX: easier check with less code? */
goto fail_arguments;
}
DUK_DDD(DUK_DDDPRINT("elem_length=%ld, byte_length=%ld",
(long) elem_length, (long) byte_length));
/* ArrayBuffer argument is handled specially above; the rest of the
* argument variants are handled by shared code below.
*/
/* Push a new ArrayBuffer (becomes view .buffer) */
h_bufarr = duk__push_arraybuffer_with_length(ctx, byte_length);
DUK_ASSERT(h_bufarr != NULL);
h_val = h_bufarr->buf;
DUK_ASSERT(h_val != NULL);
/* Push the resulting view object and attach the ArrayBuffer. */
h_bufobj = duk_push_bufferobject_raw(ctx,
DUK_HOBJECT_FLAG_EXTENSIBLE |
DUK_HOBJECT_FLAG_BUFFEROBJECT |
DUK_HOBJECT_CLASS_AS_FLAGS(class_num),
proto_bidx);
h_bufobj->buf = h_val;
DUK_HBUFFER_INCREF(thr, h_val);
DUK_ASSERT(h_bufobj->offset == 0);
h_bufobj->length = byte_length;
h_bufobj->shift = (duk_uint8_t) shift;
h_bufobj->elem_type = (duk_uint8_t) elem_type;
h_bufobj->is_view = 1;
DUK_ASSERT_HBUFFEROBJECT_VALID(h_bufobj);
/* Set .buffer */
duk_dup(ctx, -2);
duk_xdef_prop_stridx(ctx, -2, DUK_STRIDX_LC_BUFFER, DUK_PROPDESC_FLAGS_NONE);
duk_compact(ctx, -1);
/* Copy values, the copy method depends on the arguments.
*
* Copy mode decision may depend on the validity of the underlying
* buffer of the source argument; there must be no harmful side effects
* from there to here for copy_mode to still be valid.
*/
DUK_DDD(DUK_DDDPRINT("copy mode: %d", (int) copy_mode));
switch (copy_mode) {
case 0: {
/* Use byte copy. */
duk_uint8_t *p_src;
duk_uint8_t *p_dst;
DUK_ASSERT(h_bufobj != NULL);
DUK_ASSERT(h_bufobj->buf != NULL);
DUK_ASSERT(DUK_HBUFFEROBJECT_VALID_SLICE(h_bufobj));
DUK_ASSERT(h_bufarg != NULL);
DUK_ASSERT(h_bufarg->buf != NULL);
DUK_ASSERT(DUK_HBUFFEROBJECT_VALID_SLICE(h_bufarg));
p_dst = DUK_HBUFFEROBJECT_GET_SLICE_BASE(thr->heap, h_bufobj);
p_src = DUK_HBUFFEROBJECT_GET_SLICE_BASE(thr->heap, h_bufarg);
DUK_DDD(DUK_DDDPRINT("using memcpy: p_src=%p, p_dst=%p, byte_length=%ld",
(void *) p_src, (void *) p_dst, (long) byte_length));
DUK_MEMCPY((void *) p_dst, (const void *) p_src, (size_t) byte_length);
break;
}
case 1: {
/* Copy values through direct validated reads and writes. */
duk_small_uint_t src_elem_size;
duk_small_uint_t dst_elem_size;
duk_uint8_t *p_src;
duk_uint8_t *p_src_end;
duk_uint8_t *p_dst;
DUK_ASSERT(h_bufobj != NULL);
DUK_ASSERT(h_bufobj->buf != NULL);
DUK_ASSERT(DUK_HBUFFEROBJECT_VALID_SLICE(h_bufobj));
DUK_ASSERT(h_bufarg != NULL);
DUK_ASSERT(h_bufarg->buf != NULL);
DUK_ASSERT(DUK_HBUFFEROBJECT_VALID_SLICE(h_bufarg));
src_elem_size = 1 << h_bufarg->shift;
dst_elem_size = elem_size;
p_src = DUK_HBUFFEROBJECT_GET_SLICE_BASE(thr->heap, h_bufarg);
p_dst = DUK_HBUFFEROBJECT_GET_SLICE_BASE(thr->heap, h_bufobj);
p_src_end = p_src + h_bufarg->length;
DUK_DDD(DUK_DDDPRINT("using fast copy: p_src=%p, p_src_end=%p, p_dst=%p, "
"src_elem_size=%d, dst_elem_size=%d",
(void *) p_src, (void *) p_src_end, (void *) p_dst,
(int) src_elem_size, (int) dst_elem_size));
while (p_src != p_src_end) {
DUK_DDD(DUK_DDDPRINT("fast path per element copy loop: "
"p_src=%p, p_src_end=%p, p_dst=%p",
(void *) p_src, (void *) p_src_end, (void *) p_dst));
/* A validated read() is always a number, so it's write coercion
* is always side effect free an won't invalidate pointers etc.
*/
duk_hbufferobject_push_validated_read(ctx, h_bufarg, p_src, src_elem_size);
duk_hbufferobject_validated_write(ctx, h_bufobj, p_dst, dst_elem_size);
duk_pop(ctx);
p_src += src_elem_size;
p_dst += dst_elem_size;
}
break;
}
case 2: {
/* Copy values by index reads and writes. Let virtual
* property handling take care of coercion.
*/
duk_uint_t i;
DUK_DDD(DUK_DDDPRINT("using slow copy"));
for (i = 0; i < elem_length; i++) {
duk_get_prop_index(ctx, 0, (duk_uarridx_t) i);
duk_put_prop_index(ctx, -2, (duk_uarridx_t) i);
}
break;
}
default:
case 3: {
/* No copy, leave zero bytes in the buffer. There's no
* ambiguity with Float32/Float64 because zero bytes also
* represent 0.0.
*/
#if !defined(DUK_USE_ZERO_BUFFER_DATA)
/* Khronos/ES6 requires zeroing even when DUK_USE_ZERO_BUFFER_DATA
* is not set.
*/
DUK_ASSERT(!DUK_HBUFFER_HAS_DYNAMIC((duk_hbuffer *) h_val));
DUK_MEMZERO((void *) DUK_HBUFFER_FIXED_GET_DATA_PTR(thr->heap, h_val), (duk_size_t) byte_length);
#endif
DUK_DDD(DUK_DDDPRINT("using no copy"));
break;
}
}
return 1;
fail_arguments:
return DUK_RET_RANGE_ERROR;
}
#else /* DUK_USE_BUFFEROBJECT_SUPPORT */
DUK_INTERNAL duk_ret_t duk_bi_typedarray_constructor(duk_context *ctx) {
DUK_UNREF(ctx);
return DUK_RET_UNSUPPORTED_ERROR;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
DUK_INTERNAL duk_ret_t duk_bi_dataview_constructor(duk_context *ctx) {
duk_hbufferobject *h_bufarg;
duk_hbufferobject *h_bufobj;
duk_hbuffer *h_val;
duk_uint_t offset;
duk_uint_t length;
/* XXX: function flag to make this automatic? */
if (!duk_is_constructor_call(ctx)) {
return DUK_RET_TYPE_ERROR;
}
h_bufarg = duk__require_bufobj_value(ctx, 0);
DUK_ASSERT(h_bufarg != NULL);
duk__resolve_offset_opt_length(ctx, h_bufarg, 1, 2, &offset, &length, 1 /*throw_flag*/);
DUK_ASSERT(offset <= h_bufarg->length);
DUK_ASSERT(offset + length <= h_bufarg->length);
h_bufobj = duk_push_bufferobject_raw(ctx,
DUK_HOBJECT_FLAG_EXTENSIBLE |
DUK_HOBJECT_FLAG_BUFFEROBJECT |
DUK_HOBJECT_CLASS_AS_FLAGS(DUK_HOBJECT_CLASS_DATAVIEW),
DUK_BIDX_DATAVIEW_PROTOTYPE);
h_val = h_bufarg->buf;
if (h_val == NULL) {
return DUK_RET_TYPE_ERROR;
}
h_bufobj->buf = h_val;
DUK_HBUFFER_INCREF(thr, h_val);
h_bufobj->offset = h_bufarg->offset + offset;
h_bufobj->length = length;
DUK_ASSERT(h_bufobj->shift == 0);
DUK_ASSERT(h_bufobj->elem_type == DUK_HBUFFEROBJECT_ELEM_UINT8);
h_bufobj->is_view = 1;
/* The DataView .buffer property is ordinarily set to the argument
* which is an ArrayBuffer. We accept any duk_hbufferobject as
* an argument and .buffer will be set to the argument regardless
* of what it is. This may be a bit confusing if the argument
* is e.g. a DataView or another TypedArray view.
*
* XXX: Copy .buffer property from a DataView/TypedArray argument?
* Create a fresh ArrayBuffer for Duktape.Buffer and Node.js Buffer
* arguments? See: test-bug-dataview-buffer-prop.js.
*/
duk_dup(ctx, 0);
duk_xdef_prop_stridx(ctx, -2, DUK_STRIDX_LC_BUFFER, DUK_PROPDESC_FLAGS_NONE);
duk_compact(ctx, -1);
DUK_ASSERT_HBUFFEROBJECT_VALID(h_bufobj);
return 1;
}
#else /* DUK_USE_BUFFEROBJECT_SUPPORT */
DUK_INTERNAL duk_ret_t duk_bi_dataview_constructor(duk_context *ctx) {
DUK_UNREF(ctx);
return DUK_RET_UNSUPPORTED_ERROR;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
/*
* ArrayBuffer.isView()
*/
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
DUK_INTERNAL duk_ret_t duk_bi_arraybuffer_isview(duk_context *ctx) {
duk_hobject *h_obj;
duk_bool_t ret = 0;
h_obj = duk_get_hobject(ctx, 0);
if (h_obj != NULL && DUK_HOBJECT_IS_BUFFEROBJECT(h_obj)) {
ret = ((duk_hbufferobject *) h_obj)->is_view;
}
duk_push_boolean(ctx, ret);
return 1;
}
#else /* DUK_USE_BUFFEROBJECT_SUPPORT */
DUK_INTERNAL duk_ret_t duk_bi_arraybuffer_isview(duk_context *ctx) {
DUK_UNREF(ctx);
return DUK_RET_UNSUPPORTED_ERROR;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
/*
* Node.js Buffer: toString([encoding], [start], [end])
*/
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
DUK_INTERNAL duk_ret_t duk_bi_nodejs_buffer_tostring(duk_context *ctx) {
duk_hthread *thr;
duk_hbufferobject *h_this;
duk_int_t start_offset, end_offset;
duk_uint8_t *buf_slice;
duk_size_t slice_length;
thr = (duk_hthread *) ctx;
DUK_UNREF(thr);
h_this = duk__get_bufobj_this(ctx);
if (h_this == NULL) {
/* XXX: happens e.g. when evaluating: String(Buffer.prototype). */
duk_push_string(ctx, "[object Object]");
return 1;
}
DUK_ASSERT_HBUFFEROBJECT_VALID(h_this);
/* ignore encoding for now */
duk__clamp_startend_nonegidx_noshift(ctx, h_this, 1 /*idx_start*/, 2 /*idx_end*/, &start_offset, &end_offset);
slice_length = (duk_size_t) (end_offset - start_offset);
buf_slice = (duk_uint8_t *) duk_push_fixed_buffer(ctx, slice_length);
DUK_ASSERT(buf_slice != NULL);
if (h_this->buf == NULL) {
goto type_error;
}
if (DUK_HBUFFEROBJECT_VALID_BYTEOFFSET_EXCL(h_this, start_offset + slice_length)) {
DUK_MEMCPY((void *) buf_slice,
(const void *) (DUK_HBUFFEROBJECT_GET_SLICE_BASE(thr->heap, h_this) + start_offset),
(size_t) slice_length);
} else {
/* not covered, return all zeroes */
;
}
duk_to_string(ctx, -1);
return 1;
type_error:
return DUK_RET_TYPE_ERROR;
}
#else /* DUK_USE_BUFFEROBJECT_SUPPORT */
DUK_INTERNAL duk_ret_t duk_bi_nodejs_buffer_tostring(duk_context *ctx) {
DUK_UNREF(ctx);
return DUK_RET_UNSUPPORTED_ERROR;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
/*
* Duktape.Buffer: toString(), valueOf()
*/
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
DUK_INTERNAL duk_ret_t duk_bi_buffer_prototype_tostring_shared(duk_context *ctx) {
duk_hthread *thr;
duk_tval *tv;
duk_small_int_t to_string = duk_get_current_magic(ctx);
thr = (duk_hthread *) ctx;
DUK_UNREF(thr);
tv = duk_get_borrowed_this_tval(ctx);
DUK_ASSERT(tv != NULL);
if (DUK_TVAL_IS_BUFFER(tv)) {
duk_hbuffer *h_buf;
h_buf = DUK_TVAL_GET_BUFFER(tv);
DUK_ASSERT(h_buf != NULL);
duk_push_hbuffer(ctx, h_buf);
} else if (DUK_TVAL_IS_OBJECT(tv)) {
duk_hobject *h;
duk_hbufferobject *h_bufobj;
/* Accept any duk_hbufferobject, though we're only normally
* called for Duktape.Buffer values.
*/
h = DUK_TVAL_GET_OBJECT(tv);
DUK_ASSERT(h != NULL);
if (!DUK_HOBJECT_IS_BUFFEROBJECT(h)) {
DUK_DD(DUK_DDPRINT("toString/valueOf() called for a non-bufferobject object"));
goto type_error;
}
h_bufobj = (duk_hbufferobject *) h;
DUK_ASSERT_HBUFFEROBJECT_VALID(h_bufobj);
if (h_bufobj->buf == NULL) {
DUK_DD(DUK_DDPRINT("toString/valueOf() called for a bufferobject with NULL buf"));
goto type_error;
}
duk_push_hbuffer(ctx, h_bufobj->buf);
} else {
goto type_error;
}
if (to_string) {
duk_to_string(ctx, -1);
}
return 1;
type_error:
return DUK_RET_TYPE_ERROR;
}
#else /* DUK_USE_BUFFEROBJECT_SUPPORT */
DUK_INTERNAL duk_ret_t duk_bi_buffer_prototype_tostring_shared(duk_context *ctx) {
DUK_UNREF(ctx);
return DUK_RET_UNSUPPORTED_ERROR;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
/*
* Node.js Buffer.prototype: toJSON()
*/
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
DUK_INTERNAL duk_ret_t duk_bi_nodejs_buffer_tojson(duk_context *ctx) {
duk_hthread *thr;
duk_hbufferobject *h_this;
duk_uint8_t *buf;
duk_uint_t i;
thr = (duk_hthread *) ctx;
DUK_UNREF(thr);
h_this = duk__require_bufobj_this(ctx);
DUK_ASSERT(h_this != NULL);
if (h_this->buf == NULL || !DUK_HBUFFEROBJECT_VALID_SLICE(h_this)) {
/* Serialize uncovered backing buffer as a null; doesn't
* really matter as long we're memory safe.
*/
duk_push_null(ctx);
return 1;
}
duk_push_object(ctx);
duk_push_hstring_stridx(ctx, DUK_STRIDX_UC_BUFFER);
duk_put_prop_stridx(ctx, -2, DUK_STRIDX_TYPE);
duk_push_array(ctx);
for (i = 0; i < h_this->length; i++) {
/* XXX: regetting the pointer may be overkill - we're writing
* to a side-effect free array here.
*/
DUK_ASSERT(h_this->buf != NULL);
buf = DUK_HBUFFEROBJECT_GET_SLICE_BASE(thr->heap, h_this);
duk_push_uint(ctx, (duk_uint_t) buf[i]);
duk_put_prop_index(ctx, -2, (duk_idx_t) i);
}
duk_put_prop_stridx(ctx, -2, DUK_STRIDX_DATA);
return 1;
}
#else /* DUK_USE_BUFFEROBJECT_SUPPORT */
DUK_INTERNAL duk_ret_t duk_bi_nodejs_buffer_tojson(duk_context *ctx) {
DUK_UNREF(ctx);
return DUK_RET_UNSUPPORTED_ERROR;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
/*
* Node.js Buffer.prototype.equals()
* Node.js Buffer.prototype.compare()
* Node.js Buffer.compare()
*/
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
DUK_INTERNAL duk_ret_t duk_bi_buffer_compare_shared(duk_context *ctx) {
duk_hthread *thr;
duk_small_uint_t magic;
duk_hbufferobject *h_bufarg1;
duk_hbufferobject *h_bufarg2;
duk_small_int_t comp_res;
thr = (duk_hthread *) ctx;
DUK_UNREF(thr);
magic = duk_get_current_magic(ctx);
if (magic & 0x02) {
/* Static call style. */
h_bufarg1 = duk__require_bufobj_value(ctx, 0);
h_bufarg2 = duk__require_bufobj_value(ctx, 1);
} else {
h_bufarg1 = duk__require_bufobj_this(ctx);
h_bufarg2 = duk__require_bufobj_value(ctx, 0);
}
DUK_ASSERT(h_bufarg1 != NULL);
DUK_ASSERT(h_bufarg2 != NULL);
/* We want to compare the slice/view areas of the arguments.
* If either slice/view is invalid (underlying buffer is shorter)
* ensure equals() is false, but otherwise the only thing that
* matters is to be memory safe.
*/
if (DUK_HBUFFEROBJECT_VALID_SLICE(h_bufarg1) &&
DUK_HBUFFEROBJECT_VALID_SLICE(h_bufarg2)) {
comp_res = duk_js_data_compare((const duk_uint8_t *) DUK_HBUFFER_GET_DATA_PTR(thr->heap, h_bufarg1->buf) + h_bufarg1->offset,
(const duk_uint8_t *) DUK_HBUFFER_GET_DATA_PTR(thr->heap, h_bufarg2->buf) + h_bufarg2->offset,
(duk_size_t) h_bufarg1->length,
(duk_size_t) h_bufarg2->length);
} else {
comp_res = -1; /* either nonzero value is ok */
}
if (magic & 0x01) {
/* compare: similar to string comparison but for buffer data. */
duk_push_int(ctx, comp_res);
} else {
/* equals */
duk_push_boolean(ctx, (comp_res == 0));
}
return 1;
}
#else /* DUK_USE_BUFFEROBJECT_SUPPORT */
DUK_INTERNAL duk_ret_t duk_bi_buffer_compare_shared(duk_context *ctx) {
DUK_UNREF(ctx);
return DUK_RET_UNSUPPORTED_ERROR;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
/*
* Node.js Buffer.prototype.fill()
*/
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
DUK_INTERNAL duk_ret_t duk_bi_nodejs_buffer_fill(duk_context *ctx) {
duk_hthread *thr;
duk_hbufferobject *h_this;
const duk_uint8_t *fill_str_ptr;
duk_size_t fill_str_len;
duk_uint8_t fill_value;
duk_int_t fill_offset;
duk_int_t fill_end;
duk_size_t fill_length;
duk_uint8_t *p;
thr = (duk_hthread *) ctx;
DUK_UNREF(thr);
h_this = duk__require_bufobj_this(ctx);
DUK_ASSERT(h_this != NULL);
if (h_this->buf == NULL) {
return DUK_RET_TYPE_ERROR;
}
/* [ value offset end ] */
if (duk_is_string(ctx, 0)) {
fill_str_ptr = (const duk_uint8_t *) duk_get_lstring(ctx, 0, &fill_str_len);
DUK_ASSERT(fill_str_ptr != NULL);
} else {
fill_value = (duk_uint8_t) duk_to_uint32(ctx, 0);
fill_str_ptr = (const duk_uint8_t *) &fill_value;
fill_str_len = 1;
}
/* Fill offset handling is more lenient than in Node.js. */
duk__clamp_startend_nonegidx_noshift(ctx, h_this, 1 /*idx_start*/, 2 /*idx_end*/, &fill_offset, &fill_end);
DUK_DDD(DUK_DDDPRINT("fill: fill_value=%02x, fill_offset=%ld, fill_end=%ld, view length=%ld",
(unsigned int) fill_value, (long) fill_offset, (long) fill_end, (long) h_this->length));
DUK_ASSERT(fill_end - fill_offset >= 0);
DUK_ASSERT(h_this->buf != NULL);
p = (DUK_HBUFFEROBJECT_GET_SLICE_BASE(thr->heap, h_this) + fill_offset);
fill_length = (duk_size_t) (fill_end - fill_offset);
if (fill_str_len == 1) {
/* Handle single character fills as memset() even when
* the fill data comes from a one-char argument.
*/
DUK_MEMSET((void *) p, (int) fill_str_ptr[0], (size_t) fill_length);
} else if (fill_str_len > 1) {
duk_size_t i, n, t;
for (i = 0, n = (fill_end - fill_offset), t = 0; i < n; i++) {
p[i] = fill_str_ptr[t++];
if (t >= fill_str_len) {
t = 0;
}
}
} else {
DUK_DDD(DUK_DDDPRINT("zero size fill pattern, ignore silently"));
}
/* Return the Buffer to allow chaining: b.fill(0x11).fill(0x22, 3, 5).toString() */
duk_push_this(ctx);
return 1;
}
#else /* DUK_USE_BUFFEROBJECT_SUPPORT */
DUK_INTERNAL duk_ret_t duk_bi_nodejs_buffer_fill(duk_context *ctx) {
DUK_UNREF(ctx);
return DUK_RET_UNSUPPORTED_ERROR;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
/*
* Node.js Buffer.prototype.write(string, [offset], [length], [encoding])
*/
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
DUK_INTERNAL duk_ret_t duk_bi_nodejs_buffer_write(duk_context *ctx) {
duk_hthread *thr;
duk_hbufferobject *h_this;
duk_uint_t offset;
duk_uint_t length;
const duk_uint8_t *str_data;
duk_size_t str_len;
thr = (duk_hthread *) ctx;
DUK_UNREF(thr);
h_this = duk__require_bufobj_this(ctx);
DUK_ASSERT(h_this != NULL);
/* Argument must be a string, e.g. a buffer is not allowed. */
str_data = (const duk_uint8_t *) duk_require_lstring(ctx, 0, &str_len);
duk__resolve_offset_opt_length(ctx, h_this, 1, 2, &offset, &length, 0 /*throw_flag*/);
DUK_ASSERT(offset <= h_this->length);
DUK_ASSERT(offset + length <= h_this->length);
/* XXX: encoding is ignored now. */
if (length > str_len) {
length = (duk_uint_t) str_len;
}
if (DUK_HBUFFEROBJECT_VALID_SLICE(h_this)) {
/* Cannot overlap. */
DUK_MEMCPY((void *) (DUK_HBUFFEROBJECT_GET_SLICE_BASE(thr->heap, h_this) + offset),
(const void *) str_data,
(size_t) length);
} else {
DUK_DDD(DUK_DDDPRINT("write() target buffer is not covered, silent ignore"));
}
duk_push_uint(ctx, length);
return 1;
}
#else /* DUK_USE_BUFFEROBJECT_SUPPORT */
DUK_INTERNAL duk_ret_t duk_bi_nodejs_buffer_write(duk_context *ctx) {
DUK_UNREF(ctx);
return DUK_RET_UNSUPPORTED_ERROR;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
/*
* Node.js Buffer.prototype.copy()
*/
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
DUK_INTERNAL duk_ret_t duk_bi_nodejs_buffer_copy(duk_context *ctx) {
duk_hthread *thr;
duk_hbufferobject *h_this;
duk_hbufferobject *h_bufarg;
duk_int_t source_length;
duk_int_t target_length;
duk_int_t target_start, source_start, source_end;
duk_uint_t target_ustart, source_ustart, source_uend;
duk_uint_t copy_size = 0;
/* [ targetBuffer targetStart sourceStart sourceEnd ] */
thr = (duk_hthread *) ctx;
DUK_UNREF(thr);
h_this = duk__require_bufobj_this(ctx);
h_bufarg = duk__require_bufobj_value(ctx, 0);
DUK_ASSERT(h_this != NULL);
DUK_ASSERT(h_bufarg != NULL);
source_length = (duk_int_t) h_this->length;
target_length = (duk_int_t) h_bufarg->length;
target_start = duk_to_int(ctx, 1);
source_start = duk_to_int(ctx, 2);
if (duk_is_undefined(ctx, 3)) {
source_end = source_length;
} else {
source_end = duk_to_int(ctx, 3);
}
DUK_DDD(DUK_DDDPRINT("checking copy args: target_start=%ld, target_length=%ld, "
"source_start=%ld, source_end=%ld, source_length=%ld",
(long) target_start, (long) h_bufarg->length,
(long) source_start, (long) source_end, (long) source_length));
/* This behavior mostly mimics Node.js now. */
if (source_start < 0 || source_end < 0 || target_start < 0) {
/* Negative offsets cause a RangeError. */
goto fail_bounds;
}
source_ustart = (duk_uint_t) source_start;
source_uend = (duk_uint_t) source_end;
target_ustart = (duk_uint_t) target_start;
if (source_ustart >= source_uend || /* crossed offsets or zero size */
source_ustart >= (duk_uint_t) source_length || /* source out-of-bounds (but positive) */
target_ustart >= (duk_uint_t) target_length) { /* target out-of-bounds (but positive) */
goto silent_ignore;
}
if (source_uend >= (duk_uint_t) source_length) {
/* Source end clamped silently to available length. */
source_uend = source_length;
}
copy_size = source_uend - source_ustart;
if (target_ustart + copy_size > (duk_uint_t) target_length) {
/* Clamp to target's end if too long.
*
* NOTE: there's no overflow possibility in the comparison;
* both target_ustart and copy_size are >= 0 and based on
* values in duk_int_t range. Adding them as duk_uint_t
* values is then guaranteed not to overflow.
*/
DUK_ASSERT(target_ustart + copy_size >= target_ustart); /* no overflow */
DUK_ASSERT(target_ustart + copy_size >= copy_size); /* no overflow */
copy_size = (duk_uint_t) target_length - target_ustart;
}
DUK_DDD(DUK_DDDPRINT("making copy: target_ustart=%lu source_ustart=%lu copy_size=%lu",
(unsigned long) target_ustart, (unsigned long) source_ustart,
(unsigned long) copy_size));
DUK_ASSERT(copy_size >= 1);
DUK_ASSERT(source_ustart <= (duk_uint_t) source_length);
DUK_ASSERT(source_ustart + copy_size <= (duk_uint_t) source_length);
DUK_ASSERT(target_ustart <= (duk_uint_t) target_length);
DUK_ASSERT(target_ustart + copy_size <= (duk_uint_t) target_length);
/* Ensure copy is covered by underlying buffers. */
DUK_ASSERT(h_bufarg->buf != NULL); /* length check */
DUK_ASSERT(h_this->buf != NULL); /* length check */
if (DUK_HBUFFEROBJECT_VALID_BYTEOFFSET_EXCL(h_bufarg, target_ustart + copy_size) &&
DUK_HBUFFEROBJECT_VALID_BYTEOFFSET_EXCL(h_this, source_ustart + copy_size)) {
/* Must use memmove() because copy area may overlap (source and target
* buffer may be the same, or from different slices.
*/
DUK_MEMMOVE((void *) (DUK_HBUFFEROBJECT_GET_SLICE_BASE(thr->heap, h_bufarg) + target_ustart),
(const void *) (DUK_HBUFFEROBJECT_GET_SLICE_BASE(thr->heap, h_this) + source_ustart),
(size_t) copy_size);
} else {
DUK_DDD(DUK_DDDPRINT("buffer copy not covered by underlying buffer(s), ignoring"));
}
silent_ignore:
/* Return value is like write(), number of bytes written.
* The return value matters because of code like:
* "off += buf.copy(...)".
*/
duk_push_uint(ctx, copy_size);
return 1;
fail_bounds:
return DUK_RET_RANGE_ERROR;
}
#else /* DUK_USE_BUFFEROBJECT_SUPPORT */
DUK_INTERNAL duk_ret_t duk_bi_nodejs_buffer_copy(duk_context *ctx) {
DUK_UNREF(ctx);
return DUK_RET_UNSUPPORTED_ERROR;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
/*
* TypedArray.prototype.set()
*
* TypedArray set() is pretty interesting to implement because:
*
* - The source argument may be a plain array or a typedarray. If the
* source is a TypedArray, values are decoded and re-encoded into the
* target (not as a plain byte copy). This may happen even when the
* element byte size is the same, e.g. integer values may be re-encoded
* into floats.
*
* - Source and target may refer to the same underlying buffer, so that
* the set() operation may overlap. The specification requires that this
* must work as if a copy was made before the operation. Note that this
* is NOT a simple memmove() situation because the source and target
* byte sizes may be different -- e.g. a 4-byte source (Int8Array) may
* expand to a 16-byte target (Uint32Array) so that the target overlaps
* the source both from beginning and the end (unlike in typical memmove).
*
* - Even if 'buf' pointers of the source and target differ, there's no
* guarantee that their memory areas don't overlap. This may be the
* case with external buffers.
*
* Even so, it is nice to optimize for the common case:
*
* - Source and target separate buffers or non-overlapping.
*
* - Source and target have a compatible type so that a plain byte copy
* is possible. Note that while e.g. uint8 and int8 are compatible
* (coercion one way or another doesn't change the byte representation),
* e.g. int8 and uint8clamped are NOT compatible when writing int8
* values into uint8clamped typedarray (-1 would clamp to 0 for instance).
*
* See test-bi-typedarray-proto-set.js.
*/
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
DUK_INTERNAL duk_ret_t duk_bi_typedarray_set(duk_context *ctx) {
duk_hthread *thr;
duk_hbufferobject *h_this;
duk_hobject *h_obj;
duk_uarridx_t i, n;
duk_int_t offset_signed;
duk_uint_t offset_elems;
duk_uint_t offset_bytes;
thr = (duk_hthread *) ctx;
DUK_UNREF(thr);
h_this = duk__require_bufobj_this(ctx);
DUK_ASSERT(h_this != NULL);
DUK_ASSERT_HBUFFEROBJECT_VALID(h_this);
if (h_this->buf == NULL) {
DUK_DDD(DUK_DDDPRINT("source neutered, skip copy"));
return 0;
}
h_obj = duk_require_hobject(ctx, 0);
DUK_ASSERT(h_obj != NULL);
/* XXX: V8 throws a TypeError for negative values. Would it
* be more useful to interpret negative offsets here from the
* end of the buffer too?
*/
offset_signed = duk_to_int(ctx, 1);
if (offset_signed < 0) {
return DUK_RET_TYPE_ERROR;
}
offset_elems = (duk_uint_t) offset_signed;
offset_bytes = offset_elems << h_this->shift;
if ((offset_bytes >> h_this->shift) != offset_elems) {
/* Byte length would overflow. */
/* XXX: easier check with less code? */
return DUK_RET_RANGE_ERROR;
}
if (offset_bytes > h_this->length) {
/* Equality may be OK but >length not. Checking
* this explicitly avoids some overflow cases
* below.
*/
return DUK_RET_RANGE_ERROR;
}
DUK_ASSERT(offset_bytes <= h_this->length);
/* Fast path: source is a TypedArray (or any bufferobject). */
if (DUK_HOBJECT_IS_BUFFEROBJECT(h_obj)) {
duk_hbufferobject *h_bufarg;
duk_uint16_t comp_mask;
duk_small_int_t no_overlap = 0;
duk_uint_t src_length;
duk_uint_t dst_length;
duk_uint_t dst_length_elems;
duk_uint8_t *p_src_base;
duk_uint8_t *p_src_end;
duk_uint8_t *p_src;
duk_uint8_t *p_dst_base;
duk_uint8_t *p_dst;
duk_small_uint_t src_elem_size;
duk_small_uint_t dst_elem_size;
h_bufarg = (duk_hbufferobject *) h_obj;
DUK_ASSERT_HBUFFEROBJECT_VALID(h_bufarg);
if (h_bufarg->buf == NULL) {
DUK_DDD(DUK_DDDPRINT("target neutered, skip copy"));
return 0;
}
/* Nominal size check. */
src_length = h_bufarg->length; /* bytes in source */
dst_length_elems = (src_length >> h_bufarg->shift); /* elems in source and dest */
dst_length = dst_length_elems << h_this->shift; /* bytes in dest */
if ((dst_length >> h_this->shift) != dst_length_elems) {
/* Byte length would overflow. */
/* XXX: easier check with less code? */
return DUK_RET_RANGE_ERROR;
}
DUK_DDD(DUK_DDDPRINT("nominal size check: src_length=%ld, dst_length=%ld",
(long) src_length, (long) dst_length));
DUK_ASSERT(offset_bytes <= h_this->length);
if (dst_length > h_this->length - offset_bytes) {
/* Overflow not an issue because subtraction is used on the right
* side and guaranteed to be >= 0.
*/
DUK_DDD(DUK_DDDPRINT("copy exceeds target buffer nominal length"));
return DUK_RET_RANGE_ERROR;
}
if (!DUK_HBUFFEROBJECT_VALID_BYTEOFFSET_EXCL(h_this, offset_bytes + dst_length)) {
DUK_DDD(DUK_DDDPRINT("copy not covered by underlying target buffer, ignore"));
return 0;
}
p_src_base = DUK_HBUFFEROBJECT_GET_SLICE_BASE(thr->heap, h_bufarg);
p_dst_base = DUK_HBUFFEROBJECT_GET_SLICE_BASE(thr->heap, h_this) + offset_bytes;
/* Check actual underlying buffers for validity and that they
* cover the copy. No side effects are allowed after the check
* so that the validity status doesn't change.
*/
if (!DUK_HBUFFEROBJECT_VALID_SLICE(h_this) ||
!DUK_HBUFFEROBJECT_VALID_SLICE(h_bufarg)) {
/* The condition could be more narrow and check for the
* copy area only, but there's no need for fine grained
* behavior when the underlying buffer is misconfigured.
*/
DUK_DDD(DUK_DDDPRINT("source and/or target not covered by underlying buffer, skip copy"));
return 0;
}
/* We want to do a straight memory copy if possible: this is
* an important operation because .set() is the TypedArray
* way to copy chunks of memory. However, because set()
* conceptually works in terms of elements, not all views are
* compatible with direct byte copying.
*
* If we do manage a direct copy, the "overlap issue" handled
* below can just be solved using memmove() because the source
* and destination element sizes are necessarily equal.
*/
DUK_ASSERT(h_this->elem_type < sizeof(duk__buffer_elemtype_copy_compatible) / sizeof(duk_uint16_t));
comp_mask = duk__buffer_elemtype_copy_compatible[h_this->elem_type];
if (comp_mask & (1 << h_bufarg->elem_type)) {
DUK_ASSERT(src_length == dst_length);
DUK_DDD(DUK_DDDPRINT("fast path: able to use memmove() because views are compatible"));
DUK_MEMMOVE((void *) p_dst_base, (const void *) p_src_base, (size_t) dst_length);
return 0;
}
DUK_DDD(DUK_DDDPRINT("fast path: views are not compatible with a byte copy, copy by item"));
/* We want to avoid making a copy to process set() but that's
* not always possible: the source and the target may overlap
* and because element sizes are different, the overlap cannot
* always be handled with a memmove() or choosing the copy
* direction in a certain way. For example, if source type is
* uint8 and target type is uint32, the target area may exceed
* the source area from both ends!
*
* Note that because external buffers may point to the same
* memory areas, we must ultimately make this check using
* pointers.
*
* NOTE: careful with side effects: any side effect may cause
* a buffer resize (or external buffer pointer/length update)!
*/
DUK_DDD(DUK_DDDPRINT("overlap check: p_src_base=%p, src_length=%ld, "
"p_dst_base=%p, dst_length=%ld",
(void *) p_src_base, (long) src_length,
(void *) p_dst_base, (long) dst_length));
if (p_src_base >= p_dst_base + dst_length || /* source starts after dest ends */
p_src_base + src_length <= p_dst_base) { /* source ends before dest starts */
no_overlap = 1;
}
if (!no_overlap) {
/* There's overlap: the desired end result is that
* conceptually a copy is made to avoid "trampling"
* of source data by destination writes. We make
* an actual temporary copy to handle this case.
*/
duk_uint8_t *p_src_copy;
DUK_DDD(DUK_DDDPRINT("there is overlap, make a copy of the source"));
p_src_copy = (duk_uint8_t *) duk_push_fixed_buffer(ctx, src_length);
DUK_ASSERT(p_src_copy != NULL);
DUK_MEMCPY((void *) p_src_copy, (const void *) p_src_base, (size_t) src_length);
p_src_base = p_src_copy; /* use p_src_base from now on */
}
/* Value stack intentionally mixed size here. */
DUK_DDD(DUK_DDDPRINT("after overlap check: p_src_base=%p, src_length=%ld, "
"p_dst_base=%p, dst_length=%ld, valstack top=%ld",
(void *) p_src_base, (long) src_length,
(void *) p_dst_base, (long) dst_length,
(long) duk_get_top(ctx)));
/* Ready to make the copy. We must proceed element by element
* and must avoid any side effects that might cause the buffer
* validity check above to become invalid.
*
* Although we work through the value stack here, only plain
* numbers are handled which should be side effect safe.
*/
src_elem_size = 1 << h_bufarg->shift;
dst_elem_size = 1 << h_this->shift;
p_src = p_src_base;
p_dst = p_dst_base;
p_src_end = p_src_base + src_length;
while (p_src != p_src_end) {
DUK_DDD(DUK_DDDPRINT("fast path per element copy loop: "
"p_src=%p, p_src_end=%p, p_dst=%p",
(void *) p_src, (void *) p_src_end, (void *) p_dst));
/* A validated read() is always a number, so it's write coercion
* is always side effect free an won't invalidate pointers etc.
*/
duk_hbufferobject_push_validated_read(ctx, h_bufarg, p_src, src_elem_size);
duk_hbufferobject_validated_write(ctx, h_this, p_dst, dst_elem_size);
duk_pop(ctx);
p_src += src_elem_size;
p_dst += dst_elem_size;
}
return 0;
} else {
/* Slow path: quite slow, but we save space by using the property code
* to write coerce target values. We don't need to worry about overlap
* here because the source is not a TypedArray.
*
* We could use the bufferobject write coercion helper but since the
* property read may have arbitrary side effects, full validity checks
* would be needed for every element anyway.
*/
n = (duk_uarridx_t) duk_get_length(ctx, 0);
DUK_ASSERT(offset_bytes <= h_this->length);
if ((n << h_this->shift) > h_this->length - offset_bytes) {
/* Overflow not an issue because subtraction is used on the right
* side and guaranteed to be >= 0.
*/
DUK_DDD(DUK_DDDPRINT("copy exceeds target buffer nominal length"));
return DUK_RET_RANGE_ERROR;
}
/* There's no need to check for buffer validity status for the
* target here: the property access code will do that for each
* element. Moreover, if we did check the validity here, side
* effects from reading the source argument might invalidate
* the results anyway.
*/
DUK_ASSERT_TOP(ctx, 2);
duk_push_this(ctx);
for (i = 0; i < n; i++) {
duk_get_prop_index(ctx, 0, i);
duk_put_prop_index(ctx, 2, offset_elems + i);
}
}
return 0;
}
#else /* DUK_USE_BUFFEROBJECT_SUPPORT */
DUK_INTERNAL duk_ret_t duk_bi_typedarray_set(duk_context *ctx) {
DUK_UNREF(ctx);
return DUK_RET_UNSUPPORTED_ERROR;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
/*
* Node.js Buffer.prototype.slice([start], [end])
* ArrayBuffer.prototype.slice(begin, [end])
* TypedArray.prototype.slice(begin, [end])
*
* The API calls are almost identical; negative indices are counted from end
* of buffer, and final indices are clamped (allowing crossed indices). Main
* differences:
*
* - Copy/view behavior; Node.js .slice() and TypedArray .subarray() create
* views, ArrayBuffer .slice() creates a copy
*
* - Resulting object has a different class and prototype depending on the
* call (or 'this' argument)
*
* - TypedArray .subarray() arguments are element indices, not byte offsets
*/
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
DUK_INTERNAL duk_ret_t duk_bi_buffer_slice_shared(duk_context *ctx) {
duk_hthread *thr;
duk_small_int_t magic;
duk_small_uint_t res_class_num;
duk_hobject *res_proto;
duk_hbufferobject *h_this;
duk_hbufferobject *h_bufobj;
duk_hbuffer *h_val;
duk_int_t start_offset, end_offset;
duk_uint_t slice_length;
thr = (duk_hthread *) ctx;
DUK_UNREF(thr);
/* [ start end ] */
magic = duk_get_current_magic(ctx);
h_this = duk__require_bufobj_this(ctx);
/* Slice offsets are element (not byte) offsets, which only matters
* for TypedArray views, Node.js Buffer and ArrayBuffer have shift
* zero so byte and element offsets are the same. Negative indices
* are counted from end of slice, crossed indices are allowed (and
* result in zero length result), and final values are clamped
* against the current slice. There's intentionally no check
* against the underlying buffer here.
*/
duk__clamp_startend_negidx_shifted(ctx, h_this, 0 /*idx_start*/, 1 /*idx_end*/, &start_offset, &end_offset);
DUK_ASSERT(end_offset >= start_offset);
slice_length = (duk_uint_t) (end_offset - start_offset);
/* The resulting buffer object gets the same class and prototype as
* the buffer in 'this', e.g. if the input is a Node.js Buffer the
* result is a Node.js Buffer; if the input is a Float32Array, the
* result is a Float32Array.
*
* For the class number this seems correct. The internal prototype
* is not so clear: if 'this' is a bufferobject with a non-standard
* prototype object, that value gets copied over into the result
* (instead of using the standard prototype for that object type).
*/
res_class_num = DUK_HOBJECT_GET_CLASS_NUMBER((duk_hobject *) h_this);
h_bufobj = duk_push_bufferobject_raw(ctx,
DUK_HOBJECT_FLAG_EXTENSIBLE |
DUK_HOBJECT_FLAG_BUFFEROBJECT |
DUK_HOBJECT_CLASS_AS_FLAGS(res_class_num),
DUK_BIDX_OBJECT_PROTOTYPE); /* replaced */
DUK_ASSERT(h_bufobj != NULL);
res_proto = DUK_HOBJECT_GET_PROTOTYPE(thr->heap, (duk_hobject *) h_this); /* may be NULL */
DUK_HOBJECT_SET_PROTOTYPE_UPDREF(thr, (duk_hobject *) h_bufobj, res_proto);
h_bufobj->length = slice_length;
h_bufobj->shift = h_this->shift; /* inherit */
h_bufobj->elem_type = h_this->elem_type; /* inherit */
h_bufobj->is_view = magic & 0x01;
DUK_ASSERT(h_bufobj->is_view == 0 || h_bufobj->is_view == 1);
h_val = h_this->buf;
if (h_val == NULL) {
return DUK_RET_TYPE_ERROR;
}
if (magic & 0x02) {
/* non-zero: make copy */
duk_uint8_t *p_copy;
duk_size_t copy_length;
p_copy = (duk_uint8_t *) duk_push_fixed_buffer(ctx, (duk_size_t) slice_length);
DUK_ASSERT(p_copy != NULL);
/* Copy slice, respecting underlying buffer limits; remainder
* is left as zero.
*/
copy_length = DUK_HBUFFEROBJECT_CLAMP_BYTELENGTH(h_this, slice_length);
DUK_MEMCPY((void *) p_copy,
(const void *) (DUK_HBUFFEROBJECT_GET_SLICE_BASE(thr->heap, h_this) + start_offset),
copy_length);
h_val = duk_get_hbuffer(ctx, -1);
DUK_ASSERT(h_val != NULL);
h_bufobj->buf = h_val;
DUK_HBUFFER_INCREF(thr, h_val);
DUK_ASSERT(h_bufobj->offset == 0);
duk_pop(ctx); /* reachable so pop OK */
} else {
h_bufobj->buf = h_val;
DUK_HBUFFER_INCREF(thr, h_val);
h_bufobj->offset = (duk_uint_t) (h_this->offset + start_offset);
/* Copy the .buffer property, needed for TypedArray.prototype.subarray().
*
* XXX: limit copy only for TypedArray classes specifically?
*/
duk_push_this(ctx);
if (duk_get_prop_stridx(ctx, -1, DUK_STRIDX_LC_BUFFER)) {
duk_xdef_prop_stridx(ctx, -3, DUK_STRIDX_LC_BUFFER, DUK_PROPDESC_FLAGS_NONE);
duk_pop(ctx);
} else {
duk_pop_2(ctx);
}
}
/* unbalanced stack on purpose */
DUK_ASSERT_HBUFFEROBJECT_VALID(h_bufobj);
return 1;
}
#else /* DUK_USE_BUFFEROBJECT_SUPPORT */
DUK_INTERNAL duk_ret_t duk_bi_buffer_slice_shared(duk_context *ctx) {
DUK_UNREF(ctx);
return DUK_RET_UNSUPPORTED_ERROR;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
/*
* Node.js Buffer.isEncoding()
*/
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
DUK_INTERNAL duk_ret_t duk_bi_nodejs_buffer_is_encoding(duk_context *ctx) {
const char *encoding;
/* only accept lowercase 'utf8' now. */
encoding = duk_to_string(ctx, 0);
DUK_ASSERT(duk_is_string(ctx, 0)); /* guaranteed by duk_to_string() */
duk_push_boolean(ctx, DUK_STRCMP(encoding, "utf8") == 0);
return 1;
}
#else /* DUK_USE_BUFFEROBJECT_SUPPORT */
DUK_INTERNAL duk_ret_t duk_bi_nodejs_buffer_is_encoding(duk_context *ctx) {
DUK_UNREF(ctx);
return DUK_RET_UNSUPPORTED_ERROR;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
/*
* Node.js Buffer.isBuffer()
*/
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
DUK_INTERNAL duk_ret_t duk_bi_nodejs_buffer_is_buffer(duk_context *ctx) {
duk_hthread *thr;
duk_tval *tv;
duk_hobject *h;
duk_hobject *h_proto;
duk_bool_t ret = 0;
thr = (duk_hthread *) ctx;
DUK_ASSERT(duk_get_top(ctx) >= 1); /* nargs */
tv = duk_get_tval(ctx, 0);
DUK_ASSERT(tv != NULL);
if (DUK_TVAL_IS_OBJECT(tv)) {
h = DUK_TVAL_GET_OBJECT(tv);
DUK_ASSERT(h != NULL);
h_proto = thr->builtins[DUK_BIDX_NODEJS_BUFFER_PROTOTYPE];
DUK_ASSERT(h_proto != NULL);
h = DUK_HOBJECT_GET_PROTOTYPE(thr->heap, h);
if (h) {
ret = duk_hobject_prototype_chain_contains(thr, h, h_proto, 0 /*ignore_loop*/);
}
}
duk_push_boolean(ctx, ret);
return 1;
}
#else /* DUK_USE_BUFFEROBJECT_SUPPORT */
DUK_INTERNAL duk_ret_t duk_bi_nodejs_buffer_is_buffer(duk_context *ctx) {
DUK_UNREF(ctx);
return DUK_RET_UNSUPPORTED_ERROR;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
/*
* Node.js Buffer.byteLength()
*/
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
DUK_INTERNAL duk_ret_t duk_bi_nodejs_buffer_byte_length(duk_context *ctx) {
const char *str;
duk_size_t len;
/* At the moment Buffer(<str>) will just use the string bytes as
* is (ignoring encoding), so we return the string length here
* unconditionally.
*/
str = duk_to_lstring(ctx, 0, &len);
DUK_UNREF(str);
duk_push_size_t(ctx, len);
return 1;
}
#else /* DUK_USE_BUFFEROBJECT_SUPPORT */
DUK_INTERNAL duk_ret_t duk_bi_nodejs_buffer_byte_length(duk_context *ctx) {
DUK_UNREF(ctx);
return DUK_RET_UNSUPPORTED_ERROR;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
/*
* Node.js Buffer.concat()
*/
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
DUK_INTERNAL duk_ret_t duk_bi_nodejs_buffer_concat(duk_context *ctx) {
duk_hthread *thr;
duk_hobject *h_arg;
duk_int_t total_length = 0;
duk_hbufferobject *h_bufobj;
duk_hbufferobject *h_bufres;
duk_hbuffer *h_val;
duk_uint_t i, n;
duk_uint8_t *p;
duk_size_t space_left;
duk_size_t copy_size;
thr = (duk_hthread *) ctx;
DUK_UNREF(thr);
/* Node.js accepts only actual Arrays. */
h_arg = duk_require_hobject(ctx, 0);
if (DUK_HOBJECT_GET_CLASS_NUMBER(h_arg) != DUK_HOBJECT_CLASS_ARRAY) {
return DUK_RET_TYPE_ERROR;
}
/* Compute result length and validate argument buffers. */
n = (duk_uint_t) duk_get_length(ctx, 0);
for (i = 0; i < n; i++) {
/* Neutered checks not necessary here: neutered buffers have
* zero 'length' so we'll effectively skip them.
*/
DUK_ASSERT_TOP(ctx, 2); /* [ array totalLength ] */
duk_get_prop_index(ctx, 0, (duk_uarridx_t) i); /* -> [ array totalLength buf ] */
h_bufobj = duk__require_bufobj_value(ctx, 2);
DUK_ASSERT(h_bufobj != NULL);
total_length += h_bufobj->length;
duk_pop(ctx);
}
if (n == 1) {
/* For the case n==1 Node.js doesn't seem to type check
* the sole member but we do it before returning it.
* For this case only the original buffer object is
* returned (not a copy).
*/
duk_get_prop_index(ctx, 0, 0);
return 1;
}
/* User totalLength overrides a computed length, but we'll check
* every copy in the copy loop. Note that duk_to_uint() can
* technically have arbitrary side effects so we need to recheck
* the buffers in the copy loop.
*/
if (!duk_is_undefined(ctx, 1) && n > 0) {
/* For n == 0, Node.js ignores totalLength argument and
* returns a zero length buffer.
*/
total_length = duk_to_int(ctx, 1);
}
if (total_length < 0) {
return DUK_RET_RANGE_ERROR;
}
h_bufres = duk_push_bufferobject_raw(ctx,
DUK_HOBJECT_FLAG_EXTENSIBLE |
DUK_HOBJECT_FLAG_BUFFEROBJECT |
DUK_HOBJECT_CLASS_AS_FLAGS(DUK_HOBJECT_CLASS_BUFFER),
DUK_BIDX_NODEJS_BUFFER_PROTOTYPE);
DUK_ASSERT(h_bufres != NULL);
p = (duk_uint8_t *) duk_push_fixed_buffer(ctx, total_length);
DUK_ASSERT(p != NULL);
space_left = total_length;
for (i = 0; i < n; i++) {
DUK_ASSERT_TOP(ctx, 4); /* [ array totalLength bufres buf ] */
duk_get_prop_index(ctx, 0, (duk_uarridx_t) i);
h_bufobj = duk__require_bufobj_value(ctx, 4);
DUK_ASSERT(h_bufobj != NULL);
copy_size = h_bufobj->length;
if (copy_size > space_left) {
copy_size = space_left;
}
if (h_bufobj->buf != NULL &&
DUK_HBUFFEROBJECT_VALID_SLICE(h_bufobj)) {
DUK_MEMCPY((void *) p,
(const void *) DUK_HBUFFEROBJECT_GET_SLICE_BASE(thr->heap, h_bufobj),
copy_size);
} else {
/* Just skip, leaving zeroes in the result. */
;
}
p += copy_size;
space_left -= copy_size;
duk_pop(ctx);
}
h_val = duk_get_hbuffer(ctx, -1);
DUK_ASSERT(h_val != NULL);
duk__set_bufobj_buffer(ctx, h_bufres, h_val);
DUK_ASSERT_HBUFFEROBJECT_VALID(h_bufres);
duk_pop(ctx); /* pop plain buffer, now reachable through h_bufres */
return 1; /* return h_bufres */
}
#else /* DUK_USE_BUFFEROBJECT_SUPPORT */
DUK_INTERNAL duk_ret_t duk_bi_nodejs_buffer_concat(duk_context *ctx) {
DUK_UNREF(ctx);
return DUK_RET_UNSUPPORTED_ERROR;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
/*
* Shared readfield and writefield methods
*
* The readfield/writefield methods need support for endianness and field
* types. All offsets are byte based so no offset shifting is needed.
*/
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
/* Format of magic, bits:
* 0...1: field type; 0=uint8, 1=uint16, 2=uint32, 3=float, 4=double, 5=unused, 6=unused, 7=unused
* 3: endianness: 0=little, 1=big
* 4: signed: 1=yes, 0=no
* 5: typedarray: 1=yes, 0=no
*/
#define DUK__FLD_8BIT 0
#define DUK__FLD_16BIT 1
#define DUK__FLD_32BIT 2
#define DUK__FLD_FLOAT 3
#define DUK__FLD_DOUBLE 4
#define DUK__FLD_VARINT 5
#define DUK__FLD_BIGENDIAN (1 << 3)
#define DUK__FLD_SIGNED (1 << 4)
#define DUK__FLD_TYPEDARRAY (1 << 5)
/* XXX: split into separate functions for each field type? */
DUK_INTERNAL duk_ret_t duk_bi_buffer_readfield(duk_context *ctx) {
duk_hthread *thr;
duk_small_int_t magic = (duk_small_int_t) duk_get_current_magic(ctx);
duk_small_int_t magic_ftype;
duk_small_int_t magic_bigendian;
duk_small_int_t magic_signed;
duk_small_int_t magic_typedarray;
duk_small_int_t endswap;
duk_hbufferobject *h_this;
duk_bool_t no_assert;
duk_int_t offset_signed;
duk_uint_t offset;
duk_uint_t buffer_length;
duk_uint_t check_length;
duk_uint8_t *buf;
duk_double_union du;
thr = (duk_hthread *) ctx;
DUK_UNREF(thr);
magic_ftype = magic & 0x0007;
magic_bigendian = magic & 0x0008;
magic_signed = magic & 0x0010;
magic_typedarray = magic & 0x0020;
h_this = duk__require_bufobj_this(ctx);
DUK_ASSERT(h_this != NULL);
buffer_length = h_this->length;
/* [ offset noAssert ], when ftype != DUK__FLD_VARINT */
/* [ offset fieldByteLength noAssert ], when ftype == DUK__FLD_VARINT */
/* [ offset littleEndian ], when DUK__FLD_TYPEDARRAY (regardless of ftype) */
/* Handle TypedArray vs. Node.js Buffer arg differences */
if (magic_typedarray) {
no_assert = 0;
#if defined(DUK_USE_INTEGER_LE)
endswap = !duk_to_boolean(ctx, 1); /* 1=little endian */
#else
endswap = duk_to_boolean(ctx, 1); /* 1=little endian */
#endif
} else {
no_assert = duk_to_boolean(ctx, (magic_ftype == DUK__FLD_VARINT) ? 2 : 1);
#if defined(DUK_USE_INTEGER_LE)
endswap = magic_bigendian;
#else
endswap = !magic_bigendian;
#endif
}
/* Offset is coerced first to signed integer range and then to unsigned.
* This ensures we can add a small byte length (1-8) to the offset in
* bound checks and not wrap.
*/
offset_signed = duk_to_int(ctx, 0);
offset = (duk_uint_t) offset_signed;
if (offset_signed < 0) {
goto fail_bounds;
}
DUK_DDD(DUK_DDDPRINT("readfield, buffer_length=%ld, offset=%ld, no_assert=%d, "
"magic=%04x, magic_fieldtype=%d, magic_bigendian=%d, magic_signed=%d, "
"endswap=%d",
(long) buffer_length, (long) offset, (int) no_assert,
(unsigned int) magic, (int) magic_ftype, (int) (magic_bigendian >> 3),
(int) (magic_signed >> 4), (int) endswap));
/* Update 'buffer_length' to be the effective, safe limit which
* takes into account the underlying buffer. This value will be
* potentially invalidated by any side effect.
*/
check_length = DUK_HBUFFEROBJECT_CLAMP_BYTELENGTH(h_this, buffer_length);
DUK_DDD(DUK_DDDPRINT("buffer_length=%ld, check_length=%ld",
(long) buffer_length, (long) check_length));
if (h_this->buf) {
buf = DUK_HBUFFEROBJECT_GET_SLICE_BASE(thr->heap, h_this);
} else {
/* Neutered. We could go into the switch-case safely with
* buf == NULL because check_length == 0. To avoid scanbuild
* warnings, fail directly instead.
*/
DUK_ASSERT(check_length == 0);
goto fail_neutered;
}
DUK_ASSERT(buf != NULL);
switch (magic_ftype) {
case DUK__FLD_8BIT: {
duk_uint8_t tmp;
if (offset + 1U > check_length) {
goto fail_bounds;
}
tmp = buf[offset];
if (magic_signed) {
duk_push_int(ctx, (duk_int_t) ((duk_int8_t) tmp));
} else {
duk_push_uint(ctx, (duk_uint_t) tmp);
}
break;
}
case DUK__FLD_16BIT: {
duk_uint16_t tmp;
if (offset + 2U > check_length) {
goto fail_bounds;
}
DUK_MEMCPY((void *) du.uc, (const void *) (buf + offset), 2);
tmp = du.us[0];
if (endswap) {
tmp = DUK_BSWAP16(tmp);
}
if (magic_signed) {
duk_push_int(ctx, (duk_int_t) ((duk_int16_t) tmp));
} else {
duk_push_uint(ctx, (duk_uint_t) tmp);
}
break;
}
case DUK__FLD_32BIT: {
duk_uint32_t tmp;
if (offset + 4U > check_length) {
goto fail_bounds;
}
DUK_MEMCPY((void *) du.uc, (const void *) (buf + offset), 4);
tmp = du.ui[0];
if (endswap) {
tmp = DUK_BSWAP32(tmp);
}
if (magic_signed) {
duk_push_int(ctx, (duk_int_t) ((duk_int32_t) tmp));
} else {
duk_push_uint(ctx, (duk_uint_t) tmp);
}
break;
}
case DUK__FLD_FLOAT: {
duk_uint32_t tmp;
if (offset + 4U > check_length) {
goto fail_bounds;
}
DUK_MEMCPY((void *) du.uc, (const void *) (buf + offset), 4);
if (endswap) {
tmp = du.ui[0];
tmp = DUK_BSWAP32(tmp);
du.ui[0] = tmp;
}
duk_push_number(ctx, (duk_double_t) du.f[0]);
break;
}
case DUK__FLD_DOUBLE: {
if (offset + 8U > check_length) {
goto fail_bounds;
}
DUK_MEMCPY((void *) du.uc, (const void *) (buf + offset), 8);
if (endswap) {
DUK_DBLUNION_BSWAP64(&du);
}
duk_push_number(ctx, (duk_double_t) du.d);
break;
}
case DUK__FLD_VARINT: {
/* Node.js Buffer variable width integer field. We don't really
* care about speed here, so aim for shortest algorithm.
*/
duk_int_t field_bytelen;
duk_int_t i, i_step, i_end;
#if defined(DUK_USE_64BIT_OPS)
duk_int64_t tmp;
duk_small_uint_t shift_tmp;
#else
duk_double_t tmp;
duk_small_int_t highbyte;
#endif
const duk_uint8_t *p;
field_bytelen = duk_get_int(ctx, 1); /* avoid side effects! */
if (field_bytelen < 1 || field_bytelen > 6) {
goto fail_field_length;
}
if (offset + (duk_uint_t) field_bytelen > check_length) {
goto fail_bounds;
}
p = (const duk_uint8_t *) (buf + offset);
/* Slow gathering of value using either 64-bit arithmetic
* or IEEE doubles if 64-bit types not available. Handling
* of negative numbers is a bit non-obvious in both cases.
*/
if (magic_bigendian) {
/* Gather in big endian */
i = 0;
i_step = 1;
i_end = field_bytelen; /* one i_step over */
} else {
/* Gather in little endian */
i = field_bytelen - 1;
i_step = -1;
i_end = -1; /* one i_step over */
}
#if defined(DUK_USE_64BIT_OPS)
tmp = 0;
do {
DUK_ASSERT(i >= 0 && i < field_bytelen);
tmp = (tmp << 8) + (duk_int64_t) p[i];
i += i_step;
} while (i != i_end);
if (magic_signed) {
/* Shift to sign extend. */
shift_tmp = 64 - (field_bytelen * 8);
tmp = (tmp << shift_tmp) >> shift_tmp;
}
duk_push_i64(ctx, tmp);
#else
highbyte = p[i];
if (magic_signed && (highbyte & 0x80) != 0) {
/* 0xff => 255 - 256 = -1; 0x80 => 128 - 256 = -128 */
tmp = (duk_double_t) (highbyte - 256);
} else {
tmp = (duk_double_t) highbyte;
}
for (;;) {
i += i_step;
if (i == i_end) {
break;
}
DUK_ASSERT(i >= 0 && i < field_bytelen);
tmp = (tmp * 256.0) + (duk_double_t) p[i];
}
duk_push_number(ctx, tmp);
#endif
break;
}
default: { /* should never happen but default here */
goto fail_bounds;
}
}
return 1;
fail_neutered:
fail_field_length:
fail_bounds:
if (no_assert) {
/* Node.js return value for noAssert out-of-bounds reads is
* usually (but not always) NaN. Return NaN consistently.
*/
duk_push_nan(ctx);
return 1;
}
return DUK_RET_RANGE_ERROR;
}
#else /* DUK_USE_BUFFEROBJECT_SUPPORT */
DUK_INTERNAL duk_ret_t duk_bi_buffer_readfield(duk_context *ctx) {
DUK_UNREF(ctx);
return DUK_RET_UNSUPPORTED_ERROR;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
#if defined(DUK_USE_BUFFEROBJECT_SUPPORT)
/* XXX: split into separate functions for each field type? */
DUK_INTERNAL duk_ret_t duk_bi_buffer_writefield(duk_context *ctx) {
duk_hthread *thr;
duk_small_int_t magic = (duk_small_int_t) duk_get_current_magic(ctx);
duk_small_int_t magic_ftype;
duk_small_int_t magic_bigendian;
duk_small_int_t magic_signed;
duk_small_int_t magic_typedarray;
duk_small_int_t endswap;
duk_hbufferobject *h_this;
duk_bool_t no_assert;
duk_int_t offset_signed;
duk_uint_t offset;
duk_uint_t buffer_length;
duk_uint_t check_length;
duk_uint8_t *buf;
duk_double_union du;
duk_int_t nbytes = 0;
thr = (duk_hthread *) ctx;
DUK_UNREF(thr);
magic_ftype = magic & 0x0007;
magic_bigendian = magic & 0x0008;
magic_signed = magic & 0x0010;
magic_typedarray = magic & 0x0020;
DUK_UNREF(magic_signed);
h_this = duk__require_bufobj_this(ctx);
DUK_ASSERT(h_this != NULL);
buffer_length = h_this->length;
/* [ value offset noAssert ], when ftype != DUK__FLD_VARINT */
/* [ value offset fieldByteLength noAssert ], when ftype == DUK__FLD_VARINT */
/* [ offset value littleEndian ], when DUK__FLD_TYPEDARRAY (regardless of ftype) */
/* Handle TypedArray vs. Node.js Buffer arg differences */
if (magic_typedarray) {
no_assert = 0;
#if defined(DUK_USE_INTEGER_LE)
endswap = !duk_to_boolean(ctx, 2); /* 1=little endian */
#else
endswap = duk_to_boolean(ctx, 2); /* 1=little endian */
#endif
duk_swap(ctx, 0, 1); /* offset/value order different from Node.js */
} else {
no_assert = duk_to_boolean(ctx, (magic_ftype == DUK__FLD_VARINT) ? 3 : 2);
#if defined(DUK_USE_INTEGER_LE)
endswap = magic_bigendian;
#else
endswap = !magic_bigendian;
#endif
}
/* Offset is coerced first to signed integer range and then to unsigned.
* This ensures we can add a small byte length (1-8) to the offset in
* bound checks and not wrap.
*/
offset_signed = duk_to_int(ctx, 1);
offset = (duk_uint_t) offset_signed;
/* We need 'nbytes' even for a failed offset; return value must be
* (offset + nbytes) even when write fails due to invalid offset.
*/
if (magic_ftype != DUK__FLD_VARINT) {
DUK_ASSERT(magic_ftype >= 0 && magic_ftype < (duk_small_int_t) (sizeof(duk__buffer_nbytes_from_fldtype) / sizeof(duk_uint8_t)));
nbytes = duk__buffer_nbytes_from_fldtype[magic_ftype];
} else {
nbytes = duk_get_int(ctx, 2);
if (nbytes < 1 || nbytes > 6) {
goto fail_field_length;
}
}
DUK_ASSERT(nbytes >= 1 && nbytes <= 8);
/* Now we can check offset validity. */
if (offset_signed < 0) {
goto fail_bounds;
}
DUK_DDD(DUK_DDDPRINT("writefield, value=%!T, buffer_length=%ld, offset=%ld, no_assert=%d, "
"magic=%04x, magic_fieldtype=%d, magic_bigendian=%d, magic_signed=%d, "
"endswap=%d",
duk_get_tval(ctx, 0), (long) buffer_length, (long) offset, (int) no_assert,
(unsigned int) magic, (int) magic_ftype, (int) (magic_bigendian >> 3),
(int) (magic_signed >> 4), (int) endswap));
/* Coerce value to a number before computing check_length, so that
* the field type specific coercion below can't have side effects
* that would invalidate check_length.
*/
duk_to_number(ctx, 0);
/* Update 'buffer_length' to be the effective, safe limit which
* takes into account the underlying buffer. This value will be
* potentially invalidated by any side effect.
*/
check_length = DUK_HBUFFEROBJECT_CLAMP_BYTELENGTH(h_this, buffer_length);
DUK_DDD(DUK_DDDPRINT("buffer_length=%ld, check_length=%ld",
(long) buffer_length, (long) check_length));
if (h_this->buf) {
buf = DUK_HBUFFEROBJECT_GET_SLICE_BASE(thr->heap, h_this);
} else {
/* Neutered. We could go into the switch-case safely with
* buf == NULL because check_length == 0. To avoid scanbuild
* warnings, fail directly instead.
*/
DUK_ASSERT(check_length == 0);
goto fail_neutered;
}
DUK_ASSERT(buf != NULL);
switch (magic_ftype) {
case DUK__FLD_8BIT: {
if (offset + 1U > check_length) {
goto fail_bounds;
}
/* sign doesn't matter when writing */
buf[offset] = (duk_uint8_t) duk_to_uint32(ctx, 0);
break;
}
case DUK__FLD_16BIT: {
duk_uint16_t tmp;
if (offset + 2U > check_length) {
goto fail_bounds;
}
tmp = (duk_uint16_t) duk_to_uint32(ctx, 0);
if (endswap) {
tmp = DUK_BSWAP16(tmp);
}
du.us[0] = tmp;
/* sign doesn't matter when writing */
DUK_MEMCPY((void *) (buf + offset), (const void *) du.uc, 2);
break;
}
case DUK__FLD_32BIT: {
duk_uint32_t tmp;
if (offset + 4U > check_length) {
goto fail_bounds;
}
tmp = (duk_uint32_t) duk_to_uint32(ctx, 0);
if (endswap) {
tmp = DUK_BSWAP32(tmp);
}
du.ui[0] = tmp;
/* sign doesn't matter when writing */
DUK_MEMCPY((void *) (buf + offset), (const void *) du.uc, 4);
break;
}
case DUK__FLD_FLOAT: {
duk_uint32_t tmp;
if (offset + 4U > check_length) {
goto fail_bounds;
}
du.f[0] = (duk_float_t) duk_to_number(ctx, 0);
if (endswap) {
tmp = du.ui[0];
tmp = DUK_BSWAP32(tmp);
du.ui[0] = tmp;
}
/* sign doesn't matter when writing */
DUK_MEMCPY((void *) (buf + offset), (const void *) du.uc, 4);
break;
}
case DUK__FLD_DOUBLE: {
if (offset + 8U > check_length) {
goto fail_bounds;
}
du.d = (duk_double_t) duk_to_number(ctx, 0);
if (endswap) {
DUK_DBLUNION_BSWAP64(&du);
}
/* sign doesn't matter when writing */
DUK_MEMCPY((void *) (buf + offset), (const void *) du.uc, 8);
break;
}
case DUK__FLD_VARINT: {
/* Node.js Buffer variable width integer field. We don't really
* care about speed here, so aim for shortest algorithm.
*/
duk_int_t field_bytelen;
duk_int_t i, i_step, i_end;
#if defined(DUK_USE_64BIT_OPS)
duk_int64_t tmp;
#else
duk_double_t tmp;
#endif
duk_uint8_t *p;
field_bytelen = (duk_int_t) nbytes;
if (offset + (duk_uint_t) field_bytelen > check_length) {
goto fail_bounds;
}
/* Slow writing of value using either 64-bit arithmetic
* or IEEE doubles if 64-bit types not available. There's
* no special sign handling when writing varints.
*/
if (magic_bigendian) {
/* Write in big endian */
i = field_bytelen; /* one i_step added at top of loop */
i_step = -1;
i_end = 0;
} else {
/* Write in little endian */
i = -1; /* one i_step added at top of loop */
i_step = 1;
i_end = field_bytelen - 1;
}
/* XXX: The duk_to_number() cast followed by integer coercion
* is platform specific so NaN, +/- Infinity, and out-of-bounds
* values result in platform specific output now.
* See: test-bi-nodejs-buffer-proto-varint-special.js
*/
#if defined(DUK_USE_64BIT_OPS)
tmp = (duk_int64_t) duk_to_number(ctx, 0);
p = (duk_uint8_t *) (buf + offset);
do {
i += i_step;
DUK_ASSERT(i >= 0 && i < field_bytelen);
p[i] = (duk_uint8_t) (tmp & 0xff);
tmp = tmp >> 8; /* unnecessary shift for last byte */
} while (i != i_end);
#else
tmp = duk_to_number(ctx, 0);
p = (duk_uint8_t *) (buf + offset);
do {
i += i_step;
tmp = DUK_FLOOR(tmp);
DUK_ASSERT(i >= 0 && i < field_bytelen);
p[i] = (duk_uint8_t) (DUK_FMOD(tmp, 256.0));
tmp = tmp / 256.0; /* unnecessary div for last byte */
} while (i != i_end);
#endif
break;
}
default: { /* should never happen but default here */
goto fail_bounds;
}
}
/* Node.js Buffer: return offset + #bytes written (i.e. next
* write offset).
*/
if (magic_typedarray) {
/* For TypedArrays 'undefined' return value is specified
* by ES6 (matches V8).
*/
return 0;
}
duk_push_uint(ctx, offset + nbytes);
return 1;
fail_neutered:
fail_field_length:
fail_bounds:
if (no_assert) {
/* Node.js return value for failed writes is offset + #bytes
* that would have been written.
*/
/* XXX: for negative input offsets, 'offset' will be a large
* positive value so the result here is confusing.
*/
if (magic_typedarray) {
return 0;
}
duk_push_uint(ctx, offset + nbytes);
return 1;
}
return DUK_RET_RANGE_ERROR;
}
#else /* DUK_USE_BUFFEROBJECT_SUPPORT */
DUK_INTERNAL duk_ret_t duk_bi_buffer_writefield(duk_context *ctx) {
DUK_UNREF(ctx);
return DUK_RET_UNSUPPORTED_ERROR;
}
#endif /* DUK_USE_BUFFEROBJECT_SUPPORT */
#undef DUK__FLD_8BIT
#undef DUK__FLD_16BIT
#undef DUK__FLD_32BIT
#undef DUK__FLD_FLOAT
#undef DUK__FLD_DOUBLE
#undef DUK__FLD_VARINT
#undef DUK__FLD_BIGENDIAN
#undef DUK__FLD_SIGNED
#undef DUK__FLD_TYPEDARRAY