/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 4 -*- * * ***** BEGIN LICENSE BLOCK ***** * Version: MPL 1.1/GPL 2.0/LGPL 2.1 * * The contents of this file are subject to the Mozilla Public License Version * 1.1 (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * http://www.mozilla.org/MPL/ * * Software distributed under the License is distributed on an "AS IS" basis, * WITHOUT WARRANTY OF ANY KIND, either express or implied. See the License * for the specific language governing rights and limitations under the * License. * * The Original Code is Mozilla Communicator client code, released * March 31, 1998. * * The Initial Developer of the Original Code is * Netscape Communications Corporation. * Portions created by the Initial Developer are Copyright (C) 1998 * the Initial Developer. All Rights Reserved. * * Contributor(s): * * Alternatively, the contents of this file may be used under the terms of * either of the GNU General Public License Version 2 or later (the "GPL"), * or the GNU Lesser General Public License Version 2.1 or later (the "LGPL"), * in which case the provisions of the GPL or the LGPL are applicable instead * of those above. If you wish to allow use of your version of this file only * under the terms of either the GPL or the LGPL, and not to allow others to * use your version of this file under the terms of the MPL, indicate your * decision by deleting the provisions above and replace them with the notice * and other provisions required by the GPL or the LGPL. If you do not delete * the provisions above, a recipient may use your version of this file under * the terms of any one of the MPL, the GPL or the LGPL. * * ***** END LICENSE BLOCK ***** */ /* * JS atom table. */ #include #include #include "jstypes.h" #include "jsstdint.h" #include "jsutil.h" /* Added by JSIFY */ #include "jshash.h" /* Added by JSIFY */ #include "jsprf.h" #include "jsapi.h" #include "jsatom.h" #include "jsbit.h" #include "jscntxt.h" #include "jsgc.h" #include "jslock.h" #include "jsnum.h" #include "jsparse.h" #include "jsscan.h" #include "jsstr.h" #include "jsversion.h" #include "jsstrinlines.h" /* * ATOM_HASH assumes that JSHashNumber is 32-bit even on 64-bit systems. */ JS_STATIC_ASSERT(sizeof(JSHashNumber) == 4); JS_STATIC_ASSERT(sizeof(JSAtom *) == JS_BYTES_PER_WORD); /* * Start and limit offsets for atom pointers in JSAtomState must be aligned * on the word boundary. */ JS_STATIC_ASSERT(ATOM_OFFSET_START % sizeof(JSAtom *) == 0); JS_STATIC_ASSERT(ATOM_OFFSET_LIMIT % sizeof(JSAtom *) == 0); /* * JS_BOOLEAN_STR and JS_TYPE_STR assume that boolean names starts from the * index 1 and type name starts from the index 1+2 atoms in JSAtomState. */ JS_STATIC_ASSERT(1 * sizeof(JSAtom *) == offsetof(JSAtomState, booleanAtoms) - ATOM_OFFSET_START); JS_STATIC_ASSERT((1 + 2) * sizeof(JSAtom *) == offsetof(JSAtomState, typeAtoms) - ATOM_OFFSET_START); const char * js_AtomToPrintableString(JSContext *cx, JSAtom *atom) { return js_ValueToPrintableString(cx, ATOM_KEY(atom)); } #define JS_PROTO(name,code,init) const char js_##name##_str[] = #name; #include "jsproto.tbl" #undef JS_PROTO /* * String constants for common atoms defined in JSAtomState starting from * JSAtomState.emptyAtom until JSAtomState.lazy. * * The elements of the array after the first empty string define strings * corresponding to the two boolean literals, false and true, followed by the * JSType enumerators from jspubtd.h starting with "undefined" for JSTYPE_VOID * (which is special-value 2) and continuing as initialized below. The static * asserts check these relations. */ JS_STATIC_ASSERT(JSTYPE_LIMIT == 8); JS_STATIC_ASSERT(JSTYPE_VOID == 0); const char *const js_common_atom_names[] = { "", /* emptyAtom */ js_false_str, /* booleanAtoms[0] */ js_true_str, /* booleanAtoms[1] */ js_undefined_str, /* typeAtoms[JSTYPE_VOID] */ js_object_str, /* typeAtoms[JSTYPE_OBJECT] */ js_function_str, /* typeAtoms[JSTYPE_FUNCTION] */ "string", /* typeAtoms[JSTYPE_STRING] */ "number", /* typeAtoms[JSTYPE_NUMBER] */ "boolean", /* typeAtoms[JSTYPE_BOOLEAN] */ js_null_str, /* typeAtoms[JSTYPE_NULL] */ "xml", /* typeAtoms[JSTYPE_XML] */ js_null_str, /* nullAtom */ #define JS_PROTO(name,code,init) js_##name##_str, #include "jsproto.tbl" #undef JS_PROTO js_anonymous_str, /* anonymousAtom */ js_apply_str, /* applyAtom */ js_arguments_str, /* argumentsAtom */ js_arity_str, /* arityAtom */ js_call_str, /* callAtom */ js_callee_str, /* calleeAtom */ js_caller_str, /* callerAtom */ js_class_prototype_str, /* classPrototypeAtom */ js_constructor_str, /* constructorAtom */ js_count_str, /* countAtom */ js_each_str, /* eachAtom */ js_eval_str, /* evalAtom */ js_fileName_str, /* fileNameAtom */ js_get_str, /* getAtom */ js_getter_str, /* getterAtom */ js_index_str, /* indexAtom */ js_input_str, /* inputAtom */ js_iterator_str, /* iteratorAtom */ js_length_str, /* lengthAtom */ js_lineNumber_str, /* lineNumberAtom */ js_message_str, /* messageAtom */ js_name_str, /* nameAtom */ js_next_str, /* nextAtom */ js_noSuchMethod_str, /* noSuchMethodAtom */ js_parent_str, /* parentAtom */ js_proto_str, /* protoAtom */ js_set_str, /* setAtom */ js_setter_str, /* setterAtom */ js_stack_str, /* stackAtom */ js_toLocaleString_str, /* toLocaleStringAtom */ js_toSource_str, /* toSourceAtom */ js_toString_str, /* toStringAtom */ js_valueOf_str, /* valueOfAtom */ js_toJSON_str, /* toJSONAtom */ "(void 0)", /* void0Atom */ js_enumerable_str, /* enumerableAtom */ js_configurable_str, /* configurableAtom */ js_writable_str, /* writableAtom */ js_value_str, /* valueAtom */ #if JS_HAS_XML_SUPPORT js_etago_str, /* etagoAtom */ js_namespace_str, /* namespaceAtom */ js_ptagc_str, /* ptagcAtom */ js_qualifier_str, /* qualifierAtom */ js_space_str, /* spaceAtom */ js_stago_str, /* stagoAtom */ js_star_str, /* starAtom */ js_starQualifier_str, /* starQualifierAtom */ js_tagc_str, /* tagcAtom */ js_xml_str, /* xmlAtom */ #endif #ifdef NARCISSUS js___call___str, /* __call__Atom */ js___construct___str, /* __construct__Atom */ js___hasInstance___str, /* __hasInstance__Atom */ js_ExecutionContext_str, /* ExecutionContextAtom */ js_current_str, /* currentAtom */ #endif }; JS_STATIC_ASSERT(JS_ARRAY_LENGTH(js_common_atom_names) * sizeof(JSAtom *) == LAZY_ATOM_OFFSET_START - ATOM_OFFSET_START); /* * Interpreter macros called by the trace recorder assume common atom indexes * fit in one byte of immediate operand. */ JS_STATIC_ASSERT(JS_ARRAY_LENGTH(js_common_atom_names) < 256); const size_t js_common_atom_count = JS_ARRAY_LENGTH(js_common_atom_names); const char js_anonymous_str[] = "anonymous"; const char js_apply_str[] = "apply"; const char js_arguments_str[] = "arguments"; const char js_arity_str[] = "arity"; const char js_call_str[] = "call"; const char js_callee_str[] = "callee"; const char js_caller_str[] = "caller"; const char js_class_prototype_str[] = "prototype"; const char js_constructor_str[] = "constructor"; const char js_count_str[] = "__count__"; const char js_each_str[] = "each"; const char js_eval_str[] = "eval"; const char js_fileName_str[] = "fileName"; const char js_get_str[] = "get"; const char js_getter_str[] = "getter"; const char js_index_str[] = "index"; const char js_input_str[] = "input"; const char js_iterator_str[] = "__iterator__"; const char js_length_str[] = "length"; const char js_lineNumber_str[] = "lineNumber"; const char js_message_str[] = "message"; const char js_name_str[] = "name"; const char js_next_str[] = "next"; const char js_noSuchMethod_str[] = "__noSuchMethod__"; const char js_object_str[] = "object"; const char js_parent_str[] = "__parent__"; const char js_proto_str[] = "__proto__"; const char js_setter_str[] = "setter"; const char js_set_str[] = "set"; const char js_stack_str[] = "stack"; const char js_toSource_str[] = "toSource"; const char js_toString_str[] = "toString"; const char js_toLocaleString_str[] = "toLocaleString"; const char js_undefined_str[] = "undefined"; const char js_valueOf_str[] = "valueOf"; const char js_toJSON_str[] = "toJSON"; const char js_enumerable_str[] = "enumerable"; const char js_configurable_str[] = "configurable"; const char js_writable_str[] = "writable"; const char js_value_str[] = "value"; #if JS_HAS_XML_SUPPORT const char js_etago_str[] = ""; const char js_qualifier_str[] = "::"; const char js_space_str[] = " "; const char js_stago_str[] = "<"; const char js_star_str[] = "*"; const char js_starQualifier_str[] = "*::"; const char js_tagc_str[] = ">"; const char js_xml_str[] = "xml"; #endif #if JS_HAS_GENERATORS const char js_close_str[] = "close"; const char js_send_str[] = "send"; #endif #ifdef NARCISSUS const char js___call___str[] = "__call__"; const char js___construct___str[] = "__construct__"; const char js___hasInstance___str[] = "__hasInstance__"; const char js_ExecutionContext_str[] = "ExecutionContext"; const char js_current_str[] = "current"; #endif /* * JSAtomState.doubleAtoms and JSAtomState.stringAtoms hashtable entry. To * support pinned and interned string atoms, we use the lowest bits of the * keyAndFlags field to store ATOM_PINNED and ATOM_INTERNED flags. */ typedef struct JSAtomHashEntry { JSDHashEntryHdr hdr; jsuword keyAndFlags; } JSAtomHashEntry; #define ATOM_ENTRY_FLAG_MASK (ATOM_PINNED | ATOM_INTERNED) JS_STATIC_ASSERT(ATOM_ENTRY_FLAG_MASK < JSVAL_ALIGN); /* * Helper macros to access and modify JSAtomHashEntry. */ #define TO_ATOM_ENTRY(hdr) ((JSAtomHashEntry *) hdr) #define ATOM_ENTRY_KEY(entry) \ ((void *)((entry)->keyAndFlags & ~ATOM_ENTRY_FLAG_MASK)) #define ATOM_ENTRY_FLAGS(entry) \ ((uintN)((entry)->keyAndFlags & ATOM_ENTRY_FLAG_MASK)) #define INIT_ATOM_ENTRY(entry, key) \ ((void)((entry)->keyAndFlags = (jsuword)(key))) #define ADD_ATOM_ENTRY_FLAGS(entry, flags) \ ((void)((entry)->keyAndFlags |= (jsuword)(flags))) #define CLEAR_ATOM_ENTRY_FLAGS(entry, flags) \ ((void)((entry)->keyAndFlags &= ~(jsuword)(flags))) static JSDHashNumber HashDouble(JSDHashTable *table, const void *key); static JSBool MatchDouble(JSDHashTable *table, const JSDHashEntryHdr *hdr, const void *key); static JSDHashNumber HashString(JSDHashTable *table, const void *key); static JSBool MatchString(JSDHashTable *table, const JSDHashEntryHdr *hdr, const void *key); static const JSDHashTableOps DoubleHashOps = { JS_DHashAllocTable, JS_DHashFreeTable, HashDouble, MatchDouble, JS_DHashMoveEntryStub, JS_DHashClearEntryStub, JS_DHashFinalizeStub, NULL }; static const JSDHashTableOps StringHashOps = { JS_DHashAllocTable, JS_DHashFreeTable, HashString, MatchString, JS_DHashMoveEntryStub, JS_DHashClearEntryStub, JS_DHashFinalizeStub, NULL }; #define IS_DOUBLE_TABLE(table) ((table)->ops == &DoubleHashOps) #define IS_STRING_TABLE(table) ((table)->ops == &StringHashOps) #define IS_INITIALIZED_STATE(state) IS_DOUBLE_TABLE(&(state)->doubleAtoms) static JSDHashNumber HashDouble(JSDHashTable *table, const void *key) { JS_ASSERT(IS_DOUBLE_TABLE(table)); return JS_HASH_DOUBLE(*(jsdouble *)key); } static JSDHashNumber HashString(JSDHashTable *table, const void *key) { JS_ASSERT(IS_STRING_TABLE(table)); return js_HashString((JSString *)key); } static JSBool MatchDouble(JSDHashTable *table, const JSDHashEntryHdr *hdr, const void *key) { JSAtomHashEntry *entry = TO_ATOM_ENTRY(hdr); jsdouble d1, d2; JS_ASSERT(IS_DOUBLE_TABLE(table)); if (entry->keyAndFlags == 0) { /* See comments in MatchString. */ return JS_FALSE; } d1 = *(jsdouble *)ATOM_ENTRY_KEY(entry); d2 = *(jsdouble *)key; if (JSDOUBLE_IS_NaN(d1)) return JSDOUBLE_IS_NaN(d2); #if defined(XP_WIN) /* XXX MSVC miscompiles such that (NaN == 0) */ if (JSDOUBLE_IS_NaN(d2)) return JS_FALSE; #endif return d1 == d2; } static JSBool MatchString(JSDHashTable *table, const JSDHashEntryHdr *hdr, const void *key) { JSAtomHashEntry *entry = TO_ATOM_ENTRY(hdr); JS_ASSERT(IS_STRING_TABLE(table)); if (entry->keyAndFlags == 0) { /* * This happens when js_AtomizeString adds a new hash entry and * releases the lock but before it takes the lock the second time to * initialize keyAndFlags for the entry. * * We always return false for such entries so JS_DHashTableOperate * never finds them. We clean them during GC's sweep phase. * * It means that with a contested lock or when GC is triggered outside * the lock we may end up adding two entries, but this is a price for * simpler code. */ return JS_FALSE; } return js_EqualStrings((JSString *)ATOM_ENTRY_KEY(entry), (JSString *)key); } /* * For a browser build from 2007-08-09 after the browser starts up there are * just 55 double atoms, but over 15000 string atoms. Not to penalize more * economical embeddings allocating too much memory initially we initialize * atomized strings with just 1K entries. */ #define JS_STRING_HASH_COUNT 1024 #define JS_DOUBLE_HASH_COUNT 64 JSBool js_InitAtomState(JSRuntime *rt) { JSAtomState *state = &rt->atomState; /* * The caller must zero the state before calling this function. */ JS_ASSERT(!state->stringAtoms.ops); JS_ASSERT(!state->doubleAtoms.ops); if (!JS_DHashTableInit(&state->stringAtoms, &StringHashOps, NULL, sizeof(JSAtomHashEntry), JS_DHASH_DEFAULT_CAPACITY(JS_STRING_HASH_COUNT))) { state->stringAtoms.ops = NULL; return JS_FALSE; } JS_ASSERT(IS_STRING_TABLE(&state->stringAtoms)); if (!JS_DHashTableInit(&state->doubleAtoms, &DoubleHashOps, NULL, sizeof(JSAtomHashEntry), JS_DHASH_DEFAULT_CAPACITY(JS_DOUBLE_HASH_COUNT))) { state->doubleAtoms.ops = NULL; JS_DHashTableFinish(&state->stringAtoms); state->stringAtoms.ops = NULL; return JS_FALSE; } JS_ASSERT(IS_DOUBLE_TABLE(&state->doubleAtoms)); #ifdef JS_THREADSAFE js_InitLock(&state->lock); #endif JS_ASSERT(IS_INITIALIZED_STATE(state)); return JS_TRUE; } static JSDHashOperator js_string_uninterner(JSDHashTable *table, JSDHashEntryHdr *hdr, uint32 number, void *arg) { JSAtomHashEntry *entry = TO_ATOM_ENTRY(hdr); JSRuntime *rt = (JSRuntime *)arg; JSString *str; /* * Any string entry that remains at this point must be initialized, as the * last GC should clean any uninitialized ones. */ JS_ASSERT(IS_STRING_TABLE(table)); JS_ASSERT(entry->keyAndFlags != 0); str = (JSString *)ATOM_ENTRY_KEY(entry); js_FinalizeStringRT(rt, str); return JS_DHASH_NEXT; } void js_FinishAtomState(JSRuntime *rt) { JSAtomState *state = &rt->atomState; if (!IS_INITIALIZED_STATE(state)) { /* * We are called with uninitialized state when JS_NewRuntime fails and * calls JS_DestroyRuntime on a partially initialized runtime. */ return; } JS_DHashTableEnumerate(&state->stringAtoms, js_string_uninterner, rt); JS_DHashTableFinish(&state->stringAtoms); JS_DHashTableFinish(&state->doubleAtoms); #ifdef JS_THREADSAFE js_FinishLock(&state->lock); #endif #ifdef DEBUG memset(state, JS_FREE_PATTERN, sizeof *state); #endif } JSBool js_InitCommonAtoms(JSContext *cx) { JSAtomState *state = &cx->runtime->atomState; uintN i; JSAtom **atoms; atoms = COMMON_ATOMS_START(state); for (i = 0; i < JS_ARRAY_LENGTH(js_common_atom_names); i++, atoms++) { *atoms = js_Atomize(cx, js_common_atom_names[i], strlen(js_common_atom_names[i]), ATOM_PINNED); if (!*atoms) return JS_FALSE; } JS_ASSERT((uint8 *)atoms - (uint8 *)state == LAZY_ATOM_OFFSET_START); memset(atoms, 0, ATOM_OFFSET_LIMIT - LAZY_ATOM_OFFSET_START); return JS_TRUE; } static JSDHashOperator js_atom_unpinner(JSDHashTable *table, JSDHashEntryHdr *hdr, uint32 number, void *arg) { JS_ASSERT(IS_STRING_TABLE(table)); CLEAR_ATOM_ENTRY_FLAGS(TO_ATOM_ENTRY(hdr), ATOM_PINNED); return JS_DHASH_NEXT; } void js_FinishCommonAtoms(JSContext *cx) { JSAtomState *state = &cx->runtime->atomState; JS_DHashTableEnumerate(&state->stringAtoms, js_atom_unpinner, NULL); #ifdef DEBUG memset(COMMON_ATOMS_START(state), JS_FREE_PATTERN, ATOM_OFFSET_LIMIT - ATOM_OFFSET_START); #endif } static JSDHashOperator js_locked_atom_tracer(JSDHashTable *table, JSDHashEntryHdr *hdr, uint32 number, void *arg) { JSAtomHashEntry *entry = TO_ATOM_ENTRY(hdr); JSTracer *trc = (JSTracer *)arg; if (entry->keyAndFlags == 0) { /* Ignore uninitialized entries during tracing. */ return JS_DHASH_NEXT; } JS_SET_TRACING_INDEX(trc, "locked_atom", (size_t)number); JS_CallTracer(trc, ATOM_ENTRY_KEY(entry), IS_STRING_TABLE(table) ? JSTRACE_STRING : JSTRACE_DOUBLE); return JS_DHASH_NEXT; } static JSDHashOperator js_pinned_atom_tracer(JSDHashTable *table, JSDHashEntryHdr *hdr, uint32 number, void *arg) { JSAtomHashEntry *entry = TO_ATOM_ENTRY(hdr); JSTracer *trc = (JSTracer *)arg; uintN flags = ATOM_ENTRY_FLAGS(entry); JS_ASSERT(IS_STRING_TABLE(table)); if (flags & (ATOM_PINNED | ATOM_INTERNED)) { JS_SET_TRACING_INDEX(trc, flags & ATOM_PINNED ? "pinned_atom" : "interned_atom", (size_t)number); JS_CallTracer(trc, ATOM_ENTRY_KEY(entry), JSTRACE_STRING); } return JS_DHASH_NEXT; } void js_TraceAtomState(JSTracer *trc, JSBool allAtoms) { JSRuntime *rt = trc->context->runtime; JSAtomState *state = &rt->atomState; if (allAtoms) { JS_DHashTableEnumerate(&state->doubleAtoms, js_locked_atom_tracer, trc); JS_DHashTableEnumerate(&state->stringAtoms, js_locked_atom_tracer, trc); } else { JS_DHashTableEnumerate(&state->stringAtoms, js_pinned_atom_tracer, trc); } } static JSDHashOperator js_atom_sweeper(JSDHashTable *table, JSDHashEntryHdr *hdr, uint32 number, void *arg) { JSAtomHashEntry *entry = TO_ATOM_ENTRY(hdr); JSContext *cx = (JSContext *)arg; /* Remove uninitialized entries. */ if (entry->keyAndFlags == 0) return JS_DHASH_REMOVE; if (ATOM_ENTRY_FLAGS(entry) & (ATOM_PINNED | ATOM_INTERNED)) { /* Pinned or interned key cannot be finalized. */ JS_ASSERT(!js_IsAboutToBeFinalized(cx, ATOM_ENTRY_KEY(entry))); } else if (js_IsAboutToBeFinalized(cx, ATOM_ENTRY_KEY(entry))) { /* Remove entries with things about to be GC'ed. */ return JS_DHASH_REMOVE; } return JS_DHASH_NEXT; } void js_SweepAtomState(JSContext *cx) { JSAtomState *state = &cx->runtime->atomState; JS_DHashTableEnumerate(&state->doubleAtoms, js_atom_sweeper, cx); JS_DHashTableEnumerate(&state->stringAtoms, js_atom_sweeper, cx); /* * Optimize for simplicity and mutate table generation numbers even if the * sweeper has not removed any entries. */ state->doubleAtoms.generation++; state->stringAtoms.generation++; } JSAtom * js_AtomizeDouble(JSContext *cx, jsdouble d) { JSAtomState *state; JSDHashTable *table; JSAtomHashEntry *entry; uint32 gen; jsdouble *key; jsval v; state = &cx->runtime->atomState; table = &state->doubleAtoms; JS_LOCK(cx, &state->lock); entry = TO_ATOM_ENTRY(JS_DHashTableOperate(table, &d, JS_DHASH_ADD)); if (!entry) goto failed_hash_add; if (entry->keyAndFlags == 0) { gen = ++table->generation; JS_UNLOCK(cx, &state->lock); key = js_NewWeaklyRootedDouble(cx, d); if (!key) return NULL; JS_LOCK(cx, &state->lock); if (table->generation == gen) { JS_ASSERT(entry->keyAndFlags == 0); } else { entry = TO_ATOM_ENTRY(JS_DHashTableOperate(table, key, JS_DHASH_ADD)); if (!entry) goto failed_hash_add; if (entry->keyAndFlags != 0) goto finish; ++table->generation; } INIT_ATOM_ENTRY(entry, key); } finish: v = DOUBLE_TO_JSVAL((jsdouble *)ATOM_ENTRY_KEY(entry)); cx->weakRoots.lastAtom = v; JS_UNLOCK(cx, &state->lock); return (JSAtom *)v; failed_hash_add: JS_UNLOCK(cx, &state->lock); JS_ReportOutOfMemory(cx); return NULL; } JSAtom * js_AtomizeString(JSContext *cx, JSString *str, uintN flags) { jsval v; JSAtomState *state; JSDHashTable *table; JSAtomHashEntry *entry; JSString *key; uint32 gen; JS_ASSERT(!(flags & ~(ATOM_PINNED|ATOM_INTERNED|ATOM_TMPSTR|ATOM_NOCOPY))); JS_ASSERT_IF(flags & ATOM_NOCOPY, flags & ATOM_TMPSTR); if (str->isAtomized()) return (JSAtom *) STRING_TO_JSVAL(str); size_t length = str->length(); if (length == 1) { jschar c = str->chars()[0]; if (c < UNIT_STRING_LIMIT) return (JSAtom *) STRING_TO_JSVAL(JSString::unitString(c)); } /* * Here we know that JSString::intStringTable covers only 256 (or at least * not 1000 or more) chars. We rely on order here to resolve the unit vs. * int string atom identity issue by giving priority to unit strings for * '0' through '9' (see JSString::intString in jsstrinlines.h). */ JS_STATIC_ASSERT(INT_STRING_LIMIT <= 999); if (2 <= length && length <= 3) { const jschar *chars = str->chars(); if ('1' <= chars[0] && chars[0] <= '9' && '0' <= chars[1] && chars[1] <= '9' && (length == 2 || ('0' <= chars[2] && chars[2] <= '9'))) { jsint i = (chars[0] - '0') * 10 + chars[1] - '0'; if (length == 3) i = i * 10 + chars[2] - '0'; if (jsuint(i) < INT_STRING_LIMIT) return (JSAtom *) STRING_TO_JSVAL(JSString::intString(i)); } } state = &cx->runtime->atomState; table = &state->stringAtoms; JS_LOCK(cx, &state->lock); entry = TO_ATOM_ENTRY(JS_DHashTableOperate(table, str, JS_DHASH_ADD)); if (!entry) goto failed_hash_add; if (entry->keyAndFlags != 0) { key = (JSString *)ATOM_ENTRY_KEY(entry); } else { /* * We created a new hashtable entry. Unless str is already allocated * from the GC heap and flat, we have to release state->lock as * string construction is a complex operation. For example, it can * trigger GC which may rehash the table and make the entry invalid. */ ++table->generation; if (!(flags & ATOM_TMPSTR) && str->isFlat()) { str->flatClearMutable(); key = str; } else { gen = table->generation; JS_UNLOCK(cx, &state->lock); if (flags & ATOM_TMPSTR) { if (flags & ATOM_NOCOPY) { key = js_NewString(cx, str->flatChars(), str->flatLength()); if (!key) return NULL; /* Finish handing off chars to the GC'ed key string. */ str->mChars = NULL; } else { key = js_NewStringCopyN(cx, str->flatChars(), str->flatLength()); if (!key) return NULL; } } else { JS_ASSERT(str->isDependent()); if (!js_UndependString(cx, str)) return NULL; key = str; } JS_LOCK(cx, &state->lock); if (table->generation == gen) { JS_ASSERT(entry->keyAndFlags == 0); } else { entry = TO_ATOM_ENTRY(JS_DHashTableOperate(table, key, JS_DHASH_ADD)); if (!entry) goto failed_hash_add; if (entry->keyAndFlags != 0) { key = (JSString *)ATOM_ENTRY_KEY(entry); goto finish; } ++table->generation; } } INIT_ATOM_ENTRY(entry, key); key->flatSetAtomized(); } finish: ADD_ATOM_ENTRY_FLAGS(entry, flags & (ATOM_PINNED | ATOM_INTERNED)); JS_ASSERT(key->isAtomized()); v = STRING_TO_JSVAL(key); cx->weakRoots.lastAtom = v; JS_UNLOCK(cx, &state->lock); return (JSAtom *)v; failed_hash_add: JS_UNLOCK(cx, &state->lock); JS_ReportOutOfMemory(cx); return NULL; } JSAtom * js_Atomize(JSContext *cx, const char *bytes, size_t length, uintN flags) { jschar *chars; JSString str; JSAtom *atom; /* * Avoiding the malloc in js_InflateString on shorter strings saves us * over 20,000 malloc calls on mozilla browser startup. This compares to * only 131 calls where the string is longer than a 31 char (net) buffer. * The vast majority of atomized strings are already in the hashtable. So * js_AtomizeString rarely has to copy the temp string we make. */ #define ATOMIZE_BUF_MAX 32 jschar inflated[ATOMIZE_BUF_MAX]; size_t inflatedLength = ATOMIZE_BUF_MAX - 1; if (length < ATOMIZE_BUF_MAX) { js_InflateStringToBuffer(cx, bytes, length, inflated, &inflatedLength); inflated[inflatedLength] = 0; chars = inflated; } else { inflatedLength = length; chars = js_InflateString(cx, bytes, &inflatedLength); if (!chars) return NULL; flags |= ATOM_NOCOPY; } str.initFlat(chars, inflatedLength); atom = js_AtomizeString(cx, &str, ATOM_TMPSTR | flags); if (chars != inflated && str.flatChars()) cx->free(chars); return atom; } JSAtom * js_AtomizeChars(JSContext *cx, const jschar *chars, size_t length, uintN flags) { JSString str; str.initFlat((jschar *)chars, length); return js_AtomizeString(cx, &str, ATOM_TMPSTR | flags); } JSAtom * js_GetExistingStringAtom(JSContext *cx, const jschar *chars, size_t length) { JSString str, *str2; JSAtomState *state; JSDHashEntryHdr *hdr; if (length == 1) { jschar c = *chars; if (c < UNIT_STRING_LIMIT) return (JSAtom *) STRING_TO_JSVAL(JSString::unitString(c)); } str.initFlat((jschar *)chars, length); state = &cx->runtime->atomState; JS_LOCK(cx, &state->lock); hdr = JS_DHashTableOperate(&state->stringAtoms, &str, JS_DHASH_LOOKUP); str2 = JS_DHASH_ENTRY_IS_BUSY(hdr) ? (JSString *)ATOM_ENTRY_KEY(TO_ATOM_ENTRY(hdr)) : NULL; JS_UNLOCK(cx, &state->lock); return str2 ? (JSAtom *)STRING_TO_JSVAL(str2) : NULL; } JSBool js_AtomizePrimitiveValue(JSContext *cx, jsval v, JSAtom **atomp) { JSAtom *atom; if (JSVAL_IS_STRING(v)) { atom = js_AtomizeString(cx, JSVAL_TO_STRING(v), 0); if (!atom) return JS_FALSE; } else if (JSVAL_IS_DOUBLE(v)) { atom = js_AtomizeDouble(cx, *JSVAL_TO_DOUBLE(v)); if (!atom) return JS_FALSE; } else { JS_ASSERT(JSVAL_IS_INT(v) || JSVAL_IS_BOOLEAN(v) || JSVAL_IS_NULL(v) || JSVAL_IS_VOID(v)); atom = (JSAtom *)v; } *atomp = atom; return JS_TRUE; } #ifdef DEBUG static JSDHashOperator atom_dumper(JSDHashTable *table, JSDHashEntryHdr *hdr, uint32 number, void *arg) { JSAtomHashEntry *entry = TO_ATOM_ENTRY(hdr); FILE *fp = (FILE *)arg; void *key; uintN flags; fprintf(fp, "%3u %08x ", number, (uintN)entry->hdr.keyHash); if (entry->keyAndFlags == 0) { fputs("", fp); } else { key = ATOM_ENTRY_KEY(entry); if (IS_DOUBLE_TABLE(table)) { fprintf(fp, "%.16g", *(jsdouble *)key); } else { JS_ASSERT(IS_STRING_TABLE(table)); js_FileEscapedString(fp, (JSString *)key, '"'); } flags = ATOM_ENTRY_FLAGS(entry); if (flags != 0) { fputs((flags & (ATOM_PINNED | ATOM_INTERNED)) ? " pinned | interned" : (flags & ATOM_PINNED) ? " pinned" : " interned", fp); } } putc('\n', fp); return JS_DHASH_NEXT; } JS_FRIEND_API(void) js_DumpAtoms(JSContext *cx, FILE *fp) { JSAtomState *state = &cx->runtime->atomState; fprintf(fp, "stringAtoms table contents:\n"); JS_DHashTableEnumerate(&state->stringAtoms, atom_dumper, fp); #ifdef JS_DHASHMETER JS_DHashTableDumpMeter(&state->stringAtoms, atom_dumper, fp); #endif putc('\n', fp); fprintf(fp, "doubleAtoms table contents:\n"); JS_DHashTableEnumerate(&state->doubleAtoms, atom_dumper, fp); #ifdef JS_DHASHMETER JS_DHashTableDumpMeter(&state->doubleAtoms, atom_dumper, fp); #endif putc('\n', fp); } #endif static JSHashNumber js_hash_atom_ptr(const void *key) { const JSAtom *atom = (const JSAtom *) key; return ATOM_HASH(atom); } #if JS_BITS_PER_WORD == 32 # define TEMP_SIZE_START_LOG2 5 #else # define TEMP_SIZE_START_LOG2 6 #endif #define TEMP_SIZE_LIMIT_LOG2 (TEMP_SIZE_START_LOG2 + NUM_TEMP_FREELISTS) #define TEMP_SIZE_START JS_BIT(TEMP_SIZE_START_LOG2) #define TEMP_SIZE_LIMIT JS_BIT(TEMP_SIZE_LIMIT_LOG2) JS_STATIC_ASSERT(TEMP_SIZE_START >= sizeof(JSHashTable)); static void * js_alloc_temp_space(void *priv, size_t size) { JSCompiler *jsc = (JSCompiler *) priv; void *space; if (size < TEMP_SIZE_LIMIT) { int bin = JS_CeilingLog2(size) - TEMP_SIZE_START_LOG2; JS_ASSERT(unsigned(bin) < NUM_TEMP_FREELISTS); space = jsc->tempFreeList[bin]; if (space) { jsc->tempFreeList[bin] = *(void **)space; return space; } } JS_ARENA_ALLOCATE(space, &jsc->context->tempPool, size); if (!space) js_ReportOutOfScriptQuota(jsc->context); return space; } static void js_free_temp_space(void *priv, void *item, size_t size) { if (size >= TEMP_SIZE_LIMIT) return; JSCompiler *jsc = (JSCompiler *) priv; int bin = JS_CeilingLog2(size) - TEMP_SIZE_START_LOG2; JS_ASSERT(unsigned(bin) < NUM_TEMP_FREELISTS); *(void **)item = jsc->tempFreeList[bin]; jsc->tempFreeList[bin] = item; } static JSHashEntry * js_alloc_temp_entry(void *priv, const void *key) { JSCompiler *jsc = (JSCompiler *) priv; JSAtomListElement *ale; ale = jsc->aleFreeList; if (ale) { jsc->aleFreeList = ALE_NEXT(ale); return &ale->entry; } JS_ARENA_ALLOCATE_TYPE(ale, JSAtomListElement, &jsc->context->tempPool); if (!ale) { js_ReportOutOfScriptQuota(jsc->context); return NULL; } return &ale->entry; } static void js_free_temp_entry(void *priv, JSHashEntry *he, uintN flag) { JSCompiler *jsc = (JSCompiler *) priv; JSAtomListElement *ale = (JSAtomListElement *) he; ALE_SET_NEXT(ale, jsc->aleFreeList); jsc->aleFreeList = ale; } static JSHashAllocOps temp_alloc_ops = { js_alloc_temp_space, js_free_temp_space, js_alloc_temp_entry, js_free_temp_entry }; JSAtomListElement * JSAtomList::rawLookup(JSAtom *atom, JSHashEntry **&hep) { JSAtomListElement *ale; if (table) { hep = JS_HashTableRawLookup(table, ATOM_HASH(atom), atom); ale = *hep ? (JSAtomListElement *) *hep : NULL; } else { JSHashEntry **alep = &list; hep = NULL; while ((ale = (JSAtomListElement *)*alep) != NULL) { if (ALE_ATOM(ale) == atom) { /* Hit, move atom's element to the front of the list. */ *alep = ale->entry.next; ale->entry.next = list; list = &ale->entry; break; } alep = &ale->entry.next; } } return ale; } #define ATOM_LIST_HASH_THRESHOLD 12 JSAtomListElement * JSAtomList::add(JSCompiler *jsc, JSAtom *atom, AddHow how) { JS_ASSERT(!set); JSAtomListElement *ale, *ale2, *next; JSHashEntry **hep; ale = rawLookup(atom, hep); if (!ale || how != UNIQUE) { if (count < ATOM_LIST_HASH_THRESHOLD && !table) { /* Few enough for linear search and no hash table yet needed. */ ale = (JSAtomListElement *)js_alloc_temp_entry(jsc, atom); if (!ale) return NULL; ALE_SET_ATOM(ale, atom); if (how == HOIST) { ale->entry.next = NULL; hep = (JSHashEntry **) &list; while (*hep) hep = &(*hep)->next; *hep = &ale->entry; } else { ale->entry.next = list; list = &ale->entry; } } else { /* * We should hash, or else we already are hashing, but count was * reduced by JSAtomList::rawRemove below ATOM_LIST_HASH_THRESHOLD. * Check whether we should create the table. */ if (!table) { /* No hash table yet, so hep had better be null! */ JS_ASSERT(!hep); table = JS_NewHashTable(count + 1, js_hash_atom_ptr, JS_CompareValues, JS_CompareValues, &temp_alloc_ops, jsc); if (!table) return NULL; /* * Set ht->nentries explicitly, because we are moving entries * from list to ht, not calling JS_HashTable(Raw|)Add. */ table->nentries = count; /* * Insert each ale on list into the new hash table. Append to * the hash chain rather than inserting at the bucket head, to * preserve order among entries with the same key. */ for (ale2 = (JSAtomListElement *)list; ale2; ale2 = next) { next = ALE_NEXT(ale2); ale2->entry.keyHash = ATOM_HASH(ALE_ATOM(ale2)); hep = JS_HashTableRawLookup(table, ale2->entry.keyHash, ale2->entry.key); while (*hep) hep = &(*hep)->next; *hep = &ale2->entry; ale2->entry.next = NULL; } list = NULL; /* Set hep for insertion of atom's ale, immediately below. */ hep = JS_HashTableRawLookup(table, ATOM_HASH(atom), atom); } /* Finally, add an entry for atom into the hash bucket at hep. */ ale = (JSAtomListElement *) JS_HashTableRawAdd(table, hep, ATOM_HASH(atom), atom, NULL); if (!ale) return NULL; /* * If hoisting, move ale to the end of its chain after we called * JS_HashTableRawAdd, since RawAdd may have grown the table and * then recomputed hep to refer to the pointer to the first entry * with the given key. */ if (how == HOIST && ale->entry.next) { JS_ASSERT(*hep == &ale->entry); *hep = ale->entry.next; ale->entry.next = NULL; do { hep = &(*hep)->next; } while (*hep); *hep = &ale->entry; } } ALE_SET_INDEX(ale, count++); } return ale; } void JSAtomList::rawRemove(JSCompiler *jsc, JSAtomListElement *ale, JSHashEntry **hep) { JS_ASSERT(!set); JS_ASSERT(count != 0); if (table) { JS_ASSERT(hep); JS_HashTableRawRemove(table, hep, &ale->entry); } else { JS_ASSERT(!hep); hep = &list; while (*hep != &ale->entry) { JS_ASSERT(*hep); hep = &(*hep)->next; } *hep = ale->entry.next; js_free_temp_entry(jsc, &ale->entry, HT_FREE_ENTRY); } --count; } JSAtomListElement * JSAtomListIterator::operator ()() { JSAtomListElement *ale; JSHashTable *ht; if (index == uint32(-1)) return NULL; ale = next; if (!ale) { ht = list->table; if (!ht) goto done; do { if (index == JS_BIT(JS_HASH_BITS - ht->shift)) goto done; next = (JSAtomListElement *) ht->buckets[index++]; } while (!next); ale = next; } next = ALE_NEXT(ale); return ale; done: index = uint32(-1); return NULL; } static intN js_map_atom(JSHashEntry *he, intN i, void *arg) { JSAtomListElement *ale = (JSAtomListElement *)he; JSAtom **vector = (JSAtom **) arg; vector[ALE_INDEX(ale)] = ALE_ATOM(ale); return HT_ENUMERATE_NEXT; } #ifdef DEBUG static jsrefcount js_atom_map_count; static jsrefcount js_atom_map_hash_table_count; #endif void js_InitAtomMap(JSContext *cx, JSAtomMap *map, JSAtomList *al) { JSAtom **vector; JSAtomListElement *ale; uint32 count; /* Map length must already be initialized. */ JS_ASSERT(al->count == map->length); #ifdef DEBUG JS_ATOMIC_INCREMENT(&js_atom_map_count); #endif ale = (JSAtomListElement *)al->list; if (!ale && !al->table) { JS_ASSERT(!map->vector); return; } count = al->count; vector = map->vector; if (al->table) { #ifdef DEBUG JS_ATOMIC_INCREMENT(&js_atom_map_hash_table_count); #endif JS_HashTableEnumerateEntries(al->table, js_map_atom, vector); } else { do { vector[ALE_INDEX(ale)] = ALE_ATOM(ale); } while ((ale = ALE_NEXT(ale)) != NULL); } al->clear(); }