| // -*- mode: c++; c-basic-offset: 4 -*- |
| /* |
| * This file is part of the KDE libraries |
| * Copyright (C) 2003, 2004, 2005, 2006 Apple Computer, Inc. |
| * |
| * This library is free software; you can redistribute it and/or |
| * modify it under the terms of the GNU Lesser General Public |
| * License as published by the Free Software Foundation; either |
| * version 2 of the License, or (at your option) any later version. |
| * |
| * This library is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| * Lesser General Public License for more details. |
| * |
| * You should have received a copy of the GNU Lesser General Public |
| * License along with this library; if not, write to the Free Software |
| * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
| * |
| */ |
| |
| #include "config.h" |
| #include "collector.h" |
| |
| #include <wtf/FastMalloc.h> |
| #include <wtf/FastMallocInternal.h> |
| #include <wtf/HashCountedSet.h> |
| #include "internal.h" |
| #include "list.h" |
| #include "value.h" |
| |
| #include <setjmp.h> |
| #include <algorithm> |
| |
| #if PLATFORM(DARWIN) |
| |
| #include <pthread.h> |
| #include <mach/mach_port.h> |
| #include <mach/task.h> |
| #include <mach/thread_act.h> |
| |
| #elif PLATFORM(WIN_OS) |
| |
| #include <windows.h> |
| |
| #elif PLATFORM(UNIX) |
| |
| #include <pthread.h> |
| |
| #if HAVE(PTHREAD_NP_H) |
| #include <pthread_np.h> |
| #endif |
| |
| #endif |
| |
| #define DEBUG_COLLECTOR 0 |
| |
| using std::max; |
| |
| namespace KJS { |
| |
| // tunable parameters |
| const size_t MINIMUM_CELL_SIZE = 48; |
| const size_t BLOCK_SIZE = (8 * 4096); |
| const size_t SPARE_EMPTY_BLOCKS = 2; |
| const size_t MIN_ARRAY_SIZE = 14; |
| const size_t GROWTH_FACTOR = 2; |
| const size_t LOW_WATER_FACTOR = 4; |
| const size_t ALLOCATIONS_PER_COLLECTION = 1000; |
| |
| // derived constants |
| const size_t CELL_ARRAY_LENGTH = (MINIMUM_CELL_SIZE / sizeof(double)) + (MINIMUM_CELL_SIZE % sizeof(double) != 0 ? sizeof(double) : 0); |
| const size_t CELL_SIZE = CELL_ARRAY_LENGTH * sizeof(double); |
| const size_t CELLS_PER_BLOCK = ((BLOCK_SIZE * 8 - sizeof(uint32_t) * 8 - sizeof(void *) * 8) / (CELL_SIZE * 8)); |
| |
| |
| |
| struct CollectorCell { |
| union { |
| double memory[CELL_ARRAY_LENGTH]; |
| struct { |
| void *zeroIfFree; |
| ptrdiff_t next; |
| } freeCell; |
| } u; |
| }; |
| |
| |
| struct CollectorBlock { |
| CollectorCell cells[CELLS_PER_BLOCK]; |
| uint32_t usedCells; |
| CollectorCell *freeList; |
| }; |
| |
| struct CollectorHeap { |
| CollectorBlock **blocks; |
| size_t numBlocks; |
| size_t usedBlocks; |
| size_t firstBlockWithPossibleSpace; |
| |
| CollectorCell **oversizeCells; |
| size_t numOversizeCells; |
| size_t usedOversizeCells; |
| |
| size_t numLiveObjects; |
| size_t numLiveObjectsAtLastCollect; |
| }; |
| |
| static CollectorHeap heap = {NULL, 0, 0, 0, NULL, 0, 0, 0, 0}; |
| |
| bool Collector::memoryFull = false; |
| |
| void* Collector::allocate(size_t s) |
| { |
| assert(JSLock::lockCount() > 0); |
| |
| // collect if needed |
| size_t numLiveObjects = heap.numLiveObjects; |
| size_t numLiveObjectsAtLastCollect = heap.numLiveObjectsAtLastCollect; |
| size_t numNewObjects = numLiveObjects - numLiveObjectsAtLastCollect; |
| |
| if (numNewObjects >= ALLOCATIONS_PER_COLLECTION && numNewObjects >= numLiveObjectsAtLastCollect) { |
| collect(); |
| numLiveObjects = heap.numLiveObjects; |
| } |
| |
| if (s > CELL_SIZE) { |
| // oversize allocator |
| size_t usedOversizeCells = heap.usedOversizeCells; |
| size_t numOversizeCells = heap.numOversizeCells; |
| |
| if (usedOversizeCells == numOversizeCells) { |
| numOversizeCells = max(MIN_ARRAY_SIZE, numOversizeCells * GROWTH_FACTOR); |
| heap.numOversizeCells = numOversizeCells; |
| heap.oversizeCells = static_cast<CollectorCell **>(fastRealloc(heap.oversizeCells, numOversizeCells * sizeof(CollectorCell *))); |
| } |
| |
| void *newCell = fastMalloc(s); |
| heap.oversizeCells[usedOversizeCells] = static_cast<CollectorCell *>(newCell); |
| heap.usedOversizeCells = usedOversizeCells + 1; |
| heap.numLiveObjects = numLiveObjects + 1; |
| |
| return newCell; |
| } |
| |
| // slab allocator |
| |
| size_t usedBlocks = heap.usedBlocks; |
| |
| size_t i = heap.firstBlockWithPossibleSpace; |
| CollectorBlock *targetBlock; |
| size_t targetBlockUsedCells; |
| if (i != usedBlocks) { |
| targetBlock = heap.blocks[i]; |
| targetBlockUsedCells = targetBlock->usedCells; |
| assert(targetBlockUsedCells <= CELLS_PER_BLOCK); |
| while (targetBlockUsedCells == CELLS_PER_BLOCK) { |
| if (++i == usedBlocks) |
| goto allocateNewBlock; |
| targetBlock = heap.blocks[i]; |
| targetBlockUsedCells = targetBlock->usedCells; |
| assert(targetBlockUsedCells <= CELLS_PER_BLOCK); |
| } |
| heap.firstBlockWithPossibleSpace = i; |
| } else { |
| allocateNewBlock: |
| // didn't find one, need to allocate a new block |
| size_t numBlocks = heap.numBlocks; |
| if (usedBlocks == numBlocks) { |
| numBlocks = max(MIN_ARRAY_SIZE, numBlocks * GROWTH_FACTOR); |
| heap.numBlocks = numBlocks; |
| heap.blocks = static_cast<CollectorBlock **>(fastRealloc(heap.blocks, numBlocks * sizeof(CollectorBlock *))); |
| } |
| |
| targetBlock = static_cast<CollectorBlock *>(fastCalloc(1, sizeof(CollectorBlock))); |
| targetBlock->freeList = targetBlock->cells; |
| targetBlockUsedCells = 0; |
| heap.blocks[usedBlocks] = targetBlock; |
| heap.usedBlocks = usedBlocks + 1; |
| heap.firstBlockWithPossibleSpace = usedBlocks; |
| } |
| |
| // find a free spot in the block and detach it from the free list |
| CollectorCell *newCell = targetBlock->freeList; |
| |
| // "next" field is a byte offset -- 0 means next cell, so a zeroed block is already initialized |
| // could avoid the casts by using a cell offset, but this avoids a relatively-slow multiply |
| targetBlock->freeList = reinterpret_cast<CollectorCell *>(reinterpret_cast<char *>(newCell + 1) + newCell->u.freeCell.next); |
| |
| targetBlock->usedCells = static_cast<uint32_t>(targetBlockUsedCells + 1); |
| heap.numLiveObjects = numLiveObjects + 1; |
| |
| return newCell; |
| } |
| |
| #if USE(MULTIPLE_THREADS) |
| |
| struct Collector::Thread { |
| Thread(pthread_t pthread, mach_port_t mthread) : posixThread(pthread), machThread(mthread) {} |
| Thread *next; |
| pthread_t posixThread; |
| mach_port_t machThread; |
| }; |
| |
| pthread_key_t registeredThreadKey; |
| pthread_once_t registeredThreadKeyOnce = PTHREAD_ONCE_INIT; |
| Collector::Thread *registeredThreads; |
| |
| static void destroyRegisteredThread(void *data) |
| { |
| Collector::Thread *thread = (Collector::Thread *)data; |
| |
| if (registeredThreads == thread) { |
| registeredThreads = registeredThreads->next; |
| } else { |
| Collector::Thread *last = registeredThreads; |
| for (Collector::Thread *t = registeredThreads->next; t != NULL; t = t->next) { |
| if (t == thread) { |
| last->next = t->next; |
| break; |
| } |
| last = t; |
| } |
| } |
| |
| delete thread; |
| } |
| |
| static void initializeRegisteredThreadKey() |
| { |
| pthread_key_create(®isteredThreadKey, destroyRegisteredThread); |
| } |
| |
| void Collector::registerThread() |
| { |
| pthread_once(®isteredThreadKeyOnce, initializeRegisteredThreadKey); |
| |
| if (!pthread_getspecific(registeredThreadKey)) { |
| pthread_t pthread = pthread_self(); |
| WTF::fastMallocRegisterThread(pthread); |
| Collector::Thread *thread = new Collector::Thread(pthread, pthread_mach_thread_np(pthread)); |
| thread->next = registeredThreads; |
| registeredThreads = thread; |
| pthread_setspecific(registeredThreadKey, thread); |
| } |
| } |
| |
| #endif |
| |
| #define IS_POINTER_ALIGNED(p) (((intptr_t)(p) & (sizeof(char *) - 1)) == 0) |
| |
| // cells are 8-byte aligned |
| #define IS_CELL_ALIGNED(p) (((intptr_t)(p) & 7) == 0) |
| |
| void Collector::markStackObjectsConservatively(void *start, void *end) |
| { |
| if (start > end) { |
| void *tmp = start; |
| start = end; |
| end = tmp; |
| } |
| |
| assert(((char *)end - (char *)start) < 0x1000000); |
| assert(IS_POINTER_ALIGNED(start)); |
| assert(IS_POINTER_ALIGNED(end)); |
| |
| char **p = (char **)start; |
| char **e = (char **)end; |
| |
| size_t usedBlocks = heap.usedBlocks; |
| CollectorBlock **blocks = heap.blocks; |
| size_t usedOversizeCells = heap.usedOversizeCells; |
| CollectorCell **oversizeCells = heap.oversizeCells; |
| |
| const size_t lastCellOffset = sizeof(CollectorCell) * (CELLS_PER_BLOCK - 1); |
| |
| while (p != e) { |
| char *x = *p++; |
| if (IS_CELL_ALIGNED(x) && x) { |
| for (size_t block = 0; block < usedBlocks; block++) { |
| size_t offset = x - reinterpret_cast<char *>(blocks[block]); |
| if (offset <= lastCellOffset && offset % sizeof(CollectorCell) == 0) |
| goto gotGoodPointer; |
| } |
| for (size_t i = 0; i != usedOversizeCells; i++) |
| if (x == reinterpret_cast<char *>(oversizeCells[i])) |
| goto gotGoodPointer; |
| continue; |
| |
| gotGoodPointer: |
| if (((CollectorCell *)x)->u.freeCell.zeroIfFree != 0) { |
| JSCell *imp = reinterpret_cast<JSCell *>(x); |
| if (!imp->marked()) |
| imp->mark(); |
| } |
| } |
| } |
| } |
| |
| void Collector::markCurrentThreadConservatively() |
| { |
| jmp_buf registers; |
| #if COMPILER(MSVC) |
| #pragma warning(push) |
| #pragma warning(disable: 4611) |
| #endif |
| setjmp(registers); |
| #if COMPILER(MSVC) |
| #pragma warning(pop) |
| #endif |
| |
| #if PLATFORM(DARWIN) |
| pthread_t thread = pthread_self(); |
| void *stackBase = pthread_get_stackaddr_np(thread); |
| #elif PLATFORM(WIN_OS) && PLATFORM(X86) && COMPILER(MSVC) |
| NT_TIB *pTib; |
| __asm { |
| MOV EAX, FS:[18h] |
| MOV pTib, EAX |
| } |
| void *stackBase = (void *)pTib->StackBase; |
| #elif PLATFORM(UNIX) |
| static void *stackBase = 0; |
| static pthread_t stackThread; |
| pthread_t thread = pthread_self(); |
| if (stackBase == 0 || thread != stackThread) { |
| pthread_attr_t sattr; |
| #if HAVE(PTHREAD_NP_H) |
| // e.g. on FreeBSD 5.4, neundorf@kde.org |
| pthread_attr_get_np(thread, &sattr); |
| #else |
| // FIXME: this function is non-portable; other POSIX systems may have different np alternatives |
| pthread_getattr_np(thread, &sattr); |
| #endif |
| size_t stackSize; |
| int rc = pthread_attr_getstack(&sattr, &stackBase, &stackSize); |
| (void)rc; // FIXME: deal with error code somehow? seems fatal... |
| assert(stackBase); |
| stackBase = (void*)(size_t(stackBase) + stackSize); |
| stackThread = thread; |
| } |
| #else |
| #error Need a way to get the stack base on this platform |
| #endif |
| |
| void *dummy; |
| void *stackPointer = &dummy; |
| |
| markStackObjectsConservatively(stackPointer, stackBase); |
| } |
| |
| #if USE(MULTIPLE_THREADS) |
| |
| typedef unsigned long usword_t; // word size, assumed to be either 32 or 64 bit |
| |
| void Collector::markOtherThreadConservatively(Thread *thread) |
| { |
| thread_suspend(thread->machThread); |
| |
| #if PLATFORM(X86) |
| i386_thread_state_t regs; |
| unsigned user_count = sizeof(regs)/sizeof(int); |
| thread_state_flavor_t flavor = i386_THREAD_STATE; |
| #elif PLATFORM(X86_64) |
| x86_thread_state64_t regs; |
| unsigned user_count = x86_THREAD_STATE64_COUNT; |
| thread_state_flavor_t flavor = x86_THREAD_STATE64; |
| #elif PLATFORM(PPC) |
| ppc_thread_state_t regs; |
| unsigned user_count = PPC_THREAD_STATE_COUNT; |
| thread_state_flavor_t flavor = PPC_THREAD_STATE; |
| #elif PLATFORM(PPC64) |
| ppc_thread_state64_t regs; |
| unsigned user_count = PPC_THREAD_STATE64_COUNT; |
| thread_state_flavor_t flavor = PPC_THREAD_STATE64; |
| #else |
| #error Unknown Architecture |
| #endif |
| // get the thread register state |
| thread_get_state(thread->machThread, flavor, (thread_state_t)®s, &user_count); |
| |
| // scan the registers |
| markStackObjectsConservatively((void *)®s, (void *)((char *)®s + (user_count * sizeof(usword_t)))); |
| |
| // scan the stack |
| #if PLATFORM(X86) && __DARWIN_UNIX03 |
| markStackObjectsConservatively((void *)regs.__esp, pthread_get_stackaddr_np(thread->posixThread)); |
| #elif PLATFORM(X86) |
| markStackObjectsConservatively((void *)regs.esp, pthread_get_stackaddr_np(thread->posixThread)); |
| #elif PLATFORM(X86_64) && __DARWIN_UNIX03 |
| markStackObjectsConservatively((void *)regs.__rsp, pthread_get_stackaddr_np(thread->posixThread)); |
| #elif PLATFORM(X86_64) |
| markStackObjectsConservatively((void *)regs.rsp, pthread_get_stackaddr_np(thread->posixThread)); |
| #elif (PLATFORM(PPC) || PLATFORM(PPC64)) && __DARWIN_UNIX03 |
| markStackObjectsConservatively((void *)regs.__r1, pthread_get_stackaddr_np(thread->posixThread)); |
| #elif PLATFORM(PPC) || PLATFORM(PPC64) |
| markStackObjectsConservatively((void *)regs.r1, pthread_get_stackaddr_np(thread->posixThread)); |
| #else |
| #error Unknown Architecture |
| #endif |
| |
| thread_resume(thread->machThread); |
| } |
| |
| #endif |
| |
| void Collector::markStackObjectsConservatively() |
| { |
| markCurrentThreadConservatively(); |
| |
| #if USE(MULTIPLE_THREADS) |
| for (Thread *thread = registeredThreads; thread != NULL; thread = thread->next) { |
| if (thread->posixThread != pthread_self()) { |
| markOtherThreadConservatively(thread); |
| } |
| } |
| #endif |
| } |
| |
| typedef HashCountedSet<JSCell *> ProtectCounts; |
| |
| static ProtectCounts& protectedValues() |
| { |
| static ProtectCounts pv; |
| return pv; |
| } |
| |
| void Collector::protect(JSValue *k) |
| { |
| assert(k); |
| assert(JSLock::lockCount() > 0); |
| |
| if (JSImmediate::isImmediate(k)) |
| return; |
| |
| protectedValues().add(k->downcast()); |
| } |
| |
| void Collector::unprotect(JSValue *k) |
| { |
| assert(k); |
| assert(JSLock::lockCount() > 0); |
| |
| if (JSImmediate::isImmediate(k)) |
| return; |
| |
| protectedValues().remove(k->downcast()); |
| } |
| |
| void Collector::markProtectedObjects() |
| { |
| ProtectCounts& pv = protectedValues(); |
| ProtectCounts::iterator end = pv.end(); |
| for (ProtectCounts::iterator it = pv.begin(); it != end; ++it) { |
| JSCell *val = it->first; |
| if (!val->marked()) |
| val->mark(); |
| } |
| } |
| |
| bool Collector::collect() |
| { |
| assert(JSLock::lockCount() > 0); |
| |
| #if USE(MULTIPLE_THREADS) |
| bool currentThreadIsMainThread = !pthread_is_threaded_np() || pthread_main_np(); |
| #else |
| bool currentThreadIsMainThread = true; |
| #endif |
| |
| if (Interpreter::s_hook) { |
| Interpreter* scr = Interpreter::s_hook; |
| do { |
| scr->mark(currentThreadIsMainThread); |
| scr = scr->next; |
| } while (scr != Interpreter::s_hook); |
| } |
| |
| // MARK: first mark all referenced objects recursively starting out from the set of root objects |
| |
| markStackObjectsConservatively(); |
| markProtectedObjects(); |
| List::markProtectedLists(); |
| |
| // SWEEP: delete everything with a zero refcount (garbage) and unmark everything else |
| |
| size_t emptyBlocks = 0; |
| size_t numLiveObjects = heap.numLiveObjects; |
| |
| for (size_t block = 0; block < heap.usedBlocks; block++) { |
| CollectorBlock *curBlock = heap.blocks[block]; |
| |
| size_t usedCells = curBlock->usedCells; |
| CollectorCell *freeList = curBlock->freeList; |
| |
| if (usedCells == CELLS_PER_BLOCK) { |
| // special case with a block where all cells are used -- testing indicates this happens often |
| for (size_t i = 0; i < CELLS_PER_BLOCK; i++) { |
| CollectorCell *cell = curBlock->cells + i; |
| JSCell *imp = reinterpret_cast<JSCell *>(cell); |
| if (imp->m_marked) { |
| imp->m_marked = false; |
| } else if (currentThreadIsMainThread || imp->m_destructorIsThreadSafe) { |
| // special case for allocated but uninitialized object |
| // (We don't need this check earlier because nothing prior this point assumes the object has a valid vptr.) |
| if (cell->u.freeCell.zeroIfFree == 0) |
| continue; |
| |
| imp->~JSCell(); |
| --usedCells; |
| --numLiveObjects; |
| |
| // put cell on the free list |
| cell->u.freeCell.zeroIfFree = 0; |
| cell->u.freeCell.next = reinterpret_cast<char *>(freeList) - reinterpret_cast<char *>(cell + 1); |
| freeList = cell; |
| } |
| } |
| } else { |
| size_t minimumCellsToProcess = usedCells; |
| for (size_t i = 0; (i < minimumCellsToProcess) & (i < CELLS_PER_BLOCK); i++) { |
| CollectorCell *cell = curBlock->cells + i; |
| if (cell->u.freeCell.zeroIfFree == 0) { |
| ++minimumCellsToProcess; |
| } else { |
| JSCell *imp = reinterpret_cast<JSCell *>(cell); |
| if (imp->m_marked) { |
| imp->m_marked = false; |
| } else if (currentThreadIsMainThread || imp->m_destructorIsThreadSafe) { |
| imp->~JSCell(); |
| --usedCells; |
| --numLiveObjects; |
| |
| // put cell on the free list |
| cell->u.freeCell.zeroIfFree = 0; |
| cell->u.freeCell.next = reinterpret_cast<char *>(freeList) - reinterpret_cast<char *>(cell + 1); |
| freeList = cell; |
| } |
| } |
| } |
| } |
| |
| curBlock->usedCells = static_cast<uint32_t>(usedCells); |
| curBlock->freeList = freeList; |
| |
| if (usedCells == 0) { |
| emptyBlocks++; |
| if (emptyBlocks > SPARE_EMPTY_BLOCKS) { |
| #if !DEBUG_COLLECTOR |
| fastFree(curBlock); |
| #endif |
| // swap with the last block so we compact as we go |
| heap.blocks[block] = heap.blocks[heap.usedBlocks - 1]; |
| heap.usedBlocks--; |
| block--; // Don't move forward a step in this case |
| |
| if (heap.numBlocks > MIN_ARRAY_SIZE && heap.usedBlocks < heap.numBlocks / LOW_WATER_FACTOR) { |
| heap.numBlocks = heap.numBlocks / GROWTH_FACTOR; |
| heap.blocks = (CollectorBlock **)fastRealloc(heap.blocks, heap.numBlocks * sizeof(CollectorBlock *)); |
| } |
| } |
| } |
| } |
| |
| if (heap.numLiveObjects != numLiveObjects) |
| heap.firstBlockWithPossibleSpace = 0; |
| |
| size_t cell = 0; |
| while (cell < heap.usedOversizeCells) { |
| JSCell *imp = (JSCell *)heap.oversizeCells[cell]; |
| |
| if (!imp->m_marked && (currentThreadIsMainThread || imp->m_destructorIsThreadSafe)) { |
| imp->~JSCell(); |
| #if DEBUG_COLLECTOR |
| heap.oversizeCells[cell]->u.freeCell.zeroIfFree = 0; |
| #else |
| fastFree(imp); |
| #endif |
| |
| // swap with the last oversize cell so we compact as we go |
| heap.oversizeCells[cell] = heap.oversizeCells[heap.usedOversizeCells - 1]; |
| |
| heap.usedOversizeCells--; |
| numLiveObjects--; |
| |
| if (heap.numOversizeCells > MIN_ARRAY_SIZE && heap.usedOversizeCells < heap.numOversizeCells / LOW_WATER_FACTOR) { |
| heap.numOversizeCells = heap.numOversizeCells / GROWTH_FACTOR; |
| heap.oversizeCells = (CollectorCell **)fastRealloc(heap.oversizeCells, heap.numOversizeCells * sizeof(CollectorCell *)); |
| } |
| } else { |
| imp->m_marked = false; |
| cell++; |
| } |
| } |
| |
| bool deleted = heap.numLiveObjects != numLiveObjects; |
| |
| heap.numLiveObjects = numLiveObjects; |
| heap.numLiveObjectsAtLastCollect = numLiveObjects; |
| |
| memoryFull = (numLiveObjects >= KJS_MEM_LIMIT); |
| |
| return deleted; |
| } |
| |
| size_t Collector::size() |
| { |
| return heap.numLiveObjects; |
| } |
| |
| #ifdef KJS_DEBUG_MEM |
| void Collector::finalCheck() |
| { |
| } |
| #endif |
| |
| size_t Collector::numInterpreters() |
| { |
| size_t count = 0; |
| if (Interpreter::s_hook) { |
| Interpreter* scr = Interpreter::s_hook; |
| do { |
| ++count; |
| scr = scr->next; |
| } while (scr != Interpreter::s_hook); |
| } |
| return count; |
| } |
| |
| size_t Collector::numProtectedObjects() |
| { |
| return protectedValues().size(); |
| } |
| |
| static const char *typeName(JSCell *val) |
| { |
| const char *name = "???"; |
| switch (val->type()) { |
| case UnspecifiedType: |
| break; |
| case UndefinedType: |
| name = "undefined"; |
| break; |
| case NullType: |
| name = "null"; |
| break; |
| case BooleanType: |
| name = "boolean"; |
| break; |
| case StringType: |
| name = "string"; |
| break; |
| case NumberType: |
| name = "number"; |
| break; |
| case ObjectType: { |
| const ClassInfo *info = static_cast<JSObject *>(val)->classInfo(); |
| name = info ? info->className : "Object"; |
| break; |
| } |
| case GetterSetterType: |
| name = "gettersetter"; |
| break; |
| } |
| return name; |
| } |
| |
| HashCountedSet<const char*>* Collector::rootObjectTypeCounts() |
| { |
| HashCountedSet<const char*>* counts = new HashCountedSet<const char*>; |
| |
| ProtectCounts& pv = protectedValues(); |
| ProtectCounts::iterator end = pv.end(); |
| for (ProtectCounts::iterator it = pv.begin(); it != end; ++it) |
| counts->add(typeName(it->first)); |
| |
| return counts; |
| } |
| |
| } // namespace KJS |