Lines Matching refs:flags

37 	uint8				flags;
101 Allocate(guarded_heap* heap, void* pageBase, uint32 flags)
105 fFlags(flags)
113 "flags:%s%s%s%s", fHeap, fPageBase,
181 page.flags = GUARDED_HEAP_PAGE_FLAG_USED;
193 page.flags |= GUARDED_HEAP_PAGE_FLAG_FIRST;
209 page.flags |= GUARDED_HEAP_PAGE_FLAG_GUARD;
218 page.flags));
231 page.flags = 0;
233 page.flags |= GUARDED_HEAP_PAGE_FLAG_DEAD;
235 page.flags = 0;
266 size_t alignment, uint32 flags, bool& grow)
289 if ((page->flags & GUARDED_HEAP_PAGE_FLAG_USED) != 0)
300 if ((area.pages[pageIndex + j].flags & GUARDED_HEAP_PAGE_FLAG_USED)
331 size_t size, uint32 flags)
371 guarded_heap_area_create(guarded_heap& heap, uint32 flags)
384 if (guarded_heap_area_init(heap, id, baseAddress, trySize, flags))
396 guarded_heap_add_area(guarded_heap& heap, int32 counter, uint32 flags)
398 if ((flags & (HEAP_DONT_LOCK_KERNEL_SPACE | HEAP_DONT_WAIT_FOR_MEMORY))
405 return guarded_heap_area_create(heap, flags);
414 uint32 flags)
423 result = guarded_heap_area_allocate(*area, size, alignment, flags,
433 bool added = guarded_heap_add_area(heap, counter, flags);
435 return guarded_heap_allocate(heap, size, alignment, flags);
471 if ((page.flags & GUARDED_HEAP_PAGE_FLAG_USED) == 0) {
477 if ((page.flags & GUARDED_HEAP_PAGE_FLAG_GUARD) != 0) {
483 if ((page.flags & GUARDED_HEAP_PAGE_FLAG_FIRST) == 0) {
489 if ((page.flags & GUARDED_HEAP_PAGE_FLAG_DEAD) != 0) {
500 guarded_heap_area_free(guarded_heap_area& area, void* address, uint32 flags)
508 while ((page->flags & GUARDED_HEAP_PAGE_FLAG_GUARD) == 0) {
529 guarded_heap_free(void* address, uint32 flags)
540 guarded_heap_area_free(*area, address, flags);
545 guarded_heap_realloc(void* address, size_t newSize, uint32 flags)
565 void* newBlock = malloc_etc(newSize, flags);
571 free_etc(address, flags);
638 kprintf("flags:");
639 if ((page.flags & GUARDED_HEAP_PAGE_FLAG_USED) != 0)
641 if ((page.flags & GUARDED_HEAP_PAGE_FLAG_FIRST) != 0)
643 if ((page.flags & GUARDED_HEAP_PAGE_FLAG_GUARD) != 0)
645 if ((page.flags & GUARDED_HEAP_PAGE_FLAG_DEAD) != 0)
707 if ((((guarded_heap_page*)item)->flags & GUARDED_HEAP_PAGE_FLAG_USED)
724 || (page.flags & GUARDED_HEAP_PAGE_FLAG_USED) != 0) {
734 if ((area->pages[i + j].flags & GUARDED_HEAP_PAGE_FLAG_USED) != 0)
820 if ((page.flags & GUARDED_HEAP_PAGE_FLAG_FIRST) == 0)
887 if ((page.flags & GUARDED_HEAP_PAGE_FLAG_USED) != 0
888 && (page.flags & GUARDED_HEAP_PAGE_FLAG_GUARD) == 0
889 && (page.flags & GUARDED_HEAP_PAGE_FLAG_DEAD) == 0) {
938 memalign_etc(size_t alignment, size_t size, uint32 flags)
943 return guarded_heap_allocate(sGuardedHeap, size, alignment, flags);
948 free_etc(void *address, uint32 flags)
950 guarded_heap_free(address, flags);
969 realloc_etc(void* address, size_t newSize, uint32 flags)
972 free_etc(address, flags);
977 return malloc_etc(newSize, flags);
979 return guarded_heap_realloc(address, newSize, flags);
1033 object_cache_alloc(object_cache* cache, uint32 flags)
1035 return memalign_etc(0, (size_t)cache, flags);
1040 object_cache_free(object_cache* cache, void* object, uint32 flags)
1042 return free_etc(object, flags);
1047 object_cache_reserve(object_cache* cache, size_t objectCount, uint32 flags)