Lines Matching refs:flags

85 	Allocate(ObjectCache* cache, uint32 flags)
89 fFlags(flags)
96 out.Print("slab memory manager alloc: cache: %p, flags: %#" B_PRIx32,
108 Free(void* address, uint32 flags)
112 fFlags(flags)
119 out.Print("slab memory manager free: address: %p, flags: %#" B_PRIx32,
131 AllocateRaw(size_t size, uint32 flags)
135 fFlags(flags)
143 ", flags: %#" B_PRIx32, fSize, fFlags);
154 FreeRawOrReturnCache(void* address, uint32 flags)
158 fFlags(flags)
165 out.Print("slab memory manager free raw/return: address: %p, flags: %#"
177 AllocateArea(Area* area, uint32 flags)
181 fFlags(flags)
188 out.Print("slab memory manager alloc area: flags: %#" B_PRIx32
220 FreeArea(Area* area, bool areaRemoved, uint32 flags)
224 fFlags(flags),
232 out.Print("slab memory manager free area: %p%s, flags: %#" B_PRIx32,
365 Map(addr_t address, size_t size, uint32 flags)
370 fFlags(flags)
378 B_PRIuSIZE ", flags: %#" B_PRIx32, fAddress, fSize, fFlags);
390 Unmap(addr_t address, size_t size, uint32 flags)
395 fFlags(flags)
403 B_PRIuSIZE ", flags: %#" B_PRIx32, fAddress, fSize, fFlags);
533 MemoryManager::Allocate(ObjectCache* cache, uint32 flags, void*& _pages)
537 T(Allocate(cache, flags));
542 B_PRIuSIZE "\n", cache, flags, chunkSize);
549 status_t error = _AllocateChunks(chunkSize, 1, flags, metaChunk, chunk);
558 error = _MapChunk(area->vmArea, chunkAddress, chunkSize, 0, flags);
562 _FreeChunk(area, metaChunk, chunk, chunkAddress, true, flags);
577 MemoryManager::Free(void* pages, uint32 flags)
579 TRACE("MemoryManager::Free(%p, %#" B_PRIx32 ")\n", pages, flags);
581 T(Free(pages, flags));
603 _FreeChunk(area, metaChunk, chunk, (addr_t)pages, false, flags);
608 MemoryManager::AllocateRaw(size_t size, uint32 flags, void*& _pages)
612 AbstractTraceEntryWithStackTrace* traceEntry = T(AllocateRaw(size, flags));
615 T(AllocateRaw(size, flags));
622 flags);
624 if (size > SLAB_CHUNK_SIZE_LARGE || (flags & CACHE_ALIGN_ON_SIZE) != 0) {
627 if ((flags & CACHE_DONT_LOCK_KERNEL_SPACE) != 0)
632 = (flags & CACHE_ALIGN_ON_SIZE) != 0
638 ((flags & CACHE_DONT_WAIT_FOR_MEMORY) != 0
668 status_t error = _AllocateChunks(chunkSize, chunkCount, flags, metaChunk,
678 error = _MapChunk(area->vmArea, chunkAddress, size, 0, flags);
683 _FreeChunk(area, metaChunk, chunk + i, chunkAddress, true, flags);
703 MemoryManager::FreeRawOrReturnCache(void* pages, uint32 flags)
706 flags);
708 T(FreeRawOrReturnCache(pages, flags));
710 if ((flags & CACHE_DONT_LOCK_KERNEL_SPACE) != 0) {
758 _UnmapChunk(area->vmArea, (addr_t)pages, size, flags);
764 _FreeChunk(area, metaChunk, chunk + i, (addr_t)pages, true, flags);
954 uint32 flags, MetaChunk*& _metaChunk, Chunk*& _chunk)
978 if ((flags & CACHE_DONT_LOCK_KERNEL_SPACE) != 0) {
991 && (flags & CACHE_DONT_WAIT_FOR_MEMORY) == 0) {
1006 = (flags & CACHE_DONT_WAIT_FOR_MEMORY) != 0
1015 status_t error = _AllocateArea(flags, area);
1026 _FreeArea(area, true, flags);
1177 addr_t chunkAddress, bool alreadyUnmapped, uint32 flags)
1182 _UnmapChunk(area->vmArea, chunkAddress, metaChunk->chunkSize, flags);
1216 _FreeArea(area, false, flags);
1303 MemoryManager::_AllocateArea(uint32 flags, Area*& _area)
1305 TRACE("MemoryManager::_AllocateArea(%#" B_PRIx32 ")\n", flags);
1307 ASSERT((flags & CACHE_DONT_LOCK_KERNEL_SPACE) == 0);
1318 uint32 areaCreationFlags = (flags & CACHE_PRIORITY_VIP) != 0
1339 pagesNeededToMap, flags);
1385 T(AllocateArea(area, flags));
1392 MemoryManager::_FreeArea(Area* area, bool areaRemoved, uint32 flags)
1394 TRACE("MemoryManager::_FreeArea(%p, %#" B_PRIx32 ")\n", area, flags);
1396 T(FreeArea(area, areaRemoved, flags));
1422 if (area->vmArea == NULL || (flags & CACHE_DONT_LOCK_KERNEL_SPACE) != 0) {
1445 size_t reserveAdditionalMemory, uint32 flags)
1450 T(Map(address, size, flags));
1461 int priority = (flags & CACHE_PRIORITY_VIP) != 0
1465 (flags & CACHE_DONT_WAIT_FOR_MEMORY) != 0 ? 0 : 1000000);
1473 if ((flags & CACHE_DONT_WAIT_FOR_MEMORY) != 0) {
1515 uint32 flags)
1517 T(Unmap(address, size, flags));