/haiku/src/add-ons/kernel/file_systems/packagefs/util/ |
H A D | ClassCache.h | 9 #include <slab/Slab.h>
|
H A D | TwoKeyAVLTree.h | 9 #include <slab/Slab.h>
|
/haiku/src/add-ons/kernel/network/stack/ |
H A D | simple_net_buffer.cpp | 15 #include <slab/Slab.h>
|
H A D | net_buffer.cpp | 14 #include <slab/Slab.h> 2313 // and keep around half-constructed buffers in the slab
|
/haiku/src/system/kernel/arch/ppc/ |
H A D | arch_vm_translation_map.cpp | 86 #include <slab/Slab.h>
|
/haiku/src/system/kernel/cache/ |
H A D | vnode_store.cpp | 13 #include <slab/Slab.h>
|
/haiku/src/system/kernel/slab/ |
H A D | HashedObjectCache.cpp | 39 free_slab(HashedSlab* slab, uint32 flags) argument 41 slab_internal_free(slab, flags); 105 slab* 113 HashedSlab* slab = allocate_slab(flags); local 114 if (slab != NULL) { 117 && AllocateTrackingInfos(slab, slab_size, flags) == B_OK) { 119 if (InitSlab(slab, pages, slab_size, flags)) { 120 hash_table.InsertUnchecked(slab); 122 return slab; 125 FreeTrackingInfos(slab, flag 142 HashedSlab* slab = static_cast<HashedSlab*>(_slab); local 162 HashedSlab* slab = hash_table.Lookup(::lower_boundary(object, slab_size)); local [all...] |
H A D | HashedObjectCache.h | 17 struct HashedSlab : slab { 35 virtual slab* CreateSlab(uint32 flags); 36 virtual void ReturnSlab(slab* slab, uint32 flags); 37 virtual slab* ObjectSlab(void* object) const;
|
H A D | ObjectCache.h | 13 #include <slab/ObjectDepot.h> 14 #include <slab/Slab.h> 28 struct slab : DoublyLinkedListLinkImpl<slab> { struct in inherits:DoublyLinkedListLinkImpl 39 typedef DoublyLinkedList<slab> SlabList; 98 virtual slab* CreateSlab(uint32 flags) = 0; 99 virtual void ReturnSlab(slab* slab, uint32 flags) = 0; 100 virtual slab* ObjectSlab(void* object) const = 0; 102 slab* InitSla 168 AllocateTrackingInfos(slab* slab, size_t byteCount, uint32 flags) argument 175 FreeTrackingInfos(slab* slab, uint32 flags) argument [all...] |
H A D | ObjectDepot.cpp | 10 #include <slab/ObjectDepot.h> 15 #include <slab/Slab.h> 288 // we return the object directly to the slab.
|
H A D | SmallObjectCache.cpp | 19 static inline slab * 22 BytePointer<slab> pointer(pages); 23 pointer += slab_size - sizeof(slab); 67 slab* 82 slab* newSlab = slab_in_pages(pages, slab_size); 83 size_t byteCount = slab_size - sizeof(slab); 94 SmallObjectCache::ReturnSlab(slab* slab, uint32 flags) argument 96 UninitSlab(slab); 99 FreeTrackingInfos(slab, flag [all...] |
H A D | SmallObjectCache.h | 25 virtual slab* CreateSlab(uint32 flags); 26 virtual void ReturnSlab(slab* slab, uint32 flags); 27 virtual slab* ObjectSlab(void* object) const;
|
H A D | slab_debug.h | 13 #include <slab/Slab.h>
|
H A D | slab_private.h | 14 #include <slab/Slab.h>
|
/haiku/src/system/kernel/vm/ |
H A D | VMAnonymousNoSwapCache.cpp | 17 #include <slab/Slab.h>
|
H A D | VMDeviceCache.cpp | 12 #include <slab/Slab.h>
|
H A D | VMKernelArea.cpp | 10 #include <slab/Slab.h>
|
H A D | VMNullCache.cpp | 9 #include <slab/Slab.h>
|
H A D | VMTranslationMap.cpp | 9 #include <slab/Slab.h>
|
/haiku/src/tests/add-ons/kernel/kernelland_emu/ |
H A D | slab.cpp | 6 #include <slab/Slab.h>
|
/haiku/src/tests/system/kernel/slab/ |
H A D | Slab.cpp | 106 Slab *slab = fPartialSlabs.Head(); local 109 slab, slab->count); 111 ObjectLink *link = SListPop(slab->free); 112 slab->count--; 113 if (slab->count == 0) { 114 // move the partial slab to the full list 115 fPartialSlabs.Remove(slab); 116 fFullSlabs.Add(slab); 126 Slab *slab local 146 ConstructSlab(Slab *slab, void *pages, size_t byteCount, ObjectLink *(*getLink)(void *parent, void *object), void *parent) argument 181 DestructSlab(Slab *slab) argument [all...] |
H A D | Slab.h | 71 Slab *ConstructSlab(Slab *slab, void *pages, size_t byteCount, 73 void DestructSlab(Slab *slab); 179 BaseCache::Slab *_ConstructSlab(Slab *slab, void *pages, size_t tailSpace, argument 182 return fParent->ConstructSlab(slab, pages, SlabSize(tailSpace) 186 void _DestructSlab(BaseCache::Slab *slab) argument 188 fParent->DestructSlab(slab); 189 Backend::FreePages(fParent, ((Slab *)slab)->id); 196 // This slab strategy includes the ObjectLink at the end of each object and the 197 // slab at the end of the allocated pages. It uses aligned allocations to 198 // provide object to slab mappin 222 Slab *slab = _SlabInPages(LowerBoundary(object, _SlabSize())); local 244 ReturnSlab(BaseCache::Slab *slab) argument 320 BaseCache::Slab *slab; member in struct:BaseHashCacheStrategy::Link 401 Slab *slab = fSlabCache.Alloc(flags); local 424 ReturnSlab(BaseCache::Slab *slab) argument 437 _PrepareSlab(BaseCache *parent, Slab *slab, void *pages, size_t byteCount, uint32_t flags) argument [all...] |
/haiku/src/add-ons/kernel/file_systems/packagefs/ |
H A D | kernel_interface.cpp | 17 #include <slab/Slab.h>
|
/haiku/src/kits/package/hpkg/ |
H A D | PackageFileHeapAccessorBase.cpp | 15 #include <slab/Slab.h>
|
/haiku/src/libs/compat/freebsd_network/ |
H A D | mbuf.c | 12 #include <slab/Slab.h>
|