Lines Matching refs:base

53 /*!	Verifies that an area with the given aligned base and size fits into
54 the spot defined by base and limit and checks for overflows.
55 \param base Minimum base address valid for the area.
56 \param alignedBase The base address of the range to check.
61 is_valid_spot(addr_t base, addr_t alignedBase, addr_t size, addr_t limit)
63 return (alignedBase >= base && alignedBase + (size - 1) > alignedBase
71 VMKernelAddressSpace::VMKernelAddressSpace(team_id id, addr_t base, size_t size)
73 VMAddressSpace(id, base, size, "kernel address space"),
187 It will also set the "_address" argument to its base address when
212 area->SetBase(range->base);
254 && nextRange->reserved.base > range->base) {
284 nextRange->base = range->base + newSize;
290 range->base + newSize, range->size - newSize,
299 && nextRange->reserved.base > range->base)) {
317 nextRange->base = range->base + newSize;
356 range->base += sizeDiff;
362 range->base, sizeDiff, Range::RANGE_FREE);
365 range->base += sizeDiff;
370 area->SetBase(range->base);
408 range->reserved.base = range->base;
412 *_address = (void*)range->base;
435 while (range != NULL && range->base + (range->size - 1) <= endAddress) {
505 B_PRIxSIZE " flags = %#" B_PRIx32 "\n", range->base,
506 range->reserved.base, range->size, range->reserved.flags);
511 B_PRIxSIZE "\n", range->base, range->size);
522 ", %#" B_PRIxSIZE ", %d), %#" B_PRIxSIZE " (%d))\n", range, range->base,
533 ", %#" B_PRIxSIZE ", %d), %#" B_PRIxSIZE " (%d))\n", range, range->base,
544 B_PRIxSIZE ", %d))\n", range, range->base, range->size, range->type);
547 Range* insertBeforeRange = fRangeTree.FindClosest(range->base, true);
566 B_PRIxSIZE ", %d))\n", range, range->base, range->size, range->type);
590 // prepare size, alignment and the base address for the range search
634 B_PRIxADDR ", %#" B_PRIxSIZE ", %d)\n", range, range->base, range->size,
641 if (address == range->base) {
653 } else if (address + size == range->base + range->size) {
656 range->base, range->size - size, range);
660 range->base = address;
666 range->base, address - range->base, range);
677 range->base = address;
690 B_PRIxSIZE ")\n", range, range->base, range->size);
715 addr_t alignedBase = ROUNDUP(range->base, alignment);
717 range->base + (range->size - 1))) {
738 // unless there are constraints (min base address, alignment).
748 addr_t alignedBase = ROUNDUP(range->base, alignment);
750 range->base + (range->size - 1))) {
766 addr_t alignedBase = ROUNDUP(range->base, alignment);
768 range->base + (range->size - 1))) {
770 // -- adjust the base accordingly
774 range->base + (range->size - size), alignment);
792 || range->base + (range->size - 1) < start + (size - 1)) {
819 B_PRIxSIZE ", %d))\n", range, range->base, range->size, range->type);
888 if (range->base != nextBase) {
889 panic("VMKernelAddressSpace::_CheckStructures(): range base %#"
890 B_PRIxADDR ", expected: %#" B_PRIxADDR, range->base, nextBase);
901 range->base, range->size);
904 if (range->size > spaceSize - (range->base - fBase)) {
907 B_PRIxADDR, range->base, range->size, fEndAddress);
917 previousRange->base, previousRange->size, range,
918 range->base, range->size);
923 nextBase = range->base + range->size;
945 "free list %d", range, range->base, range->size,
949 if (fRangeTree.Find(range->base) != range) {
952 "free list %d", range, range->base, range->size,
959 "wrong free list %d", range, range->base, range->size,