/haiku/src/kits/tracker/ |
H A D | TrackerSettingsWindow.h | 64 void ShowPage(SettingsPage page);
|
/haiku/src/add-ons/kernel/drivers/audio/ac97/auvia/ |
H A D | auvia.c | 164 uint32* page; local 168 page = stream->dmaops_log_base; 171 page[2 * i] = stream->buffer->phy_base + i * stream->blksize; 172 page[2 * i + 1] = AUVIA_DMAOP_FLAG | stream->blksize; 175 page[2 * stream->bufcount - 1] &= ~AUVIA_DMAOP_FLAG; 176 page[2 * stream->bufcount - 1] |= AUVIA_DMAOP_EOL;
|
/haiku/src/add-ons/kernel/drivers/audio/ac97/auich/ |
H A D | auich.c | 192 uint32 *page; local 213 page = stream->dmaops_log_base; 216 page[2 * i] = ((uint32)stream->buffer->phy_base) 218 page[2 * i + 1] = AUICH_DMAF_IOC | (stream->blksize
|
/haiku/src/apps/showimage/ |
H A D | ShowImageWindow.cpp | 206 B_TRANSLATE("Previous page")); 208 B_TRANSLATE("Next page")); 471 _AddItemMenu(menu, B_TRANSLATE("First page"), 473 _AddItemMenu(menu, B_TRANSLATE("Last page"), 475 _AddItemMenu(menu, B_TRANSLATE("Previous page"), 477 _AddItemMenu(menu, B_TRANSLATE("Next page"), 479 fGoToPageMenu = new BMenu(B_TRANSLATE("Go to page")); 669 int32 page = message->FindInt32("page"); local 671 if (!first && page ! 1295 int32 page = fNavigator.CurrentPage(); local [all...] |
/haiku/src/system/kernel/vm/ |
H A D | vm.cpp | 306 out.Print("page fault %#lx %s %s, pc: %#lx", fAddress, 318 // page fault errors 344 out.Print("page fault error: no area"); 347 out.Print("page fault error: area: %ld, kernel only", fArea); 350 out.Print("page fault error: area: %ld, write protected", 354 out.Print("page fault error: area: %ld, read protected", fArea); 357 out.Print("page fault error: area: %ld, execute protected", 361 out.Print("page fault error: kernel touching bad user memory"); 364 out.Print("page fault error: no address space"); 367 out.Print("page faul 381 PageFaultDone(area_id area, VMCache* topCache, VMCache* cache, vm_page* page) argument 419 increment_page_wired_count(vm_page* page) argument 430 decrement_page_wired_count(vm_page* page) argument 439 virtual_page_address(VMArea* area, vm_page* page) argument 447 is_page_in_area(VMArea* area, vm_page* page) argument 553 map_page(VMArea* area, vm_page* page, addr_t address, uint32 protection, vm_page_reservation* reservation) argument 1467 vm_page* page = NULL; local 1713 vm_page* page = vm_page_allocate_page(&reservation, local 3070 vm_test_map_modification(vm_page* page) argument 3098 vm_clear_map_flags(vm_page* page, uint32 flags) argument 3125 vm_remove_all_page_mappings(vm_page* page) argument 3137 vm_clear_page_mapping_accessed_flags(struct vm_page *page) argument 3178 vm_remove_all_page_mappings_if_unaccessed(struct vm_page *page) argument 3823 vm_page* page = (vm_page*)(addr_t)addressValue; local 3985 vm_page* page = vm_lookup_page(physicalAddress / B_PAGE_SIZE); local 4670 vm_page* page; member in struct:PageFaultContext 4726 vm_page* page = NULL; local 5569 vm_page* page = NULL; local 5748 vm_page* page; local 5932 vm_page* page; local 6089 vm_page* page; local 6936 vm_page* page = vm_lookup_page(physicalAddress / B_PAGE_SIZE); local [all...] |
H A D | vm_page.cpp | 63 #define PAGE_ASSERT(page, condition) \ 64 ASSERT_PRINT((condition), "page: %p", (page)) 67 // this many pages will be cleared at once in the page scrubber thread 70 // maximum I/O priority of the page writer 76 // The page reserve an allocation of the certain priority must not touch. 83 // Minimum number of free pages the page daemon will try to achieve. 88 // Wait interval between page daemon runs. 98 // vm_page::usage_count buff an accessed page receives in a scan. 100 // vm_page::usage_count debuff an unaccessed page receive 434 ActivatePage(vm_page* page) argument 455 DeactivatePage(vm_page* page) argument 476 FreedPageSwap(vm_page* page) argument 509 WritePage(vm_page* page) argument 542 SetPageState(vm_page* page, uint8 newState) argument 764 list_page(vm_page* page) argument 823 struct vm_page *page; local 932 struct vm_page* page; local 1059 struct vm_page *page = queue->Head(); local 1359 track_page_usage(vm_page* page) argument 1522 free_page(vm_page* page, bool clear) argument 1593 set_page_state(vm_page *page, int pageState) argument 1694 move_page_to_appropriate_queue(vm_page *page) argument 1714 clear_page(struct vm_page *page) argument 1749 vm_page *page = &sPages[startPage + i]; local 1821 vm_page *page[SCRUB_SIZE]; local 1906 vm_page* page = sModifiedPageQueue.Head(); local 2009 SetTo(vm_page* page) argument 2111 SetTo(PageWriterRun* run, vm_page* page, int32 maxPages) argument 2129 AddPage(vm_page* page) argument 2254 AddPage(vm_page* page) argument 2406 vm_page *page = next_modified_page(maxPagesToSee); local 2557 vm_page *page; local 2590 free_cached_page(vm_page *page, bool dontWait) argument 2630 vm_page *page = find_cached_page_candidate(marker); local 2670 vm_page* page = queue.Head(); local 2763 vm_page* page = nextPage; local 2892 vm_page* page = nextPage; local 3161 vm_page* page = it.Next(); local 3246 vm_page_schedule_write_page(vm_page *page) argument 3498 vm_mark_page_inuse(page_num_t page) argument 3594 vm_page* page = queue->RemoveHeadUnlocked(); local 3716 vm_page& page = sPages[start + i]; local 3781 vm_page& page = sPages[nextIndex]; local 4000 vm_page_is_dummy(struct vm_page *page) argument 4019 vm_page_free_etc(VMCache* cache, vm_page* page, vm_page_reservation* reservation) argument 4035 vm_page_set_state(vm_page *page, int pageState) argument 4053 vm_page_requeue(struct vm_page *page, bool tail) argument [all...] |
H A D | VMCache.cpp | 71 vm_page* page; member in struct:VMCache::PageEventWaiter 340 InsertPage(VMCache* cache, vm_page* page, off_t offset) argument 343 fPage(page), 351 out.Print("vm cache insert page: cache: %p, page: %p, offset: %" 363 RemovePage(VMCache* cache, vm_page* page) argument 366 fPage(page) 373 out.Print("vm cache remove page: cache: %p, page: %p", fCache, 559 vm_cache_acquire_locked_page_cache(vm_page* page, boo argument 788 vm_page* page = pages.Lookup((page_num_t)(offset >> PAGE_SHIFT)); local 800 InsertPage(vm_page* page, off_t offset) argument 838 RemovePage(vm_page* page) argument 864 MovePage(vm_page* page, off_t offset) argument 896 MovePage(vm_page* page) argument 945 WaitForPageEvents(vm_page* page, uint32 events, bool relock) argument 1503 _NotifyPageEvents(vm_page* page, uint32 events) argument [all...] |
/haiku/src/apps/bootmanager/ |
H A D | WizardView.cpp | 43 WizardView::SetPage(WizardPageView* page) argument 45 if (fPage == page) 53 fPage = page; 54 if (page == NULL) 57 fPageContainer->AddChild(page); 121 fPageContainer = new BGroupView("page container");
|
/haiku/src/add-ons/kernel/bus_managers/virtio/ |
H A D | VirtioBalloonDevice.cpp | 150 //TRACE("allocating page %" B_PRIu32 " pages\n", i); 151 vm_page* page = vm_page_allocate_page(&reservation, local 153 if (page == NULL) { 159 info->page = page; 160 fBuffer[i] = (phys_addr_t)page->physical_page_number 178 vm_page* page = info->page; local 179 fBuffer[i] = (phys_addr_t)page->physical_page_number 181 vm_page_free(NULL, page); [all...] |
H A D | VirtioBalloonPrivate.h | 32 vm_page *page; member in struct:PageInfo
|
/haiku/src/system/kernel/arch/riscv64/ |
H A D | RISCV64VMTranslationMap.cpp | 38 WriteVmPage(vm_page* page) argument 41 (addr_t)(page->physical_page_number * B_PAGE_SIZE)); 42 switch (page->State()) { 69 if (page->busy) 74 if (page->busy_writing) 79 if (page->accessed) 84 if (page->modified) 89 if (page->unused) 94 dprintf(" usage:%3u", page->usage_count); 95 dprintf(" wired:%5u", page 129 vm_page* page = vm_lookup_page(ppn); local 473 vm_page* page = vm_lookup_page(oldPte.ppn); local 566 vm_page* page = mapping->page; local [all...] |
H A D | arch_cpu.cpp | 135 addr_t page = pages[i]; local 136 if (IS_KERNEL_ADDRESS(page)) 137 FlushTlbPage(page); 139 FlushTlbPageAsid(page, 0);
|
/haiku/headers/private/system/arch/riscv64/ |
H A D | arch_cpu_defs.h | 220 // address translation and protection (pointer to page table and flags) 255 static B_ALWAYS_INLINE void FlushTlbPageAsid(uint64 page, uint64 asid) { argument 256 asm volatile("sfence.vma %0, %1" : : "r" (page), "r" (asid) : "memory");}
|
/haiku/src/apps/terminal/ |
H A D | TermParse.cpp | 557 char page = _NextParseChar(); local 558 switch (page) {
|
/haiku/src/system/kernel/cache/ |
H A D | file_cache.cpp | 159 vm_page* page = vm_page_allocate_page(reservation, local 162 fCache->InsertPage(page, fOffset + pos); 165 page->physical_page_number * B_PAGE_SIZE, B_PAGE_SIZE); 166 fPages[i++] = page; 204 // clear partial page 298 // have been written back in order to relieve the page pressure 312 vm_page* page; local 314 (page = it.Next()) != NULL && left > 0;) { 315 if (page->State() == PAGE_STATE_CACHED && !page 395 vm_page* page = pages[pageIndex++] = vm_page_allocate_page( local 519 vm_page* page = pages[pageIndex++] = vm_page_allocate_page( local 781 vm_page* page = cache->LookupPage(offset); local 998 vm_page* page = cache->LookupPage(offset); local 1119 vm_page* page = vm_page_allocate_page(&reservation, local 1274 vm_page* page = cache->LookupPage(newSize - partialBytes); local [all...] |
/haiku/src/add-ons/kernel/file_systems/ramfs/ |
H A D | DataContainer.cpp | 295 vm_page* page = pages[index]; local 296 phys_addr_t at = (page != NULL) 297 ? (page->physical_page_number * B_PAGE_SIZE) : 0; 307 page->modified = true; 310 if (page != NULL) { 352 vm_page* page = fCache->LookupPage(offset); local 353 if (page != NULL) { 354 if (page->busy) { 355 fCache->WaitForPageEvents(page, PAGE_EVENT_NOT_BUSY, true); 359 DEBUG_PAGE_ACCESS_START(page); 405 vm_page* page = pages[index++]; local [all...] |
/haiku/src/add-ons/kernel/drivers/disk/scsi/scsi_cd/ |
H A D | scsi_cd.cpp | 130 vm_page* page = vm_page_allocate_page(&reservation, local 133 entries[numEntries].address = page->physical_page_number * B_PAGE_SIZE; 340 scsi_modepage_audio *page; local 388 // find control page and retrieve values 389 page = (scsi_modepage_audio *)((char *)buffer + sizeof(header) 392 TRACE(" page=%p, codepage=%d", page, page->header.page_code); 395 volume->port0_channel = page->ports[0].channel; 396 volume->port0_volume = page [all...] |
/haiku/src/system/kernel/ |
H A D | heap.cpp | 103 heap_page * page_list; // sorted so that the desired page is always first 160 B_PAGE_SIZE, /* page size */ 163 8, /* min count per page */ 164 16 /* max waste per page */ 170 B_PAGE_SIZE * 8, /* page size */ 173 4, /* min count per page */ 174 64 /* max waste per page */ 180 B_PAGE_SIZE * 16, /* page size */ 183 1, /* min count per page */ 184 256 /* max waste per page */ 309 dump_page(heap_page *page) argument 462 heap_page *page = &area->page_table[i]; local 570 heap_page *page = &area->page_table[i]; local 707 heap_page *page = &area->page_table[i]; local 863 heap_page *page = area->free_pages; local 937 heap_page *page = bin->page_list; local 1306 heap_link_page(heap_page *page, heap_page **list) argument 1317 heap_unlink_page(heap_page *page, heap_page **list) argument 1378 heap_page *page = &area->page_table[i]; local 1443 heap_page *page = bin->page_list; local 1625 heap_page *page = &area->page_table[((addr_t)address - area->base) local 1806 heap_page *page = &area->page_table[((addr_t)address - area->base) local [all...] |
H A D | guarded_heap.cpp | 112 out.Print("guarded heap allocate: heap: %p; page: %p; " 138 out.Print("guarded heap free: heap: %p; page: %p", fHeap, 180 guarded_heap_page& page = area.pages[startPageIndex + i]; local 181 page.flags = GUARDED_HEAP_PAGE_FLAG_USED; 183 page.team = (gKernelStartup ? 0 : team_get_current_team_id()); 184 page.thread = find_thread(NULL); 186 page.stack_trace_depth = arch_debug_get_stack_trace( 187 page.stack_trace, GUARDED_HEAP_STACK_TRACE_DEPTH, 0, 4, 190 page.allocation_size = allocationSize; 191 page 227 guarded_heap_page& page = area.pages[pageIndex]; local 470 guarded_heap_page& page = area.pages[pageIndex]; local 507 guarded_heap_page* page = &area.pages[pageIndex]; local 581 dump_guarded_heap_stack_trace(guarded_heap_page& page) argument 635 guarded_heap_page& page = area->pages[pageIndex]; local 722 guarded_heap_page& page = area->pages[i]; local 819 guarded_heap_page& page = area->pages[i]; local 886 guarded_heap_page& page = area->pages[i]; local [all...] |
/haiku/src/system/kernel/arch/arm/paging/32bit/ |
H A D | ARMVMTranslationMap32Bit.cpp | 68 vm_page* page = vm_lookup_page(address / B_PAGE_SIZE); local 69 if (!page) 70 panic("destroy_tmap: didn't find pgtable page\n"); 71 DEBUG_PAGE_ACCESS_START(page); 72 vm_page_set_state(page, PAGE_STATE_FREE); 96 // allocate a physical page mapper 102 // allocate the page directory 120 // look up the page directory's physical address 129 // get the physical page mapper 172 // check to see if a page tabl 176 vm_page *page; local 456 vm_page* page = vm_lookup_page( local 545 vm_page* page = mapping->page; local [all...] |
H A D | ARMPagingMethod32Bit.cpp | 96 "physical page pool space in virtual address space!"); 100 // allocate memory for the page table and data 106 "Failed to allocate memory for page table!"); 110 // prepare the page table 127 area_id area = create_area("physical page pool", &temp, 132 "create area for physical page pool."); 140 "physical page pool space", &temp, B_EXACT_ADDRESS, 144 "create area for physical page pool space."); 202 // create an area that can contain the page table and the slot 209 area_id dataArea = create_area_etc(B_SYSTEM_TEAM, "physical page poo 502 page_table_entry page = (physicalAddress & ARM_PTE_ADDRESS_MASK) local [all...] |
/haiku/src/system/kernel/arch/arm/ |
H A D | arch_cpu.cpp | 78 arch_cpu_invalidate_TLB_page(addr_t page) argument 83 // TLBIMVAIS(page) 85 : : "r" (page));
|
/haiku/src/system/boot/platform/efi/arch/arm64/ |
H A D | arch_mmu.cpp | 153 // Set it as block or page 185 uint64* page = NULL; local 188 page = CurrentRegime.AllocatePage(); 189 ttd.SetToTable(page, flags); 192 page = ttd.Dereference(); 198 phys_addr + offset, remainingSize, level + 1, flags, page); 332 uint64* page = NULL; local 339 page = reinterpret_cast<uint64*>(ttbr1); 348 if (page == NULL) { 349 page [all...] |
/haiku/src/add-ons/print/drivers/postscript/ |
H A D | PS.cpp | 128 PSDriver::StartPage(int page) argument 130 page ++; 131 _WritePSString("%%%%Page: %d %d\n", page, page); 169 // and rotate page contents
|
/haiku/src/add-ons/kernel/bus_managers/agp_gart/ |
H A D | agp_gart.cpp | 78 vm_page *page; member in union:aperture_memory::__anon108 544 memory->page = vm_page_allocate_page_run( 547 if (memory->page == NULL) { 563 vm_page* page = vm_page_allocate_page_run( local 566 if (page == NULL) { 573 memory->pages[i] = page + i; 677 vm_page *page; local 679 page = memory->page + index; 681 page 718 vm_page *page = memory->page; local [all...] |