Function report |
Source Code:arch\x86\include\asm\pgtable.h |
Create Date:2022-07-28 06:00:13 |
Last Modify:2020-03-12 14:18:49 | Copyright©Brick |
home page | Tree |
Annotation kernel can get tool activity | Download SCCT | Chinese |
Name:pte_none
Proto:static inline int pte_none(pte_t pte)
Type:int
Parameter:
Type | Parameter | Name |
---|---|---|
pte_t | pte |
721 | Return Not (pte & ~_PAGE_KNL_ERRATUM_MASK) |
Name | Describe |
---|---|
ioremap_pte_range | |
follow_page_pte | |
get_gate_page | |
copy_pte_range | |
zap_pte_range | |
insert_page | This is the old fallback for page remapping.* For historical reasons, it only allows reserved pages. Only* old drivers should use this, and they needed to mark their* pages reserved for the old functions anyway. |
insert_pfn | |
remap_pte_range | maps a range of physical memory into the requested pages. the old* mappings are removed. any references to nonexistent pages results* in null mappings (currently treated as "copy-on-access") |
apply_to_pte_range | |
do_anonymous_page | We enter with non-exclusive mmap_sem (to exclude vma changes,* but allow concurrent faults), and pte mapped but not yet locked.* We return with mmap_sem still held, but pte unmapped and unlocked. |
alloc_set_pte | alloc_set_pte - setup new PTE entry for given page and add reverse page* mapping |
do_fault_around | do_fault_around() tries to map few pages around the fault address. The hope* is that the pages will be needed soon and this will lower the number of* faults to handle.* It uses vm_ops->map_pages() to map the pages, which skips the page if it's |
do_fault | We enter with non-exclusive mmap_sem (to exclude vma changes,* but allow concurrent faults) |
handle_pte_fault | These routines also need to handle stuff like marking pages dirty* and/or accessed for architectures that don't do it in hardware (most* RISC architectures) |
mincore_pte_range | |
move_ptes | |
page_vma_mapped_walk | page_vma_mapped_walk - check if @pvmw->page is mapped in @pvmw->vma at*@pvmw->address*@pvmw: pointer to struct page_vma_mapped_walk. page, vma, address and flags* must be set. pmd, pte and ptl must be NULL.* Returns true if the page is mapped in the vma |
vunmap_pte_range | Page table manipulation functions ** |
vmap_pte_range | |
madvise_cold_or_pageout_pte_range | |
madvise_free_pte_range | |
swap_vma_readahead | swap_vma_readahead - swap in pages in hope we need them soon*@entry: swap entry of this memory*@gfp_mask: memory allocation flags*@vmf: fault information* Returns the struct page for entry and addr, after queueing swapin.* Primitive swap readahead code |
vmemmap_pte_populate | |
shadow_mapped | |
kasan_free_pte | |
do_huge_pmd_wp_page_fallback | |
__split_huge_zero_page_pmd | |
__split_huge_pmd_locked | |
release_pte_pages | |
__collapse_huge_page_isolate | |
__collapse_huge_page_copy | |
khugepaged_scan_pmd | |
get_mctgt_type | get_mctgt_type - get target type of moving charge*@vma: the vma the pte to be checked belongs*@addr: the address corresponding to the pte to be checked*@ptent: the pte to be checked*@target: the pointer the target page or swap ent will be stored(can be |
mcopy_atomic_pte | |
mfill_zeropage_pte | |
pte_to_hmm_pfn_flags | |
hmm_vma_handle_pte | |
userfaultfd_must_wait | Verify the pagetables are still not ok after having reigstered into* the fault_pending_wqh to avoid userland having to UFFDIO_WAKE any* userfault that has already been resolved, if userfaultfd_read and* UFFDIO_COPY|ZEROPAGE are being run simultaneously on |
huge_pte_none |
Source code conversion tool public plug-in interface | X |
---|---|
Support c/c++/esqlc/java Oracle/Informix/Mysql Plug-in can realize: logical Report Code generation and batch code conversion |