/openbmc/linux/drivers/platform/surface/aggregator/ |
H A D | ssh_packet_layer.c | 741 struct ssh_ptl *ptl = packet->ptl; in __ssh_ptl_queue_push() local 776 struct ssh_ptl *ptl = packet->ptl; in ssh_ptl_queue_remove() local 793 struct ssh_ptl *ptl = p->ptl; in ssh_ptl_pending_push() local 833 struct ssh_ptl *ptl = packet->ptl; in ssh_ptl_pending_remove() local 853 struct ssh_ptl *ptl = READ_ONCE(p->ptl); in __ssh_ptl_complete() local 885 struct ssh_ptl *ptl = packet->ptl; in ssh_ptl_tx_can_process() local 975 struct ssh_ptl *ptl = packet->ptl; in ssh_ptl_tx_compl_success() local 1336 WRITE_ONCE(p->ptl, ptl); in ssh_ptl_submit() 1633 ptl->rx.blocked.offset = (ptl->rx.blocked.offset + 1) in ssh_ptl_rx_retransmit_check() 1646 ptl->ops.data_received(ptl, payload); in ssh_ptl_rx_dataframe() [all …]
|
H A D | ssh_packet_layer.h | 142 void ssh_ptl_destroy(struct ssh_ptl *ptl); 153 return ptl->serdev ? &ptl->serdev->dev : NULL; in ssh_ptl_get_device() 156 int ssh_ptl_tx_start(struct ssh_ptl *ptl); 157 int ssh_ptl_tx_stop(struct ssh_ptl *ptl); 158 int ssh_ptl_rx_start(struct ssh_ptl *ptl); 159 int ssh_ptl_rx_stop(struct ssh_ptl *ptl); 160 void ssh_ptl_shutdown(struct ssh_ptl *ptl); 162 int ssh_ptl_submit(struct ssh_ptl *ptl, struct ssh_packet *p); 176 static inline void ssh_ptl_tx_wakeup_transfer(struct ssh_ptl *ptl) in ssh_ptl_tx_wakeup_transfer() argument 178 if (test_bit(SSH_PTL_SF_SHUTDOWN_BIT, &ptl->state)) in ssh_ptl_tx_wakeup_transfer() [all …]
|
H A D | ssh_request_layer.h | 66 struct ssh_ptl ptl; member 94 #define rtl_dbg(r, fmt, ...) ptl_dbg(&(r)->ptl, fmt, ##__VA_ARGS__) 95 #define rtl_info(p, fmt, ...) ptl_info(&(p)->ptl, fmt, ##__VA_ARGS__) 96 #define rtl_warn(r, fmt, ...) ptl_warn(&(r)->ptl, fmt, ##__VA_ARGS__) 97 #define rtl_err(r, fmt, ...) ptl_err(&(r)->ptl, fmt, ##__VA_ARGS__) 112 return ssh_ptl_get_device(&rtl->ptl); in ssh_rtl_get_device() 123 struct ssh_ptl *ptl; in ssh_request_rtl() local 125 ptl = READ_ONCE(rqst->packet.ptl); in ssh_request_rtl() 126 return likely(ptl) ? to_ssh_rtl(ptl, ptl) : NULL; in ssh_request_rtl()
|
H A D | ssh_request_layer.c | 258 status = ssh_ptl_submit(&rtl->ptl, &rqst->packet); in ssh_rtl_tx_try_process_one() 391 if (cmpxchg(&rqst->packet.ptl, NULL, &rtl->ptl)) { in ssh_rtl_submit() 619 if (flags == fixed && !READ_ONCE(r->packet.ptl)) { in ssh_rtl_cancel_nonpending() 674 if (!READ_ONCE(r->packet.ptl)) { in ssh_rtl_cancel_pending() 911 struct ssh_rtl *rtl = to_ssh_rtl(p, ptl); in ssh_rtl_rx_command() 1033 status = ssh_ptl_init(&rtl->ptl, serdev, &ptl_ops); in ssh_rtl_init() 1067 ssh_ptl_destroy(&rtl->ptl); in ssh_rtl_destroy() 1080 status = ssh_ptl_tx_start(&rtl->ptl); in ssh_rtl_start() 1086 status = ssh_ptl_rx_start(&rtl->ptl); in ssh_rtl_start() 1089 ssh_ptl_tx_stop(&rtl->ptl); in ssh_rtl_start() [all …]
|
/openbmc/linux/drivers/media/platform/allegro-dvt/ |
H A D | nal-hevc.c | 115 if (ptl->general_profile_idc == 4 || in nal_hevc_rbsp_profile_tier_level() 117 ptl->general_profile_idc == 5 || in nal_hevc_rbsp_profile_tier_level() 119 ptl->general_profile_idc == 6 || in nal_hevc_rbsp_profile_tier_level() 121 ptl->general_profile_idc == 7 || in nal_hevc_rbsp_profile_tier_level() 123 ptl->general_profile_idc == 8 || in nal_hevc_rbsp_profile_tier_level() 125 ptl->general_profile_idc == 9 || in nal_hevc_rbsp_profile_tier_level() 127 ptl->general_profile_idc == 10 || in nal_hevc_rbsp_profile_tier_level() 138 if (ptl->general_profile_idc == 5 || in nal_hevc_rbsp_profile_tier_level() 140 ptl->general_profile_idc == 9 || in nal_hevc_rbsp_profile_tier_level() 161 if ((ptl->general_profile_idc >= 1 && ptl->general_profile_idc <= 5) || in nal_hevc_rbsp_profile_tier_level() [all …]
|
/openbmc/linux/mm/ |
H A D | page_vma_mapped.c | 24 *ptlp = pvmw->ptl; in map_pte() 70 pvmw->ptl = *ptlp; in map_pte() 71 spin_lock(pvmw->ptl); in map_pte() 178 spinlock_t *ptl; in page_vma_mapped_walk() local 263 pvmw->ptl = NULL; in page_vma_mapped_walk() 275 spin_unlock(ptl); in page_vma_mapped_walk() 295 if (pvmw->ptl) { in page_vma_mapped_walk() 297 pvmw->ptl = NULL; in page_vma_mapped_walk() 306 if (!pvmw->ptl) { in page_vma_mapped_walk() 307 pvmw->ptl = ptl; in page_vma_mapped_walk() [all …]
|
H A D | huge_memory.c | 841 spinlock_t *ptl; in insert_pfn_pmd() local 940 spinlock_t *ptl; in insert_pfn_pud() local 1583 spinlock_t *ptl; in madvise_free_huge_pmd() local 1592 if (!ptl) in madvise_free_huge_pmd() 1663 spinlock_t *ptl; in zap_huge_pmd() local 1668 if (!ptl) in zap_huge_pmd() 1809 spinlock_t *ptl; in change_huge_pmd() local 1822 if (!ptl) in change_huge_pmd() 1947 return ptl; in __pmd_trans_huge_lock() 1964 return ptl; in __pud_trans_huge_lock() [all …]
|
H A D | mincore.c | 103 spinlock_t *ptl; in mincore_pte_range() local 109 ptl = pmd_trans_huge_lock(pmd, vma); in mincore_pte_range() 110 if (ptl) { in mincore_pte_range() 112 spin_unlock(ptl); in mincore_pte_range() 116 ptep = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); in mincore_pte_range() 151 pte_unmap_unlock(ptep - 1, ptl); in mincore_pte_range()
|
H A D | memory.c | 432 spin_unlock(ptl); in pmd_install() 1402 spinlock_t *ptl; in zap_pte_range() local 1861 spinlock_t *ptl; in insert_page() local 2113 spinlock_t *ptl; in insert_pfn() local 2339 spinlock_t *ptl; in remap_pte_range() local 2568 spinlock_t *ptl; in apply_to_pte_range() local 3615 &vmf->ptl); in remove_device_exclusive_entry() 3921 &vmf->ptl); in do_swap_page() 4154 &vmf->ptl); in do_anonymous_page() 6134 if (!ptl) in ptlock_alloc() [all …]
|
H A D | khugepaged.c | 702 spin_lock(ptl); in __collapse_huge_page_copy_succeeded() 716 spin_lock(ptl); in __collapse_huge_page_copy_succeeded() 996 spinlock_t *ptl; in __collapse_huge_page_swapin() local 1021 vmf.ptl = ptl; in __collapse_huge_page_swapin() 1251 spinlock_t *ptl; in hpage_collapse_scan_pmd() local 1591 spin_lock(ptl); in collapse_pte_mapped_thp() 1640 if (ptl != pml) in collapse_pte_mapped_thp() 1645 if (ptl != pml) in collapse_pte_mapped_thp() 1690 spinlock_t *ptl; in retract_page_tables() local 1729 if (ptl != pml) in retract_page_tables() [all …]
|
H A D | hmm.c | 418 spinlock_t *ptl = pud_trans_huge_lock(pudp, walk->vma); in hmm_vma_walk_pud() local 420 if (!ptl) in hmm_vma_walk_pud() 428 spin_unlock(ptl); in hmm_vma_walk_pud() 439 spin_unlock(ptl); in hmm_vma_walk_pud() 451 spin_unlock(ptl); in hmm_vma_walk_pud() 465 spin_unlock(ptl); in hmm_vma_walk_pud() 484 spinlock_t *ptl; in hmm_vma_walk_hugetlb_entry() local 487 ptl = huge_pte_lock(hstate_vma(vma), walk->mm, pte); in hmm_vma_walk_hugetlb_entry() 499 spin_unlock(ptl); in hmm_vma_walk_hugetlb_entry() 518 spin_unlock(ptl); in hmm_vma_walk_hugetlb_entry()
|
H A D | madvise.c | 195 spinlock_t *ptl; in swapin_walk_pmd_entry() local 216 pte_unmap_unlock(ptep, ptl); in swapin_walk_pmd_entry() 226 pte_unmap_unlock(ptep, ptl); in swapin_walk_pmd_entry() 352 spinlock_t *ptl; in madvise_cold_or_pageout_pte_range() local 370 if (!ptl) in madvise_cold_or_pageout_pte_range() 396 spin_unlock(ptl); in madvise_cold_or_pageout_pte_range() 428 spin_unlock(ptl); in madvise_cold_or_pageout_pte_range() 470 pte_unmap_unlock(start_pte, ptl); in madvise_cold_or_pageout_pte_range() 530 pte_unmap_unlock(start_pte, ptl); in madvise_cold_or_pageout_pte_range() 632 spinlock_t *ptl; in madvise_free_pte_range() local [all …]
|
H A D | migrate_device.c | 66 spinlock_t *ptl; in migrate_vma_collect_pmd() local 76 ptl = pmd_lock(mm, pmdp); in migrate_vma_collect_pmd() 78 spin_unlock(ptl); in migrate_vma_collect_pmd() 84 spin_unlock(ptl); in migrate_vma_collect_pmd() 90 spin_unlock(ptl); in migrate_vma_collect_pmd() 103 ptep = pte_offset_map_lock(mm, pmdp, addr, &ptl); in migrate_vma_collect_pmd() 274 pte_unmap_unlock(ptep - 1, ptl); in migrate_vma_collect_pmd() 570 spinlock_t *ptl; in migrate_vma_insert_page() local 630 ptep = pte_offset_map_lock(mm, pmdp, addr, &ptl); in migrate_vma_insert_page() 671 pte_unmap_unlock(ptep, ptl); in migrate_vma_insert_page() [all …]
|
H A D | pgtable-generic.c | 365 spinlock_t *ptl; in __pte_offset_map_lock() local 372 ptl = pte_lockptr(mm, &pmdval); in __pte_offset_map_lock() 373 spin_lock(ptl); in __pte_offset_map_lock() 375 *ptlp = ptl; in __pte_offset_map_lock() 378 pte_unmap_unlock(pte, ptl); in __pte_offset_map_lock()
|
H A D | userfaultfd.c | 81 spinlock_t *ptl; in mfill_atomic_install_pte() local 94 dst_pte = pte_offset_map_lock(dst_mm, dst_pmd, dst_addr, &ptl); in mfill_atomic_install_pte() 135 pte_unmap_unlock(dst_pte, ptl); in mfill_atomic_install_pte() 221 spinlock_t *ptl; in mfill_atomic_pte_zeropage() local 227 dst_pte = pte_offset_map_lock(dst_vma->vm_mm, dst_pmd, dst_addr, &ptl); in mfill_atomic_pte_zeropage() 242 pte_unmap_unlock(dst_pte, ptl); in mfill_atomic_pte_zeropage() 300 spinlock_t *ptl; in mfill_atomic_pte_poison() local 304 dst_pte = pte_offset_map_lock(dst_mm, dst_pmd, dst_addr, &ptl); in mfill_atomic_pte_poison() 324 pte_unmap_unlock(dst_pte, ptl); in mfill_atomic_pte_poison()
|
/openbmc/linux/arch/arm/lib/ |
H A D | uaccess_with_memcpy.c | 31 spinlock_t *ptl; in pin_page_for_write() local 61 spin_lock(ptl); in pin_page_for_write() 64 spin_unlock(ptl); in pin_page_for_write() 69 *ptlp = ptl; in pin_page_for_write() 82 pte_unmap_unlock(pte, ptl); in pin_page_for_write() 87 *ptlp = ptl; in pin_page_for_write() 105 spinlock_t *ptl; in __copy_to_user_memcpy() local 129 pte_unmap_unlock(pte, ptl); in __copy_to_user_memcpy() 131 spin_unlock(ptl); in __copy_to_user_memcpy() 169 spinlock_t *ptl; in __clear_user_memset() local [all …]
|
/openbmc/linux/arch/arm/mm/ |
H A D | fault-armv.c | 70 static inline void do_pte_lock(spinlock_t *ptl) in do_pte_lock() argument 76 spin_lock_nested(ptl, SINGLE_DEPTH_NESTING); in do_pte_lock() 79 static inline void do_pte_unlock(spinlock_t *ptl) in do_pte_unlock() argument 81 spin_unlock(ptl); in do_pte_unlock() 84 static inline void do_pte_lock(spinlock_t *ptl) {} in do_pte_lock() argument 85 static inline void do_pte_unlock(spinlock_t *ptl) {} in do_pte_unlock() argument 91 spinlock_t *ptl; in adjust_pte() local 120 pte = pte_offset_map_nolock(vma->vm_mm, pmd, address, &ptl); in adjust_pte() 124 do_pte_lock(ptl); in adjust_pte() 128 do_pte_unlock(ptl); in adjust_pte()
|
/openbmc/linux/mm/damon/ |
H A D | vaddr.c | 305 spinlock_t *ptl; in damon_mkold_pmd_entry() local 312 spin_unlock(ptl); in damon_mkold_pmd_entry() 318 spin_unlock(ptl); in damon_mkold_pmd_entry() 321 spin_unlock(ptl); in damon_mkold_pmd_entry() 372 spinlock_t *ptl; in damon_mkold_hugetlb_entry() local 383 spin_unlock(ptl); in damon_mkold_hugetlb_entry() 442 spinlock_t *ptl; in damon_young_pmd_entry() local 454 spin_unlock(ptl); in damon_young_pmd_entry() 472 spin_unlock(ptl); in damon_young_pmd_entry() 508 spinlock_t *ptl; in damon_young_hugetlb_entry() local [all …]
|
/openbmc/linux/Documentation/translations/zh_CN/mm/ |
H A D | split_page_table_lock.rst | 63 面。这个区域与page->ptl共享存储。 79 page->ptl 82 page->ptl用于访问分割页表锁,其中'page'是包含该表的页面struct page。它 89 - 如果spinlock_t的大小大于long的大小,我们使用page->ptl作为spinlock_t 96 请不要直接访问page->ptl - -使用适当的辅助函数。
|
/openbmc/linux/arch/s390/mm/ |
H A D | pgtable.c | 815 spinlock_t *ptl; in set_guest_storage_key() local 835 spin_unlock(ptl); in set_guest_storage_key() 847 spin_unlock(ptl); in set_guest_storage_key() 850 spin_unlock(ptl); in set_guest_storage_key() 924 spinlock_t *ptl; in reset_guest_reference_bit() local 957 spin_unlock(ptl); in reset_guest_reference_bit() 988 spinlock_t *ptl; in get_guest_storage_key() local 1021 spin_unlock(ptl); in get_guest_storage_key() 1056 spinlock_t *ptl; in pgste_perform_essa() local 1161 spinlock_t *ptl; in set_pgste_bits() local [all …]
|
H A D | gmap.c | 554 spinlock_t *ptl; in __gmap_link() local 610 ptl = pmd_lock(mm, pmd); in __gmap_link() 632 spin_unlock(ptl); in __gmap_link() 688 spinlock_t *ptl; in __gmap_zap() local 867 spinlock_t **ptl) in gmap_pte_op_walk() argument 1026 spinlock_t *ptl; in gmap_protect_pte() local 1149 spinlock_t *ptl; in gmap_read_table() local 1235 spinlock_t *ptl; in gmap_protect_rmap() local 2140 spinlock_t *ptl; in gmap_shadow_page() local 2495 spinlock_t *ptl; in gmap_sync_dirty_log_pmd() local [all …]
|
/openbmc/linux/arch/powerpc/mm/ |
H A D | hugetlbpage.c | 47 unsigned int pshift, spinlock_t *ptl) in __hugepte_alloc() argument 82 spin_lock(ptl); in __hugepte_alloc() 102 spin_unlock(ptl); in __hugepte_alloc() 120 spinlock_t *ptl; in huge_pte_alloc() local 134 ptl = &mm->page_table_lock; in huge_pte_alloc() 144 ptl = pud_lockptr(mm, pu); in huge_pte_alloc() 155 ptl = pmd_lockptr(mm, pm); in huge_pte_alloc() 162 ptl = &mm->page_table_lock; in huge_pte_alloc() 170 ptl = pud_lockptr(mm, pu); in huge_pte_alloc() 177 ptl = pmd_lockptr(mm, pm); in huge_pte_alloc() [all …]
|
/openbmc/linux/arch/s390/pci/ |
H A D | pci_mmio.c | 126 spinlock_t *ptl; in SYSCALL_DEFINE3() local 172 ret = follow_pte(vma->vm_mm, mmio_addr, &ptep, &ptl); in SYSCALL_DEFINE3() 184 pte_unmap_unlock(ptep, ptl); in SYSCALL_DEFINE3() 268 spinlock_t *ptl; in SYSCALL_DEFINE3() local 311 ret = follow_pte(vma->vm_mm, mmio_addr, &ptep, &ptl); in SYSCALL_DEFINE3() 325 pte_unmap_unlock(ptep, ptl); in SYSCALL_DEFINE3()
|
/openbmc/linux/arch/powerpc/lib/ |
H A D | code-patching.c | 153 spinlock_t *ptl; in text_area_cpu_up_mm() local 174 pte = get_locked_pte(mm, addr, &ptl); in text_area_cpu_up_mm() 177 pte_unmap_unlock(pte, ptl); in text_area_cpu_up_mm() 290 spinlock_t *ptl; in __do_patch_instruction_mm() local 296 pte = get_locked_pte(patching_mm, text_poke_addr, &ptl); in __do_patch_instruction_mm() 326 pte_unmap_unlock(pte, ptl); in __do_patch_instruction_mm()
|
/openbmc/linux/fs/proc/ |
H A D | task_mmu.c | 614 spinlock_t *ptl; in smaps_pte_range() local 617 if (ptl) { in smaps_pte_range() 619 spin_unlock(ptl); in smaps_pte_range() 1163 spinlock_t *ptl; in clear_refs_pte_range() local 1167 if (ptl) { in clear_refs_pte_range() 1183 spin_unlock(ptl); in clear_refs_pte_range() 1464 spinlock_t *ptl; in pagemap_pmd_range() local 1471 if (ptl) { in pagemap_pmd_range() 1533 spin_unlock(ptl); in pagemap_pmd_range() 1871 spinlock_t *ptl; in gather_pte_stats() local [all …]
|