Home
last modified time | relevance | path

Searched refs:cache_op (Results 1 – 25 of 26) sorted by relevance

12

/linux/arch/loongarch/include/asm/
H A Dcacheflush.h51 #define cache_op(op, addr) \ macro
61 cache_op(Index_Writeback_Inv_LEAF0, addr); in flush_cache_line()
64 cache_op(Index_Writeback_Inv_LEAF1, addr); in flush_cache_line()
67 cache_op(Index_Writeback_Inv_LEAF2, addr); in flush_cache_line()
70 cache_op(Index_Writeback_Inv_LEAF3, addr); in flush_cache_line()
73 cache_op(Index_Writeback_Inv_LEAF4, addr); in flush_cache_line()
76 cache_op(Index_Writeback_Inv_LEAF5, addr); in flush_cache_line()
/linux/arch/mips/include/asm/
H A Dr4kcache.h57 #define cache_op(op, addr) \ macro
62 cache_op(Index_Invalidate_I, addr); in flush_icache_line_indexed()
67 cache_op(Index_Writeback_Inv_D, addr); in flush_dcache_line_indexed()
72 cache_op(Index_Writeback_Inv_SD, addr); in flush_scache_line_indexed()
83 cache_op(Hit_Invalidate_I, addr); in flush_icache_line()
90 cache_op(Hit_Writeback_Inv_D, addr); in flush_dcache_line()
95 cache_op(Hit_Invalidate_D, addr); in invalidate_dcache_line()
100 cache_op(Hit_Invalidate_SD, addr); in invalidate_scache_line()
105 cache_op(Hit_Writeback_Inv_SD, addr); in flush_scache_line()
192 cache_op(Page_Invalidate_T, addr); in invalidate_tcache_page()
[all …]
H A Dbmips.h98 cache_op(Index_Load_Tag_S, ZSCM_REG_BASE + offset); in bmips_read_zscm_reg()
119 cache_op(Index_Store_Tag_S, ZSCM_REG_BASE + offset); in bmips_write_zscm_reg()
/linux/arch/csky/mm/
H A Ddma-mapping.c14 static inline void cache_op(phys_addr_t paddr, size_t size, in cache_op() function
55 cache_op(page_to_phys(page), size, dma_wbinv_set_zero_range); in arch_dma_prep_coherent()
63 cache_op(paddr, size, dma_wb_range); in arch_sync_dma_for_device()
67 cache_op(paddr, size, dma_wbinv_range); in arch_sync_dma_for_device()
82 cache_op(paddr, size, dma_inv_range); in arch_sync_dma_for_cpu()
/linux/arch/mips/mm/
H A Dsc-rm7k.c99 cache_op(Page_Invalidate_T, start); in blast_rm7k_tcache()
117 cache_op(Index_Store_Tag_T, CKSEG0ADDR(i)); in __rm7k_tc_enable()
143 cache_op(Index_Store_Tag_SD, CKSEG0ADDR(i)); in __rm7k_sc_enable()
210 cache_op(Index_Store_Tag_T, begin); in __probe_tcache()
215 cache_op(Index_Load_Tag_T, addr); in __probe_tcache()
H A Dsc-r5k.c30 cache_op(R5K_Page_Invalidate_S, start); in blast_r5000_scache()
54 cache_op(R5K_Page_Invalidate_S, a); in r5k_dma_cache_inv_sc()
H A Dsc-mips.c40 cache_op(Hit_Writeback_Inv_SD, addr & almask); in mips_sc_inv()
41 cache_op(Hit_Writeback_Inv_SD, (addr + size - 1) & almask); in mips_sc_inv()
H A Dc-r4k.c1417 cache_op(Index_Store_Tag_I, begin); in probe_scache()
1418 cache_op(Index_Store_Tag_D, begin); in probe_scache()
1419 cache_op(Index_Store_Tag_SD, begin); in probe_scache()
1424 cache_op(Index_Load_Tag_SD, addr); in probe_scache()
H A Duasm-mips.c68 [insn_cache] = {M(cache_op, 0, 0, 0, 0, 0), RS | RT | SIMM},
/linux/arch/mips/kernel/
H A Dspram.c55 cache_op(Index_Store_Tag_I, CKSEG0|offset); in ispram_store_tag()
71 cache_op(Index_Load_Tag_I, CKSEG0 | offset); in ispram_load_tag()
90 cache_op(Index_Store_Tag_D, CKSEG0 | offset); in dspram_store_tag()
104 cache_op(Index_Load_Tag_D, CKSEG0 | offset); in dspram_load_tag()
H A Dmips-mt.c209 cache_op(Index_Load_Tag_D, INDEX_8); in mips_mt_set_cpuoptions()
219 cache_op(Index_Store_Tag_D, INDEX_8); in mips_mt_set_cpuoptions()
224 cache_op(Index_Store_Tag_D, INDEX_0); in mips_mt_set_cpuoptions()
H A Dperf_event_mipsxx.c733 unsigned int cache_type, cache_op, cache_result; in mipspmu_map_cache_event() local
740 cache_op = (config >> 8) & 0xff; in mipspmu_map_cache_event()
741 if (cache_op >= PERF_COUNT_HW_CACHE_OP_MAX) in mipspmu_map_cache_event()
750 [cache_op] in mipspmu_map_cache_event()
/linux/arch/mips/sgi-ip22/
H A Dip28-berr.c74 cache_op(Index_Load_Tag_S, addr); in save_cache_tags()
77 cache_op(Index_Load_Tag_S, addr | 1L); in save_cache_tags()
92 cache_op(Index_Load_Tag_D, addr); in save_cache_tags()
95 cache_op(Index_Load_Tag_D, addr | 1L); in save_cache_tags()
108 cache_op(Index_Load_Tag_I, addr); in save_cache_tags()
111 cache_op(Index_Load_Tag_I, addr | 1L); in save_cache_tags()
/linux/arch/xtensa/kernel/
H A Dperf_event.c115 unsigned int cache_type, cache_op, cache_result; in xtensa_pmu_cache_event() local
119 cache_op = (config >> 8) & 0xff; in xtensa_pmu_cache_event()
123 cache_op >= C(OP_MAX) || in xtensa_pmu_cache_event()
127 ret = xtensa_cache_ctl[cache_type][cache_op][cache_result]; in xtensa_pmu_cache_event()
/linux/arch/arc/kernel/
H A Dperf_event.c303 unsigned int cache_type, cache_op, cache_result; in arc_pmu_cache_event() local
307 cache_op = (config >> 8) & 0xff; in arc_pmu_cache_event()
311 if (cache_op >= PERF_COUNT_HW_CACHE_OP_MAX) in arc_pmu_cache_event()
316 ret = arc_pmu_cache_map[cache_type][cache_op][cache_result]; in arc_pmu_cache_event()
322 cache_type, cache_op, cache_result, ret, in arc_pmu_cache_event()
/linux/arch/loongarch/kernel/
H A Dperf_event.c604 unsigned int cache_type, cache_op, cache_result; in loongarch_pmu_map_cache_event() local
611 cache_op = (config >> 8) & 0xff; in loongarch_pmu_map_cache_event()
612 if (cache_op >= PERF_COUNT_HW_CACHE_OP_MAX) in loongarch_pmu_map_cache_event()
621 [cache_op] in loongarch_pmu_map_cache_event()
/linux/drivers/perf/
H A Darm_pmu.c127 unsigned int cache_type, cache_op, cache_result, ret; in armpmu_map_cache_event() local
133 cache_op = (config >> 8) & 0xff; in armpmu_map_cache_event()
134 if (cache_op >= PERF_COUNT_HW_CACHE_OP_MAX) in armpmu_map_cache_event()
144 ret = (int)(*cache_map)[cache_type][cache_op][cache_result]; in armpmu_map_cache_event()
H A Driscv_pmu_sbi.c443 unsigned int cache_type, cache_op, cache_result, ret; in pmu_event_find_cache() local
449 cache_op = (config >> 8) & 0xff; in pmu_event_find_cache()
450 if (cache_op >= PERF_COUNT_HW_CACHE_OP_MAX) in pmu_event_find_cache()
457 ret = pmu_cache_event_map[cache_type][cache_op][cache_result].event_idx; in pmu_event_find_cache()
/linux/arch/mips/txx9/generic/
H A Dsetup.c156 cache_op(Index_Writeback_Inv_D, addr | 0); in early_flush_dcache()
157 cache_op(Index_Writeback_Inv_D, addr | 1); in early_flush_dcache()
158 cache_op(Index_Writeback_Inv_D, addr | 2); in early_flush_dcache()
159 cache_op(Index_Writeback_Inv_D, addr | 3); in early_flush_dcache()
/linux/arch/riscv/kvm/
H A Dvcpu_pmu.c95 unsigned int cache_type, cache_op, cache_result; in kvm_pmu_get_perf_event_cache_config() local
100 cache_op = (sbi_event_code & SBI_PMU_EVENT_CACHE_OP_ID_CODE_MASK) >> in kvm_pmu_get_perf_event_cache_config()
105 cache_op >= PERF_COUNT_HW_CACHE_OP_MAX || in kvm_pmu_get_perf_event_cache_config()
109 config = cache_type | (cache_op << 8) | (cache_result << 16); in kvm_pmu_get_perf_event_cache_config()
/linux/tools/perf/util/
H A Dparse-events.c378 int len, cache_type = -1, cache_op = -1, cache_result = -1; in parse_events__decode_legacy_cache() local
388 cache_op = parse_aliases(str, evsel__hw_cache_op, in parse_events__decode_legacy_cache()
390 if (cache_op >= 0) { in parse_events__decode_legacy_cache()
391 if (!evsel__is_cache_op_valid(cache_type, cache_op)) in parse_events__decode_legacy_cache()
402 if (cache_op < 0) { in parse_events__decode_legacy_cache()
403 cache_op = parse_aliases(str, evsel__hw_cache_op, in parse_events__decode_legacy_cache()
405 if (cache_op >= 0) { in parse_events__decode_legacy_cache()
406 if (!evsel__is_cache_op_valid(cache_type, cache_op)) in parse_events__decode_legacy_cache()
418 if (cache_op == -1) in parse_events__decode_legacy_cache()
419 cache_op = PERF_COUNT_HW_CACHE_OP_READ; in parse_events__decode_legacy_cache()
[all …]
/linux/arch/csky/kernel/
H A Dperf_event.c952 unsigned int cache_type, cache_op, cache_result; in csky_pmu_cache_event() local
955 cache_op = (config >> 8) & 0xff; in csky_pmu_cache_event()
960 if (cache_op >= PERF_COUNT_HW_CACHE_OP_MAX) in csky_pmu_cache_event()
965 return csky_pmu_cache_map[cache_type][cache_op][cache_result]; in csky_pmu_cache_event()
/linux/arch/x86/events/
H A Dcore.c363 unsigned int cache_type, cache_op, cache_result; in set_ext_hw_attr() local
373 cache_op = (config >> 8) & 0xff; in set_ext_hw_attr()
374 if (cache_op >= PERF_COUNT_HW_CACHE_OP_MAX) in set_ext_hw_attr()
376 cache_op = array_index_nospec(cache_op, PERF_COUNT_HW_CACHE_OP_MAX); in set_ext_hw_attr()
383 val = hybrid_var(event->pmu, hw_cache_event_ids)[cache_type][cache_op][cache_result]; in set_ext_hw_attr()
391 attr->config1 = hybrid_var(event->pmu, hw_cache_extra_regs)[cache_type][cache_op][cache_result]; in set_ext_hw_attr()
/linux/arch/sparc/kernel/
H A Dperf_event.c1198 unsigned int cache_type, cache_op, cache_result; in sparc_map_cache_event() local
1208 cache_op = (config >> 8) & 0xff; in sparc_map_cache_event()
1209 if (cache_op >= PERF_COUNT_HW_CACHE_OP_MAX) in sparc_map_cache_event()
1216 pmap = &((*sparc_pmu->cache_map)[cache_type][cache_op][cache_result]); in sparc_map_cache_event()
/linux/arch/mips/include/uapi/asm/
H A Dinst.h34 sdl_op, sdr_op, swr_op, cache_op, enumerator

12