/linux/arch/x86/hyperv/ |
H A D | mmu.c | 64 struct hv_tlb_flush *flush; in hyperv_flush_tlb_multi() local 78 if (unlikely(!flush)) { in hyperv_flush_tlb_multi() 90 flush->flags = 0; in hyperv_flush_tlb_multi() 92 flush->address_space = 0; in hyperv_flush_tlb_multi() 96 flush->processor_mask = 0; in hyperv_flush_tlb_multi() 143 max_gvas = (PAGE_SIZE - sizeof(*flush)) / sizeof(flush->gva_list[0]); in hyperv_flush_tlb_multi() 148 flush, NULL); in hyperv_flush_tlb_multi() 152 flush, NULL); in hyperv_flush_tlb_multi() 192 flush->flags = 0; in hyperv_flush_tlb_others_ex() 194 flush->address_space = 0; in hyperv_flush_tlb_others_ex() [all …]
|
H A D | nested.c | 22 struct hv_guest_mapping_flush *flush; in hyperv_flush_guest_mapping() local 32 flush = *this_cpu_ptr(hyperv_pcpu_input_arg); in hyperv_flush_guest_mapping() 34 if (unlikely(!flush)) { in hyperv_flush_guest_mapping() 39 flush->address_space = as; in hyperv_flush_guest_mapping() 40 flush->flags = 0; in hyperv_flush_guest_mapping() 43 flush, NULL); in hyperv_flush_guest_mapping() 56 struct hv_guest_mapping_flush_list *flush, in hyperv_fill_flush_guest_mapping_list() argument 102 if (unlikely(!flush)) { in hyperv_flush_guest_mapping_range() 107 flush->address_space = as; in hyperv_flush_guest_mapping_range() 108 flush->flags = 0; in hyperv_flush_guest_mapping_range() [all …]
|
/linux/drivers/md/dm-vdo/ |
H A D | flush.c | 114 if (flush != NULL) { in allocate_flush() 121 return flush; in allocate_flush() 251 struct vdo_flush *flush = in notify_flush() local 279 vdo_complete_flush(flush); in flush_vdo() 327 struct vdo_flush *flush = in vdo_complete_flushes() local 339 vdo_complete_flush(flush); in vdo_complete_flushes() 371 bio_list_init(&flush->bios); in initialize_flush() 410 if (flush == NULL) { in vdo_launch_flush() 420 launch_flush(flush); in vdo_launch_flush() 448 launch_flush(flush); in release_flush() [all …]
|
/linux/fs/btrfs/ |
H A D | space-info.c | 463 flush)) { in btrfs_try_granting_tickets() 1079 space_info->flush = 0; in btrfs_async_reclaim_metadata_space() 1091 space_info->flush = 0; in btrfs_async_reclaim_metadata_space() 1207 flush = FLUSH_DELALLOC; in btrfs_preempt_reclaim_metadata_space() 1212 flush = COMMIT_TRANS; in btrfs_preempt_reclaim_metadata_space() 1296 space_info->flush = 0; in btrfs_async_reclaim_data_space() 1307 space_info->flush = 0; in btrfs_async_reclaim_data_space() 1324 space_info->flush = 0; in btrfs_async_reclaim_data_space() 1357 space_info->flush = 0; in btrfs_async_reclaim_data_space() 1529 switch (flush) { in handle_reserve_ticket() [all …]
|
H A D | delalloc-space.c | 118 enum btrfs_reserve_flush_enum flush = BTRFS_RESERVE_FLUSH_DATA; in btrfs_alloc_data_chunk_ondemand() local 124 flush = BTRFS_RESERVE_FLUSH_FREE_SPACE_INODE; in btrfs_alloc_data_chunk_ondemand() 126 return btrfs_reserve_data_bytes(fs_info, bytes, flush); in btrfs_alloc_data_chunk_ondemand() 134 enum btrfs_reserve_flush_enum flush = BTRFS_RESERVE_FLUSH_DATA; in btrfs_check_data_free_space() local 143 flush = BTRFS_RESERVE_NO_FLUSH; in btrfs_check_data_free_space() 145 flush = BTRFS_RESERVE_FLUSH_FREE_SPACE_INODE; in btrfs_check_data_free_space() 147 ret = btrfs_reserve_data_bytes(fs_info, len, flush); in btrfs_check_data_free_space() 313 enum btrfs_reserve_flush_enum flush = BTRFS_RESERVE_FLUSH_ALL; in btrfs_delalloc_reserve_metadata() local 326 flush = BTRFS_RESERVE_NO_FLUSH; in btrfs_delalloc_reserve_metadata() 329 flush = BTRFS_RESERVE_FLUSH_LIMIT; in btrfs_delalloc_reserve_metadata() [all …]
|
/linux/lib/ |
H A D | decompress_inflate.c | 44 long (*flush)(void*, unsigned long), in __gunzip() 53 if (flush) { in __gunzip() 82 strm->workspace = malloc(flush ? zlib_inflate_workspacesize() : in __gunzip() 138 if (!flush) { in __gunzip() 159 if (flush && strm->next_out > out_buf) { in __gunzip() 161 if (l != flush(out_buf, l)) { in __gunzip() 193 if (flush) in __gunzip() 202 long (*flush)(void*, unsigned long), in gunzip() 207 return __gunzip(buf, len, fill, flush, out_buf, 0, pos, error); in gunzip() 212 long (*flush)(void*, unsigned long), in __decompress() [all …]
|
H A D | decompress_unxz.c | 252 long (*flush)(void *src, unsigned long size), in unxz() 268 if (fill == NULL && flush == NULL) in unxz() 276 if (flush == NULL) { in unxz() 298 if (fill == NULL && flush == NULL) { in unxz() 324 if (flush != NULL && (b.out_pos == b.out_size in unxz() 331 if (flush(b.out, b.out_pos) != (long)b.out_pos) in unxz() 341 if (flush != NULL) in unxz() 381 if (flush != NULL) in unxz() 399 long (*flush)(void*, unsigned long), in __decompress() 404 return unxz(buf, len, fill, flush, out_buf, pos, error); in __decompress()
|
H A D | decompress_unzstd.c | 167 long (*flush)(void*, unsigned long), in __unzstd() 191 if (fill == NULL && flush == NULL) in __unzstd() 227 if (flush != NULL) { in __unzstd() 308 if (flush != NULL && out.pos > 0) { in __unzstd() 309 if (out.pos != flush(out.dst, out.pos)) { in __unzstd() 335 long (*flush)(void*, unsigned long), in unzstd() 340 return __unzstd(buf, len, fill, flush, out_buf, 0, pos, error); in unzstd() 345 long (*flush)(void*, unsigned long), in __decompress() 350 return __unzstd(buf, len, fill, flush, out_buf, out_len, pos, error); in __decompress()
|
H A D | decompress_unlz4.c | 33 long (*flush)(void *, unsigned long), in unlz4() 52 } else if (!flush) { in unlz4() 174 if (flush && flush(outp, dest_len) != dest_len) in unlz4() 209 long (*flush)(void*, unsigned long), in __decompress() 215 return unlz4(buf, in_len - 4, fill, flush, output, posp, error); in __decompress()
|
H A D | decompress_unlzo.c | 99 long (*flush)(void *, unsigned long), in unlzo() 112 } else if (!flush) { in unlzo() 243 if (flush && flush(out_buf, dst_len) != dst_len) in unlzo() 279 long (*flush)(void*, unsigned long), in __decompress() 284 return unlzo(buf, len, fill, flush, out_buf, pos, error); in __decompress()
|
/linux/net/ipv4/ |
H A D | tcp_offload.c | 306 int flush = 1; in tcp_gro_receive() local 322 flush |= *(u32 *)((u8 *)th + i) ^ in tcp_gro_receive() 336 flush |= (len - 1) >= mss; in tcp_gro_receive() 339 flush |= skb_cmp_decrypted(p, skb); in tcp_gro_receive() 365 flush = len < mss; in tcp_gro_receive() 374 NAPI_GRO_CB(skb)->flush |= (flush != 0); in tcp_gro_receive() 435 if (!NAPI_GRO_CB(skb)->flush && in tcp4_gro_receive() 438 goto flush; in tcp4_gro_receive() 442 goto flush; in tcp4_gro_receive() 448 flush: in tcp4_gro_receive() [all …]
|
H A D | udp_offload.c | 447 int flush; in udp_gro_receive_segment() local 451 NAPI_GRO_CB(skb)->flush = 1; in udp_gro_receive_segment() 458 NAPI_GRO_CB(skb)->flush = 1; in udp_gro_receive_segment() 477 NAPI_GRO_CB(skb)->flush = 1; in udp_gro_receive_segment() 529 int flush = 1; in udp_gro_receive() local 564 flush = 0; in udp_gro_receive() 614 goto flush; in udp4_gro_receive() 617 if (NAPI_GRO_CB(skb)->flush) in udp4_gro_receive() 622 goto flush; in udp4_gro_receive() 635 flush: in udp4_gro_receive() [all …]
|
/linux/include/net/ |
H A D | gro.h | 37 u16 flush; member 368 NAPI_GRO_CB(skb)->flush |= flush; in skb_gro_flush_final() 372 int flush, in skb_gro_flush_final_remcsum() argument 376 NAPI_GRO_CB(skb)->flush |= flush; in skb_gro_flush_final_remcsum() 384 NAPI_GRO_CB(skb)->flush |= flush; in skb_gro_flush_final() 388 int flush, in skb_gro_flush_final_remcsum() argument 391 NAPI_GRO_CB(skb)->flush |= flush; in skb_gro_flush_final_remcsum() 452 int flush; in inet_gro_flush() local 458 return flush; in inet_gro_flush() 498 int flush; in gro_receive_network_flush() local [all …]
|
/linux/lib/zlib_deflate/ |
H A D | deflate.c | 331 int flush in zlib_deflate() argument 338 flush > Z_FINISH || flush < 0) { in zlib_deflate() 351 s->last_flush = flush; in zlib_deflate() 394 flush != Z_FINISH) { in zlib_deflate() 453 if (flush != Z_FINISH) return Z_OK; in zlib_deflate() 866 int flush in deflate_stored() argument 912 FLUSH_BLOCK(s, flush == Z_FINISH); in deflate_stored() 925 int flush in deflate_fast() argument 1008 FLUSH_BLOCK(s, flush == Z_FINISH); in deflate_fast() 1019 int flush in deflate_slow() argument [all …]
|
/linux/tools/testing/selftests/drivers/net/netdevsim/ |
H A D | nexthop.sh | 147 $IP nexthop flush &> /dev/null 182 $IP nexthop flush &> /dev/null 202 $IP nexthop flush &> /dev/null 247 $IP nexthop flush &> /dev/null 267 $IP nexthop flush &> /dev/null 289 $IP nexthop flush &> /dev/null 314 $IP nexthop flush &> /dev/null 343 $IP nexthop flush &> /dev/null 373 $IP nexthop flush &> /dev/null 422 $IP nexthop flush &> /dev/null [all …]
|
/linux/Documentation/arch/x86/ |
H A D | tlb.rst | 12 from areas other than the one we are trying to flush will be 21 1. The size of the flush being performed. A flush of the entire 25 be no collateral damage caused by doing the global flush, and 26 all of the individual flush will have ended up being wasted 29 damage we do with a full flush. So, the larger the TLB, the 30 more attractive an individual flush looks. Data and 37 especially the contents of the TLB during a given flush. The 38 sizes of the flush will vary greatly depending on the workload as 48 This will cause us to do the global flush for more cases. 53 Despite the fact that a single individual flush on x86 is [all …]
|
/linux/lib/zlib_dfltcc/ |
H A D | dfltcc_deflate.c | 121 int flush, in dfltcc_deflate() argument 136 if (flush == Z_FULL_FLUSH) in dfltcc_deflate() 144 no_flush = flush == Z_NO_FLUSH; in dfltcc_deflate() 159 if (flush == Z_FINISH) in dfltcc_deflate() 162 if (flush == Z_FULL_FLUSH) in dfltcc_deflate() 217 need_empty_block = flush == Z_FINISH && param->bcf && !param->bhf; in dfltcc_deflate() 225 if (flush == Z_FINISH && !param->bcf) in dfltcc_deflate() 292 if (flush == Z_FINISH) { in dfltcc_deflate() 301 if (flush == Z_FULL_FLUSH) in dfltcc_deflate() 303 *result = flush == Z_NO_FLUSH ? need_more : block_done; in dfltcc_deflate()
|
H A D | dfltcc_inflate.h | 16 int flush, int *ret); 20 #define INFLATE_TYPEDO_HOOK(strm, flush) \ argument 25 action = dfltcc_inflate((strm), (flush), &ret); \
|
/linux/tools/testing/selftests/kvm/x86_64/ |
H A D | hyperv_tlb_flush.c | 217 flush->flags = HV_FLUSH_ALL_VIRTUAL_ADDRESS_SPACES; in sender_guest_code() 218 flush->processor_mask = BIT(WORKER_VCPU_ID_1); in sender_guest_code() 229 flush->flags = HV_FLUSH_ALL_VIRTUAL_ADDRESS_SPACES; in sender_guest_code() 230 flush->processor_mask = BIT(WORKER_VCPU_ID_1); in sender_guest_code() 231 flush->gva_list[0] = (u64)data->test_pages; in sender_guest_code() 245 flush->processor_mask = 0; in sender_guest_code() 258 flush->gva_list[0] = (u64)data->test_pages; in sender_guest_code() 373 flush->processor_mask = BIT(WORKER_VCPU_ID_1); in sender_guest_code() 386 flush->processor_mask = BIT(WORKER_VCPU_ID_1); in sender_guest_code() 387 flush->gva_list[0] = (u64)data->test_pages; in sender_guest_code() [all …]
|
/linux/drivers/md/ |
H A D | dm-delay.c | 41 struct delay_class flush; member 167 if (dc->flush.dev) in delay_dtr() 168 dm_put_device(ti, dc->flush.dev); in delay_dtr() 244 ret = delay_class_ctr(ti, &dc->flush, argv); in delay_ctr() 256 ret = delay_class_ctr(ti, &dc->flush, argv + 3); in delay_ctr() 262 ret = delay_class_ctr(ti, &dc->flush, argv + 6); in delay_ctr() 265 max_delay = max(max_delay, dc->flush.delay); in delay_ctr() 359 c = &dc->flush; in delay_map() 383 DMEMIT("%u %u %u", dc->read.ops, dc->write.ops, dc->flush.ops); in delay_status() 394 DMEMIT_DELAY_CLASS(&dc->flush); in delay_status() [all …]
|
/linux/drivers/gpu/drm/etnaviv/ |
H A D | etnaviv_buffer.c | 93 u32 flush = 0; in etnaviv_cmd_select_pipe() local 104 flush = VIVS_GL_FLUSH_CACHE_PE2D; in etnaviv_cmd_select_pipe() 106 flush = VIVS_GL_FLUSH_CACHE_DEPTH | VIVS_GL_FLUSH_CACHE_COLOR; in etnaviv_cmd_select_pipe() 108 CMD_LOAD_STATE(buffer, VIVS_GL_FLUSH_CACHE, flush); in etnaviv_cmd_select_pipe() 237 u32 link_target, flush = 0; in etnaviv_buffer_end() local 244 flush = VIVS_GL_FLUSH_CACHE_PE2D; in etnaviv_buffer_end() 246 flush = VIVS_GL_FLUSH_CACHE_DEPTH | in etnaviv_buffer_end() 252 if (flush) { in etnaviv_buffer_end() 268 CMD_LOAD_STATE(buffer, VIVS_GL_FLUSH_CACHE, flush); in etnaviv_buffer_end() 415 u32 flush = VIVS_MMUv2_CONFIGURATION_MODE_MASK | in etnaviv_buffer_queue() local [all …]
|
/linux/Documentation/block/ |
H A D | stat.rst | 44 flush I/Os requests number of flush I/Os processed 45 flush ticks milliseconds total wait time for flush requests 53 flush I/Os 56 These values increment when an flush I/O request completes. 58 Block layer combines flush requests and executes at most one at a time. 59 This counts flush requests executed by disk. Not tracked for partitions. 75 read ticks, write ticks, discard ticks, flush ticks
|
/linux/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_overlay.c | 63 SVGAEscapeVideoFlush flush; member 77 fill_escape(&cmd->escape, sizeof(cmd->flush)); in fill_flush() 78 cmd->flush.cmdType = SVGA_ESCAPE_VMWARE_VIDEO_FLUSH; in fill_flush() 79 cmd->flush.streamId = stream_id; in fill_flush() 93 struct vmw_escape_video_flush *flush; in vmw_overlay_send_put() local 117 fifo_size = sizeof(*cmds) + sizeof(*flush) + sizeof(*items) * num_items; in vmw_overlay_send_put() 125 flush = (struct vmw_escape_video_flush *)&items[num_items]; in vmw_overlay_send_put() 164 fill_flush(flush, arg->stream_id); in vmw_overlay_send_put() 184 struct vmw_escape_video_flush flush; in vmw_overlay_send_stop() member 206 fill_flush(&cmds->flush, stream_id); in vmw_overlay_send_stop()
|
/linux/arch/arm/mm/ |
H A D | cache-v4.S | 41 mcr p15, 0, r0, c7, c7, 0 @ flush ID cache 61 mcr p15, 0, ip, c7, c7, 0 @ flush ID cache 123 mcr p15, 0, r0, c7, c7, 0 @ flush ID cache
|
/linux/block/ |
H A D | blk-flush.c | 122 return 1 << ffz(rq->flush.seq); in blk_flush_cur_seq() 138 rq->end_io = rq->flush.saved_end_io; in blk_flush_restore_request() 173 BUG_ON(rq->flush.seq & seq); in blk_flush_complete_seq() 174 rq->flush.seq |= seq; in blk_flush_complete_seq() 387 rq->flush.seq = 0; in blk_rq_init_flush() 389 rq->flush.saved_end_io = rq->end_io; /* Usually NULL */ in blk_rq_init_flush() 446 rq->flush.seq |= REQ_FSEQ_PREFLUSH; in blk_insert_flush()
|