/dragonfly/contrib/gmp/mpf/ |
H A D | mul_ui.c | 100 mpf_t vf; in mpf_mul_ui() local 104 PTR(vf) = vp; in mpf_mul_ui() 105 SIZ(vf) = 2; in mpf_mul_ui() 106 ASSERT_CODE (PREC(vf) = 2); in mpf_mul_ui() 107 EXP(vf) = 2; in mpf_mul_ui() 108 mpf_mul (r, u, vf); in mpf_mul_ui()
|
H A D | div_ui.c | 41 mpf_t vf; in mpf_div_ui() local 43 SIZ(vf) = 2; in mpf_div_ui() 44 EXP(vf) = 2; in mpf_div_ui() 45 PTR(vf) = vl; in mpf_div_ui() 48 mpf_div (r, u, vf); in mpf_div_ui()
|
/dragonfly/contrib/mpfr/src/ |
H A D | agm.c | 34 mpfr_t u, v, uf, vf, sc1, sc2; in mpfr_agm() local 128 MPFR_TMP_INIT (vfp, vf, p, s); in mpfr_agm() 220 mpfr_add (vf, u, v, MPFR_RNDN); /* No overflow? */ in mpfr_agm() 221 mpfr_div_2ui (vf, vf, 1, MPFR_RNDN); in mpfr_agm() 237 mpfr_div (w, w, vf, MPFR_RNDN); /* 1/4*e^2/V(k) */ in mpfr_agm() 241 mpfr_sub (v, vf, w, MPFR_RNDN); in mpfr_agm() 242 err = MPFR_GET_EXP (vf) - MPFR_GET_EXP (v); /* 0 or 1 */ in mpfr_agm() 271 mpfr_swap (v, vf); in mpfr_agm()
|
/dragonfly/sys/netgraph7/ |
H A D | ng_vlan.c | 199 struct ng_vlan_filter *vf; in ng_vlan_rcvmsg() local 212 if (msg->header.arglen != sizeof(*vf)) { in ng_vlan_rcvmsg() 216 vf = (struct ng_vlan_filter *)msg->data; in ng_vlan_rcvmsg() 218 if (vf->vlan & ~EVL_VLID_MASK) { in ng_vlan_rcvmsg() 223 hook = ng_findhook(node, vf->hook); in ng_vlan_rcvmsg() 240 if (ng_vlan_findentry(priv, vf->vlan)) { in ng_vlan_rcvmsg() 253 f->vlan = vf->vlan; in ng_vlan_rcvmsg() 288 vf = &t->filter[0]; in ng_vlan_rcvmsg() 291 vf->vlan = f->vlan; in ng_vlan_rcvmsg() 292 strncpy(vf->hook, NG_HOOK_NAME(f->hook), in ng_vlan_rcvmsg() [all …]
|
/dragonfly/contrib/gcc-8.0/gcc/ |
H A D | tree-vectorizer.c | 94 poly_uint64 vf; member 164 poly_uint64 vf = 1; in adjust_simduid_builtins() local 218 vf = p->vf; in adjust_simduid_builtins() 223 t = build_int_cst (unsigned_type_node, vf); in adjust_simduid_builtins() 341 poly_uint64 vf = 1; in shrink_simd_arrays() local 348 vf = p->vf; in shrink_simd_arrays() 351 = build_array_type_nelts (TREE_TYPE (TREE_TYPE (decl)), vf); in shrink_simd_arrays() 752 simduid_to_vf_data->vf = loop_vinfo->vectorization_factor; in vectorize_loops()
|
H A D | omp-offload.c | 375 if (!targetm.simt.vf in oacc_xform_loop() 1822 int vf = targetm.simt.vf ? targetm.simt.vf () : 1; in execute_omp_device_lower() local 1843 if (vf != 1) in execute_omp_device_lower() 1845 rhs = vf == 1 ? null_pointer_node : NULL_TREE; in execute_omp_device_lower() 1849 if (vf != 1) in execute_omp_device_lower() 1855 rhs = vf == 1 ? build_zero_cst (type) : NULL_TREE; in execute_omp_device_lower() 1858 rhs = build_int_cst (type, vf); in execute_omp_device_lower() 1861 rhs = vf == 1 ? integer_zero_node : NULL_TREE; in execute_omp_device_lower() 1872 rhs = vf != 1 ? build_zero_cst (type) : NULL_TREE; in execute_omp_device_lower() 1875 rhs = vf != 1 ? build_one_cst (type) : NULL_TREE; in execute_omp_device_lower() [all …]
|
H A D | tree-vect-loop-manip.c | 413 rgroup_masks *rgm, tree vf, in vect_set_loop_masks_directly() argument 427 tree nscalars_step = vf; in vect_set_loop_masks_directly() 692 tree vf = build_int_cst (compare_type, in vect_set_loop_condition_masked() local 728 loop_cond_gsi, rgm, vf, in vect_set_loop_condition_masked() 1935 if (vf.is_constant (&const_vf) in vect_gen_vector_loop_niters() 1952 build_int_cst (type, vf)), in vect_gen_vector_loop_niters() 1960 step_vector = build_int_cst (type, vf); in vect_gen_vector_loop_niters() 1998 tree log_vf = build_int_cst (type, exact_log2 (vf)); in vect_gen_vector_loop_niters_mult_vf() 2420 bound_epilog += vf - 1; in vect_do_peeling() 2477 || !vf.is_constant ()); in vect_do_peeling() [all …]
|
H A D | graphds.c | 73 struct vertex *vf = &g->vertices[f], *vt = &g->vertices[t]; in add_edge() local 81 e->succ_next = vf->succ; in add_edge() 82 vf->succ = e; in add_edge()
|
H A D | omp-general.c | 436 poly_uint64 vf = 0; in omp_max_vf() local 438 vf = ordered_max (vf, sizes[i]); in omp_max_vf() 439 return vf; in omp_max_vf()
|
H A D | tree-vect-data-refs.c | 971 poly_uint64 vf = LOOP_VINFO_VECT_FACTOR (loop_vinfo); in vect_compute_data_ref_alignment() local 973 = multiple_p (DR_STEP_ALIGNMENT (dr) * vf, vector_alignment); in vect_compute_data_ref_alignment() 1707 poly_uint64 vf = LOOP_VINFO_VECT_FACTOR (loop_vinfo); in vect_enhance_data_refs_alignment() local 1810 ? vf * GROUP_SIZE (stmt_info) : vf); in vect_enhance_data_refs_alignment() 3244 unsigned int loop_depth, poly_uint64 vf) in dependence_distance_ge_vf() argument 3260 && maybe_lt ((unsigned HOST_WIDE_INT) abs_hwi (dist), vf)) in dependence_distance_ge_vf() 4020 poly_uint64 vf; in vect_analyze_data_refs() local 4406 *min_vf = upper_bound (*min_vf, vf); in vect_analyze_data_refs() 6154 unsigned HOST_WIDE_INT nelt, vf; in vect_shift_permute_load_chain() local 6167 if (pow2p_hwi (length) && vf > 4) in vect_shift_permute_load_chain() [all …]
|
/dragonfly/contrib/wpa_supplicant/src/pae/ |
H A D | ieee802_1x_secy_ops.c | 21 enum validate_frames vf) in secy_cp_control_validate_frames() argument 23 kay->vf = vf; in secy_cp_control_validate_frames() 256 return ops->create_receive_sc(ops->ctx, rxsc, kay->vf, kay->co); in secy_create_receive_sc()
|
H A D | ieee802_1x_kay.h | 155 enum validate_frames vf, 233 enum validate_frames vf; member
|
H A D | ieee802_1x_secy_ops.h | 22 enum validate_frames vf);
|
/dragonfly/contrib/gcc-4.7/gcc/ |
H A D | graphds.c | 75 struct vertex *vf = &g->vertices[f], *vt = &g->vertices[t]; in add_edge() local 84 e->succ_next = vf->succ; in add_edge() 85 vf->succ = e; in add_edge()
|
H A D | tree-vect-loop.c | 2464 int vf = LOOP_VINFO_VECT_FACTOR (loop_vinfo); in vect_get_known_peeling_cost() local 2468 *peel_iters_epilogue = vf/2; in vect_get_known_peeling_cost() 2487 *peel_iters_epilogue = vf; in vect_get_known_peeling_cost() 2515 int vf = LOOP_VINFO_VECT_FACTOR (loop_vinfo); in vect_estimate_min_profitable_iters() local 2641 peel_iters_prologue = vf/2; in vect_estimate_min_profitable_iters() 2648 peel_iters_epilogue = vf/2; in vect_estimate_min_profitable_iters() 2763 if ((scalar_single_iter_cost * vf) > vec_inside_cost) in vect_estimate_min_profitable_iters() 2811 min_profitable_iters < vf ? vf : min_profitable_iters; in vect_estimate_min_profitable_iters() 2997 int vf = LOOP_VINFO_VECT_FACTOR (loop_vinfo); in get_initial_def_for_induction() local 3041 ncopies = vf / nunits; in get_initial_def_for_induction() [all …]
|
H A D | tree-vect-slp.c | 2185 unsigned int i, vf = LOOP_VINFO_VECT_FACTOR (loop_vinfo); in vect_update_slp_costs_according_to_vf() local 2194 SLP_INSTANCE_INSIDE_OF_LOOP_COST (instance) *= vf in vect_update_slp_costs_according_to_vf() 2689 gimple_stmt_iterator *gsi, int vf, in vect_transform_slp_perm_load() argument 2740 ncopies = vf / SLP_INSTANCE_UNROLLING_FACTOR (slp_node_instance); in vect_transform_slp_perm_load() 2985 unsigned int i, j, vf; in vect_schedule_slp() local 2991 vf = LOOP_VINFO_VECT_FACTOR (loop_vinfo); in vect_schedule_slp() 2996 vf = 1; in vect_schedule_slp() 3003 instance, vf); in vect_schedule_slp()
|
H A D | tree-vect-stmts.c | 2978 int vf; in vectorizable_shift() local 3049 vf = LOOP_VINFO_VECT_FACTOR (loop_vinfo); in vectorizable_shift() 3051 vf = 1; in vectorizable_shift() 3185 && vf < vect_min_worthwhile_factor (code) in vectorizable_shift() 3337 int vf; in vectorizable_operation() local 3440 vf = LOOP_VINFO_VECT_FACTOR (loop_vinfo); in vectorizable_operation() 3442 vf = 1; in vectorizable_operation() 3476 || (vf < vect_min_worthwhile_factor (code) in vectorizable_operation() 3485 && vf < vect_min_worthwhile_factor (code) in vectorizable_operation() 4194 int vf; in vectorizable_load() local [all …]
|
/dragonfly/sys/dev/netif/ix/ |
H A D | ixgbe_x550.h | 64 bool enable, int vf); 80 void ixgbe_restore_mdd_vf_X550(struct ixgbe_hw *hw, u32 vf);
|
H A D | ixgbe_api.h | 210 unsigned int vf); 212 int vf); 220 void ixgbe_restore_mdd_vf(struct ixgbe_hw *hw, u32 vf);
|
H A D | ixgbe_common.h | 155 void ixgbe_set_mac_anti_spoofing(struct ixgbe_hw *hw, bool enable, int vf); 156 void ixgbe_set_vlan_anti_spoofing(struct ixgbe_hw *hw, bool enable, int vf);
|
H A D | ixgbe_api.c | 1223 void ixgbe_set_ethertype_anti_spoofing(struct ixgbe_hw *hw, bool enable, int vf) in ixgbe_set_ethertype_anti_spoofing() argument 1226 hw->mac.ops.set_ethertype_anti_spoofing(hw, enable, vf); in ixgbe_set_ethertype_anti_spoofing() 1302 void ixgbe_restore_mdd_vf(struct ixgbe_hw *hw, u32 vf) in ixgbe_restore_mdd_vf() argument 1305 hw->mac.ops.restore_mdd_vf(hw, vf); in ixgbe_restore_mdd_vf()
|
/dragonfly/contrib/wpa_supplicant/src/drivers/ |
H A D | driver_macsec_qca.c | 661 fal_rx_sc_validate_frame_e vf; in macsec_qca_create_receive_sc() local 688 vf = FAL_RX_SC_VALIDATE_FRAME_STRICT; in macsec_qca_create_receive_sc() 690 vf = FAL_RX_SC_VALIDATE_FRAME_CHECK; in macsec_qca_create_receive_sc() 692 vf = FAL_RX_SC_VALIDATE_FRAME_DISABLED; in macsec_qca_create_receive_sc() 697 vf); in macsec_qca_create_receive_sc()
|
/dragonfly/sys/dev/netif/ig_hal/ |
H A D | e1000_mbx.c | 326 v2p_mailbox |= hw->dev_spec.vf.v2p_mailbox; in e1000_read_v2p_mailbox() 327 hw->dev_spec.vf.v2p_mailbox |= v2p_mailbox & E1000_V2PMAILBOX_R2C_BITS; in e1000_read_v2p_mailbox() 348 hw->dev_spec.vf.v2p_mailbox &= ~mask; in e1000_check_for_bit_vf()
|
/dragonfly/contrib/wpa_supplicant/wpa_supplicant/ |
H A D | wpas_kay.c | 116 enum validate_frames vf, in wpas_create_receive_sc() argument 119 return wpa_drv_create_receive_sc(wpa_s, sc, conf_offset_val(co), vf); in wpas_create_receive_sc()
|
/dragonfly/contrib/wpa_supplicant/src/ap/ |
H A D | wpa_auth_kay.c | 151 enum validate_frames vf, in hapd_create_receive_sc() argument 159 conf_offset_val(co), vf); in hapd_create_receive_sc()
|