1 /* $KAME: if_altq.h,v 1.11 2003/07/10 12:07:50 kjc Exp $ */ 2 3 /* 4 * Copyright (C) 1997-2003 5 * Sony Computer Science Laboratories Inc. All rights reserved. 6 * 7 * Redistribution and use in source and binary forms, with or without 8 * modification, are permitted provided that the following conditions 9 * are met: 10 * 1. Redistributions of source code must retain the above copyright 11 * notice, this list of conditions and the following disclaimer. 12 * 2. Redistributions in binary form must reproduce the above copyright 13 * notice, this list of conditions and the following disclaimer in the 14 * documentation and/or other materials provided with the distribution. 15 * 16 * THIS SOFTWARE IS PROVIDED BY SONY CSL AND CONTRIBUTORS ``AS IS'' AND 17 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 18 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 19 * ARE DISCLAIMED. IN NO EVENT SHALL SONY CSL OR CONTRIBUTORS BE LIABLE 20 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 21 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 22 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 23 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 24 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 25 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 26 * SUCH DAMAGE. 27 */ 28 #ifndef _NET_ALTQ_IF_ALTQ_H_ 29 #define _NET_ALTQ_IF_ALTQ_H_ 30 31 #ifndef _SYS_SERIALIZE_H_ 32 #include <sys/serialize.h> 33 #endif 34 35 /* Default subqueue */ 36 #define ALTQ_SUBQ_INDEX_DEFAULT 0 37 38 struct mbuf; 39 struct altq_pktattr; 40 41 struct ifaltq_subque; 42 struct ifaltq; 43 44 typedef int (*altq_mapsubq_t)(struct ifaltq *, int); 45 46 typedef int (*ifsq_enqueue_t)(struct ifaltq_subque *, struct mbuf *, 47 struct altq_pktattr *); 48 typedef struct mbuf *(*ifsq_dequeue_t)(struct ifaltq_subque *, 49 struct mbuf *, int); 50 typedef int (*ifsq_request_t)(struct ifaltq_subque *, int, void *); 51 52 struct ifsubq_stage { 53 struct ifaltq_subque *stg_subq; 54 int stg_cnt; 55 int stg_len; 56 uint32_t stg_flags; 57 TAILQ_ENTRY(ifsubq_stage) stg_link; 58 } __cachealign; 59 60 #define IFSQ_STAGE_FLAG_QUED 0x1 61 #define IFSQ_STAGE_FLAG_SCHED 0x2 62 63 struct ifaltq_subque { 64 struct lwkt_serialize ifsq_lock; 65 int ifsq_index; 66 67 struct ifaltq *ifsq_altq; 68 struct ifnet *ifsq_ifp; 69 void *ifsq_hw_priv; /* hw private data */ 70 71 /* fields compatible with IFQ_ macros */ 72 struct mbuf *ifq_head; 73 struct mbuf *ifq_tail; 74 int ifq_len; 75 int ifq_maxlen; 76 77 ifsq_enqueue_t ifsq_enqueue; 78 ifsq_dequeue_t ifsq_dequeue; 79 ifsq_request_t ifsq_request; 80 81 struct lwkt_serialize *ifsq_hw_serialize; 82 /* hw serializer */ 83 struct mbuf *ifsq_prepended;/* mbuf dequeued, but not yet xmit */ 84 int ifsq_started; /* ifnet.if_start interlock */ 85 int ifsq_hw_oactive;/* hw too busy, protected by driver */ 86 int ifsq_cpuid; /* owner cpu */ 87 struct ifsubq_stage *ifsq_stage;/* packet staging information */ 88 struct netmsg_base *ifsq_ifstart_nmsg; 89 /* percpu msgs to sched if_start */ 90 } __cachealign; 91 92 #ifdef _KERNEL 93 94 #define ALTQ_SQ_ASSERT_LOCKED(ifsq) ASSERT_SERIALIZED(&(ifsq)->ifsq_lock) 95 #define ALTQ_SQ_LOCK_INIT(ifsq) lwkt_serialize_init(&(ifsq)->ifsq_lock) 96 #define ALTQ_SQ_LOCK(ifsq) \ 97 lwkt_serialize_adaptive_enter(&(ifsq)->ifsq_lock) 98 #define ALTQ_SQ_UNLOCK(ifsq) lwkt_serialize_exit(&(ifsq)->ifsq_lock) 99 100 #define ASSERT_ALTQ_SQ_SERIALIZED_HW(ifsq) \ 101 ASSERT_SERIALIZED((ifsq)->ifsq_hw_serialize) 102 #define ASSERT_ALTQ_SQ_NOT_SERIALIZED_HW(ifsq) \ 103 ASSERT_NOT_SERIALIZED((ifsq)->ifsq_hw_serialize) 104 105 #endif /* _KERNEL */ 106 107 /* 108 * Structure defining a queue for a network interface. 109 */ 110 struct ifaltq { 111 /* alternate queueing related fields */ 112 int altq_type; /* discipline type */ 113 int altq_flags; /* flags (e.g. ready, in-use) */ 114 void *altq_disc; /* for discipline-specific use */ 115 struct ifnet *altq_ifp; /* back pointer to interface */ 116 117 /* classifier fields */ 118 void *altq_clfier; /* classifier-specific use */ 119 void *(*altq_classify)(struct ifaltq *, struct mbuf *, 120 struct altq_pktattr *); 121 122 /* token bucket regulator */ 123 struct tb_regulator *altq_tbr; 124 125 /* Sub-queues mapping */ 126 altq_mapsubq_t altq_mapsubq; 127 uint32_t altq_subq_mask; 128 129 /* Sub-queues */ 130 int altq_subq_cnt; 131 struct ifaltq_subque *altq_subq; 132 133 int altq_maxlen; 134 }; 135 136 #ifdef _KERNRL 137 /* COMPAT */ 138 #define ALTQ_LOCK(ifq) \ 139 ALTQ_SQ_LOCK(&(ifq)->altq_subq[ALTQ_SUBQ_INDEX_DEFAULT]) 140 /* COMPAT */ 141 #define ALTQ_UNLOCK(ifq) \ 142 ALTQ_SQ_UNLOCK(&(ifq)->altq_subq[ALTQ_SUBQ_INDEX_DEFAULT]) 143 #endif 144 145 #ifdef _KERNEL 146 147 /* 148 * packet attributes used by queueing disciplines. 149 * pattr_class is a discipline-dependent scheduling class that is 150 * set by a classifier. 151 * pattr_hdr and pattr_af may be used by a discipline to access 152 * the header within a mbuf. (e.g. ECN needs to update the CE bit) 153 * note that pattr_hdr could be stale after m_pullup, though link 154 * layer output routines usually don't use m_pullup. link-level 155 * compression also invalidates these fields. thus, pattr_hdr needs 156 * to be verified when a discipline touches the header. 157 */ 158 struct altq_pktattr { 159 void *pattr_class; /* sched class set by classifier */ 160 int pattr_af; /* address family */ 161 caddr_t pattr_hdr; /* saved header position in mbuf */ 162 }; 163 164 /* 165 * a token-bucket regulator limits the rate that a network driver can 166 * dequeue packets from the output queue. 167 * modern cards are able to buffer a large amount of packets and dequeue 168 * too many packets at a time. this bursty dequeue behavior makes it 169 * impossible to schedule packets by queueing disciplines. 170 * a token-bucket is used to control the burst size in a device 171 * independent manner. 172 */ 173 struct tb_regulator { 174 int64_t tbr_rate; /* (scaled) token bucket rate */ 175 int64_t tbr_depth; /* (scaled) token bucket depth */ 176 177 int64_t tbr_token; /* (scaled) current token */ 178 int64_t tbr_filluptime; /* (scaled) time to fill up bucket */ 179 uint64_t tbr_last; /* last time token was updated */ 180 181 int tbr_lastop; /* last dequeue operation type 182 needed for poll-and-dequeue */ 183 }; 184 185 /* if_altqflags */ 186 #define ALTQF_READY 0x01 /* driver supports alternate queueing */ 187 #define ALTQF_ENABLED 0x02 /* altq is in use */ 188 #define ALTQF_CLASSIFY 0x04 /* classify packets */ 189 #define ALTQF_DRIVER1 0x40 /* driver specific */ 190 191 /* if_altqflags set internally only: */ 192 #define ALTQF_CANTCHANGE (ALTQF_READY) 193 194 /* altq_dequeue 2nd arg */ 195 #define ALTDQ_REMOVE 1 /* dequeue mbuf from the queue */ 196 #define ALTDQ_POLL 2 /* don't dequeue mbuf from the queue */ 197 198 /* altq request types (currently only purge is defined) */ 199 #define ALTRQ_PURGE 1 /* purge all packets */ 200 201 int altq_attach(struct ifaltq *, int, void *, altq_mapsubq_t, 202 ifsq_enqueue_t, ifsq_dequeue_t, ifsq_request_t, void *, 203 void *(*)(struct ifaltq *, struct mbuf *, struct altq_pktattr *)); 204 int altq_detach(struct ifaltq *); 205 int altq_enable(struct ifaltq *); 206 int altq_disable(struct ifaltq *); 207 struct mbuf *tbr_dequeue(struct ifaltq_subque *, struct mbuf *, int); 208 extern int (*altq_input)(struct mbuf *, int); 209 #endif /* _KERNEL */ 210 211 #endif /* _NET_ALTQ_IF_ALTQ_H_ */ 212