1 /* $NetBSD: altq_rio.c,v 1.23 2016/05/16 09:53:59 ozaki-r Exp $ */ 2 /* $KAME: altq_rio.c,v 1.19 2005/04/13 03:44:25 suz Exp $ */ 3 4 /* 5 * Copyright (C) 1998-2003 6 * Sony Computer Science Laboratories Inc. All rights reserved. 7 * 8 * Redistribution and use in source and binary forms, with or without 9 * modification, are permitted provided that the following conditions 10 * are met: 11 * 1. Redistributions of source code must retain the above copyright 12 * notice, this list of conditions and the following disclaimer. 13 * 2. Redistributions in binary form must reproduce the above copyright 14 * notice, this list of conditions and the following disclaimer in the 15 * documentation and/or other materials provided with the distribution. 16 * 17 * THIS SOFTWARE IS PROVIDED BY SONY CSL AND CONTRIBUTORS ``AS IS'' AND 18 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 19 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 20 * ARE DISCLAIMED. IN NO EVENT SHALL SONY CSL OR CONTRIBUTORS BE LIABLE 21 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 22 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 23 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 24 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 25 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 26 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 27 * SUCH DAMAGE. 28 */ 29 /* 30 * Copyright (c) 1990-1994 Regents of the University of California. 31 * All rights reserved. 32 * 33 * Redistribution and use in source and binary forms, with or without 34 * modification, are permitted provided that the following conditions 35 * are met: 36 * 1. Redistributions of source code must retain the above copyright 37 * notice, this list of conditions and the following disclaimer. 38 * 2. Redistributions in binary form must reproduce the above copyright 39 * notice, this list of conditions and the following disclaimer in the 40 * documentation and/or other materials provided with the distribution. 41 * 3. All advertising materials mentioning features or use of this software 42 * must display the following acknowledgement: 43 * This product includes software developed by the Computer Systems 44 * Engineering Group at Lawrence Berkeley Laboratory. 45 * 4. Neither the name of the University nor of the Laboratory may be used 46 * to endorse or promote products derived from this software without 47 * specific prior written permission. 48 * 49 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND 50 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 51 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 52 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE 53 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 54 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 55 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 56 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 57 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 58 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 59 * SUCH DAMAGE. 60 */ 61 62 #include <sys/cdefs.h> 63 __KERNEL_RCSID(0, "$NetBSD: altq_rio.c,v 1.23 2016/05/16 09:53:59 ozaki-r Exp $"); 64 65 #ifdef _KERNEL_OPT 66 #include "opt_altq.h" 67 #include "opt_inet.h" 68 #include "pf.h" 69 #endif 70 71 #ifdef ALTQ_RIO /* rio is enabled by ALTQ_RIO option in opt_altq.h */ 72 73 #include <sys/param.h> 74 #include <sys/malloc.h> 75 #include <sys/mbuf.h> 76 #include <sys/socket.h> 77 #include <sys/systm.h> 78 #include <sys/errno.h> 79 #include <sys/kauth.h> 80 #if 1 /* ALTQ3_COMPAT */ 81 #include <sys/proc.h> 82 #include <sys/sockio.h> 83 #include <sys/kernel.h> 84 #endif 85 86 #include <net/if.h> 87 88 #include <netinet/in.h> 89 #include <netinet/in_systm.h> 90 #include <netinet/ip.h> 91 #ifdef INET6 92 #include <netinet/ip6.h> 93 #endif 94 95 #if NPF > 0 96 #include <net/pfvar.h> 97 #endif 98 #include <altq/altq.h> 99 #include <altq/altq_cdnr.h> 100 #include <altq/altq_red.h> 101 #include <altq/altq_rio.h> 102 #ifdef ALTQ3_COMPAT 103 #include <altq/altq_conf.h> 104 #endif 105 106 /* 107 * RIO: RED with IN/OUT bit 108 * described in 109 * "Explicit Allocation of Best Effort Packet Delivery Service" 110 * David D. Clark and Wenjia Fang, MIT Lab for Computer Science 111 * http://diffserv.lcs.mit.edu/Papers/exp-alloc-ddc-wf.{ps,pdf} 112 * 113 * this implementation is extended to support more than 2 drop precedence 114 * values as described in RFC2597 (Assured Forwarding PHB Group). 115 * 116 */ 117 /* 118 * AF DS (differentiated service) codepoints. 119 * (classes can be mapped to CBQ or H-FSC classes.) 120 * 121 * 0 1 2 3 4 5 6 7 122 * +---+---+---+---+---+---+---+---+ 123 * | CLASS |DropPre| 0 | CU | 124 * +---+---+---+---+---+---+---+---+ 125 * 126 * class 1: 001 127 * class 2: 010 128 * class 3: 011 129 * class 4: 100 130 * 131 * low drop prec: 01 132 * medium drop prec: 10 133 * high drop prec: 11 134 */ 135 136 /* normal red parameters */ 137 #define W_WEIGHT 512 /* inverse of weight of EWMA (511/512) */ 138 /* q_weight = 0.00195 */ 139 140 /* red parameters for a slow link */ 141 #define W_WEIGHT_1 128 /* inverse of weight of EWMA (127/128) */ 142 /* q_weight = 0.0078125 */ 143 144 /* red parameters for a very slow link (e.g., dialup) */ 145 #define W_WEIGHT_2 64 /* inverse of weight of EWMA (63/64) */ 146 /* q_weight = 0.015625 */ 147 148 /* fixed-point uses 12-bit decimal places */ 149 #define FP_SHIFT 12 /* fixed-point shift */ 150 151 /* red parameters for drop probability */ 152 #define INV_P_MAX 10 /* inverse of max drop probability */ 153 #define TH_MIN 5 /* min threshold */ 154 #define TH_MAX 15 /* max threshold */ 155 156 #define RIO_LIMIT 60 /* default max queue lenght */ 157 #define RIO_STATS /* collect statistics */ 158 159 #define TV_DELTA(a, b, delta) { \ 160 register int xxs; \ 161 \ 162 delta = (a)->tv_usec - (b)->tv_usec; \ 163 if ((xxs = (a)->tv_sec - (b)->tv_sec) != 0) { \ 164 if (xxs < 0) { \ 165 delta = 60000000; \ 166 } else if (xxs > 4) { \ 167 if (xxs > 60) \ 168 delta = 60000000; \ 169 else \ 170 delta += xxs * 1000000; \ 171 } else while (xxs > 0) { \ 172 delta += 1000000; \ 173 xxs--; \ 174 } \ 175 } \ 176 } 177 178 #ifdef ALTQ3_COMPAT 179 /* rio_list keeps all rio_queue_t's allocated. */ 180 static rio_queue_t *rio_list = NULL; 181 #endif 182 /* default rio parameter values */ 183 static struct redparams default_rio_params[RIO_NDROPPREC] = { 184 /* th_min, th_max, inv_pmax */ 185 { TH_MAX * 2 + TH_MIN, TH_MAX * 3, INV_P_MAX }, /* low drop precedence */ 186 { TH_MAX + TH_MIN, TH_MAX * 2, INV_P_MAX }, /* medium drop precedence */ 187 { TH_MIN, TH_MAX, INV_P_MAX } /* high drop precedence */ 188 }; 189 190 /* internal function prototypes */ 191 static int dscp2index(u_int8_t); 192 #ifdef ALTQ3_COMPAT 193 static int rio_enqueue(struct ifaltq *, struct mbuf *); 194 static struct mbuf *rio_dequeue(struct ifaltq *, int); 195 static int rio_request(struct ifaltq *, int, void *); 196 static int rio_detach(rio_queue_t *); 197 198 /* 199 * rio device interface 200 */ 201 altqdev_decl(rio); 202 203 #endif /* ALTQ3_COMPAT */ 204 205 rio_t * 206 rio_alloc(int weight, struct redparams *params, int flags, int pkttime) 207 { 208 rio_t *rp; 209 int w, i; 210 int npkts_per_sec; 211 212 rp = malloc(sizeof(rio_t), M_DEVBUF, M_WAITOK|M_ZERO); 213 if (rp == NULL) 214 return (NULL); 215 216 rp->rio_flags = flags; 217 if (pkttime == 0) 218 /* default packet time: 1000 bytes / 10Mbps * 8 * 1000000 */ 219 rp->rio_pkttime = 800; 220 else 221 rp->rio_pkttime = pkttime; 222 223 if (weight != 0) 224 rp->rio_weight = weight; 225 else { 226 /* use default */ 227 rp->rio_weight = W_WEIGHT; 228 229 /* when the link is very slow, adjust red parameters */ 230 npkts_per_sec = 1000000 / rp->rio_pkttime; 231 if (npkts_per_sec < 50) { 232 /* up to about 400Kbps */ 233 rp->rio_weight = W_WEIGHT_2; 234 } else if (npkts_per_sec < 300) { 235 /* up to about 2.4Mbps */ 236 rp->rio_weight = W_WEIGHT_1; 237 } 238 } 239 240 /* calculate wshift. weight must be power of 2 */ 241 w = rp->rio_weight; 242 for (i = 0; w > 1; i++) 243 w = w >> 1; 244 rp->rio_wshift = i; 245 w = 1 << rp->rio_wshift; 246 if (w != rp->rio_weight) { 247 printf("invalid weight value %d for red! use %d\n", 248 rp->rio_weight, w); 249 rp->rio_weight = w; 250 } 251 252 /* allocate weight table */ 253 rp->rio_wtab = wtab_alloc(rp->rio_weight); 254 255 for (i = 0; i < RIO_NDROPPREC; i++) { 256 struct dropprec_state *prec = &rp->rio_precstate[i]; 257 258 prec->avg = 0; 259 prec->idle = 1; 260 261 if (params == NULL || params[i].inv_pmax == 0) 262 prec->inv_pmax = default_rio_params[i].inv_pmax; 263 else 264 prec->inv_pmax = params[i].inv_pmax; 265 if (params == NULL || params[i].th_min == 0) 266 prec->th_min = default_rio_params[i].th_min; 267 else 268 prec->th_min = params[i].th_min; 269 if (params == NULL || params[i].th_max == 0) 270 prec->th_max = default_rio_params[i].th_max; 271 else 272 prec->th_max = params[i].th_max; 273 274 /* 275 * th_min_s and th_max_s are scaled versions of th_min 276 * and th_max to be compared with avg. 277 */ 278 prec->th_min_s = prec->th_min << (rp->rio_wshift + FP_SHIFT); 279 prec->th_max_s = prec->th_max << (rp->rio_wshift + FP_SHIFT); 280 281 /* 282 * precompute probability denominator 283 * probd = (2 * (TH_MAX-TH_MIN) / pmax) in fixed-point 284 */ 285 prec->probd = (2 * (prec->th_max - prec->th_min) 286 * prec->inv_pmax) << FP_SHIFT; 287 288 microtime(&prec->last); 289 } 290 291 return (rp); 292 } 293 294 void 295 rio_destroy(rio_t *rp) 296 { 297 wtab_destroy(rp->rio_wtab); 298 free(rp, M_DEVBUF); 299 } 300 301 void 302 rio_getstats(rio_t *rp, struct redstats *sp) 303 { 304 int i; 305 306 for (i = 0; i < RIO_NDROPPREC; i++) { 307 memcpy(sp, &rp->q_stats[i], sizeof(struct redstats)); 308 sp->q_avg = rp->rio_precstate[i].avg >> rp->rio_wshift; 309 sp++; 310 } 311 } 312 313 #if (RIO_NDROPPREC == 3) 314 /* 315 * internally, a drop precedence value is converted to an index 316 * starting from 0. 317 */ 318 static int 319 dscp2index(u_int8_t dscp) 320 { 321 int dpindex = dscp & AF_DROPPRECMASK; 322 323 if (dpindex == 0) 324 return (0); 325 return ((dpindex >> 3) - 1); 326 } 327 #endif 328 329 int 330 rio_addq(rio_t *rp, class_queue_t *q, struct mbuf *m, 331 struct altq_pktattr *pktattr) 332 { 333 int avg, droptype; 334 u_int8_t dsfield, odsfield; 335 int dpindex, i, n, t; 336 struct timeval now; 337 struct dropprec_state *prec; 338 339 dsfield = odsfield = read_dsfield(m, pktattr); 340 dpindex = dscp2index(dsfield); 341 342 /* 343 * update avg of the precedence states whose drop precedence 344 * is larger than or equal to the drop precedence of the packet 345 */ 346 now.tv_sec = 0; 347 for (i = dpindex; i < RIO_NDROPPREC; i++) { 348 prec = &rp->rio_precstate[i]; 349 avg = prec->avg; 350 if (prec->idle) { 351 prec->idle = 0; 352 if (now.tv_sec == 0) 353 microtime(&now); 354 t = (now.tv_sec - prec->last.tv_sec); 355 if (t > 60) 356 avg = 0; 357 else { 358 t = t * 1000000 + 359 (now.tv_usec - prec->last.tv_usec); 360 n = t / rp->rio_pkttime; 361 /* calculate (avg = (1 - Wq)^n * avg) */ 362 if (n > 0) 363 avg = (avg >> FP_SHIFT) * 364 pow_w(rp->rio_wtab, n); 365 } 366 } 367 368 /* run estimator. (avg is scaled by WEIGHT in fixed-point) */ 369 avg += (prec->qlen << FP_SHIFT) - (avg >> rp->rio_wshift); 370 prec->avg = avg; /* save the new value */ 371 /* 372 * count keeps a tally of arriving traffic that has not 373 * been dropped. 374 */ 375 prec->count++; 376 } 377 378 prec = &rp->rio_precstate[dpindex]; 379 avg = prec->avg; 380 381 /* see if we drop early */ 382 droptype = DTYPE_NODROP; 383 if (avg >= prec->th_min_s && prec->qlen > 1) { 384 if (avg >= prec->th_max_s) { 385 /* avg >= th_max: forced drop */ 386 droptype = DTYPE_FORCED; 387 } else if (prec->old == 0) { 388 /* first exceeds th_min */ 389 prec->count = 1; 390 prec->old = 1; 391 } else if (drop_early((avg - prec->th_min_s) >> rp->rio_wshift, 392 prec->probd, prec->count)) { 393 /* unforced drop by red */ 394 droptype = DTYPE_EARLY; 395 } 396 } else { 397 /* avg < th_min */ 398 prec->old = 0; 399 } 400 401 /* 402 * if the queue length hits the hard limit, it's a forced drop. 403 */ 404 if (droptype == DTYPE_NODROP && qlen(q) >= qlimit(q)) 405 droptype = DTYPE_FORCED; 406 407 if (droptype != DTYPE_NODROP) { 408 /* always drop incoming packet (as opposed to randomdrop) */ 409 for (i = dpindex; i < RIO_NDROPPREC; i++) 410 rp->rio_precstate[i].count = 0; 411 #ifdef RIO_STATS 412 if (droptype == DTYPE_EARLY) 413 rp->q_stats[dpindex].drop_unforced++; 414 else 415 rp->q_stats[dpindex].drop_forced++; 416 PKTCNTR_ADD(&rp->q_stats[dpindex].drop_cnt, m_pktlen(m)); 417 #endif 418 m_freem(m); 419 return (-1); 420 } 421 422 for (i = dpindex; i < RIO_NDROPPREC; i++) 423 rp->rio_precstate[i].qlen++; 424 425 /* save drop precedence index in mbuf hdr */ 426 M_SETCTX(m, (intptr_t)dpindex); 427 428 if (rp->rio_flags & RIOF_CLEARDSCP) 429 dsfield &= ~DSCP_MASK; 430 431 if (dsfield != odsfield) 432 write_dsfield(m, pktattr, dsfield); 433 434 _addq(q, m); 435 436 #ifdef RIO_STATS 437 PKTCNTR_ADD(&rp->q_stats[dpindex].xmit_cnt, m_pktlen(m)); 438 #endif 439 return (0); 440 } 441 442 struct mbuf * 443 rio_getq(rio_t *rp, class_queue_t *q) 444 { 445 struct mbuf *m; 446 int dpindex, i; 447 448 if ((m = _getq(q)) == NULL) 449 return NULL; 450 451 dpindex = M_GETCTX(m, intptr_t); 452 for (i = dpindex; i < RIO_NDROPPREC; i++) { 453 if (--rp->rio_precstate[i].qlen == 0) { 454 if (rp->rio_precstate[i].idle == 0) { 455 rp->rio_precstate[i].idle = 1; 456 microtime(&rp->rio_precstate[i].last); 457 } 458 } 459 } 460 return (m); 461 } 462 463 #ifdef ALTQ3_COMPAT 464 int 465 rioopen(dev_t dev, int flag, int fmt, 466 struct lwp *l) 467 { 468 /* everything will be done when the queueing scheme is attached. */ 469 return 0; 470 } 471 472 int 473 rioclose(dev_t dev, int flag, int fmt, 474 struct lwp *l) 475 { 476 rio_queue_t *rqp; 477 int err, error = 0; 478 479 while ((rqp = rio_list) != NULL) { 480 /* destroy all */ 481 err = rio_detach(rqp); 482 if (err != 0 && error == 0) 483 error = err; 484 } 485 486 return error; 487 } 488 489 int 490 rioioctl(dev_t dev, ioctlcmd_t cmd, void *addr, int flag, 491 struct lwp *l) 492 { 493 rio_queue_t *rqp; 494 struct rio_interface *ifacep; 495 struct ifnet *ifp; 496 int error = 0; 497 498 /* check super-user privilege */ 499 switch (cmd) { 500 case RIO_GETSTATS: 501 break; 502 default: 503 #if (__FreeBSD_version > 400000) 504 if ((error = suser(p)) != 0) 505 return (error); 506 #else 507 if ((error = kauth_authorize_network(l->l_cred, 508 KAUTH_NETWORK_ALTQ, KAUTH_REQ_NETWORK_ALTQ_RIO, NULL, 509 NULL, NULL)) != 0) 510 return (error); 511 #endif 512 break; 513 } 514 515 switch (cmd) { 516 517 case RIO_ENABLE: 518 ifacep = (struct rio_interface *)addr; 519 if ((rqp = altq_lookup(ifacep->rio_ifname, ALTQT_RIO)) == NULL) { 520 error = EBADF; 521 break; 522 } 523 error = altq_enable(rqp->rq_ifq); 524 break; 525 526 case RIO_DISABLE: 527 ifacep = (struct rio_interface *)addr; 528 if ((rqp = altq_lookup(ifacep->rio_ifname, ALTQT_RIO)) == NULL) { 529 error = EBADF; 530 break; 531 } 532 error = altq_disable(rqp->rq_ifq); 533 break; 534 535 case RIO_IF_ATTACH: 536 ifp = ifunit(((struct rio_interface *)addr)->rio_ifname); 537 if (ifp == NULL) { 538 error = ENXIO; 539 break; 540 } 541 542 /* allocate and initialize rio_queue_t */ 543 rqp = malloc(sizeof(rio_queue_t), M_DEVBUF, M_WAITOK|M_ZERO); 544 if (rqp == NULL) { 545 error = ENOMEM; 546 break; 547 } 548 549 rqp->rq_q = malloc(sizeof(class_queue_t), M_DEVBUF, 550 M_WAITOK|M_ZERO); 551 if (rqp->rq_q == NULL) { 552 free(rqp, M_DEVBUF); 553 error = ENOMEM; 554 break; 555 } 556 557 rqp->rq_rio = rio_alloc(0, NULL, 0, 0); 558 if (rqp->rq_rio == NULL) { 559 free(rqp->rq_q, M_DEVBUF); 560 free(rqp, M_DEVBUF); 561 error = ENOMEM; 562 break; 563 } 564 565 rqp->rq_ifq = &ifp->if_snd; 566 qtail(rqp->rq_q) = NULL; 567 qlen(rqp->rq_q) = 0; 568 qlimit(rqp->rq_q) = RIO_LIMIT; 569 qtype(rqp->rq_q) = Q_RIO; 570 571 /* 572 * set RIO to this ifnet structure. 573 */ 574 error = altq_attach(rqp->rq_ifq, ALTQT_RIO, rqp, 575 rio_enqueue, rio_dequeue, rio_request, 576 NULL, NULL); 577 if (error) { 578 rio_destroy(rqp->rq_rio); 579 free(rqp->rq_q, M_DEVBUF); 580 free(rqp, M_DEVBUF); 581 break; 582 } 583 584 /* add this state to the rio list */ 585 rqp->rq_next = rio_list; 586 rio_list = rqp; 587 break; 588 589 case RIO_IF_DETACH: 590 ifacep = (struct rio_interface *)addr; 591 if ((rqp = altq_lookup(ifacep->rio_ifname, ALTQT_RIO)) == NULL) { 592 error = EBADF; 593 break; 594 } 595 error = rio_detach(rqp); 596 break; 597 598 case RIO_GETSTATS: 599 do { 600 struct rio_stats *q_stats; 601 rio_t *rp; 602 int i; 603 604 q_stats = (struct rio_stats *)addr; 605 if ((rqp = altq_lookup(q_stats->iface.rio_ifname, 606 ALTQT_RIO)) == NULL) { 607 error = EBADF; 608 break; 609 } 610 611 rp = rqp->rq_rio; 612 613 q_stats->q_limit = qlimit(rqp->rq_q); 614 q_stats->weight = rp->rio_weight; 615 q_stats->flags = rp->rio_flags; 616 617 for (i = 0; i < RIO_NDROPPREC; i++) { 618 q_stats->q_len[i] = rp->rio_precstate[i].qlen; 619 memcpy(&q_stats->q_stats[i], &rp->q_stats[i], 620 sizeof(struct redstats)); 621 q_stats->q_stats[i].q_avg = 622 rp->rio_precstate[i].avg >> rp->rio_wshift; 623 624 q_stats->q_params[i].inv_pmax 625 = rp->rio_precstate[i].inv_pmax; 626 q_stats->q_params[i].th_min 627 = rp->rio_precstate[i].th_min; 628 q_stats->q_params[i].th_max 629 = rp->rio_precstate[i].th_max; 630 } 631 } while (/*CONSTCOND*/ 0); 632 break; 633 634 case RIO_CONFIG: 635 do { 636 struct rio_conf *fc; 637 rio_t *new; 638 int s, limit, i; 639 640 fc = (struct rio_conf *)addr; 641 if ((rqp = altq_lookup(fc->iface.rio_ifname, 642 ALTQT_RIO)) == NULL) { 643 error = EBADF; 644 break; 645 } 646 647 new = rio_alloc(fc->rio_weight, &fc->q_params[0], 648 fc->rio_flags, fc->rio_pkttime); 649 if (new == NULL) { 650 error = ENOMEM; 651 break; 652 } 653 654 s = splnet(); 655 _flushq(rqp->rq_q); 656 limit = fc->rio_limit; 657 if (limit < fc->q_params[RIO_NDROPPREC-1].th_max) 658 limit = fc->q_params[RIO_NDROPPREC-1].th_max; 659 qlimit(rqp->rq_q) = limit; 660 661 rio_destroy(rqp->rq_rio); 662 rqp->rq_rio = new; 663 664 splx(s); 665 666 /* write back new values */ 667 fc->rio_limit = limit; 668 for (i = 0; i < RIO_NDROPPREC; i++) { 669 fc->q_params[i].inv_pmax = 670 rqp->rq_rio->rio_precstate[i].inv_pmax; 671 fc->q_params[i].th_min = 672 rqp->rq_rio->rio_precstate[i].th_min; 673 fc->q_params[i].th_max = 674 rqp->rq_rio->rio_precstate[i].th_max; 675 } 676 } while (/*CONSTCOND*/ 0); 677 break; 678 679 case RIO_SETDEFAULTS: 680 do { 681 struct redparams *rp; 682 int i; 683 684 rp = (struct redparams *)addr; 685 for (i = 0; i < RIO_NDROPPREC; i++) 686 default_rio_params[i] = rp[i]; 687 } while (/*CONSTCOND*/ 0); 688 break; 689 690 default: 691 error = EINVAL; 692 break; 693 } 694 695 return error; 696 } 697 698 static int 699 rio_detach(rio_queue_t *rqp) 700 { 701 rio_queue_t *tmp; 702 int error = 0; 703 704 if (ALTQ_IS_ENABLED(rqp->rq_ifq)) 705 altq_disable(rqp->rq_ifq); 706 707 if ((error = altq_detach(rqp->rq_ifq))) 708 return (error); 709 710 if (rio_list == rqp) 711 rio_list = rqp->rq_next; 712 else { 713 for (tmp = rio_list; tmp != NULL; tmp = tmp->rq_next) 714 if (tmp->rq_next == rqp) { 715 tmp->rq_next = rqp->rq_next; 716 break; 717 } 718 if (tmp == NULL) 719 printf("rio_detach: no state found in rio_list!\n"); 720 } 721 722 rio_destroy(rqp->rq_rio); 723 free(rqp->rq_q, M_DEVBUF); 724 free(rqp, M_DEVBUF); 725 return (error); 726 } 727 728 /* 729 * rio support routines 730 */ 731 static int 732 rio_request(struct ifaltq *ifq, int req, void *arg) 733 { 734 rio_queue_t *rqp = (rio_queue_t *)ifq->altq_disc; 735 736 switch (req) { 737 case ALTRQ_PURGE: 738 _flushq(rqp->rq_q); 739 if (ALTQ_IS_ENABLED(ifq)) 740 ifq->ifq_len = 0; 741 break; 742 } 743 return (0); 744 } 745 746 /* 747 * enqueue routine: 748 * 749 * returns: 0 when successfully queued. 750 * ENOBUFS when drop occurs. 751 */ 752 static int 753 rio_enqueue(struct ifaltq *ifq, struct mbuf *m) 754 { 755 struct altq_pktattr pktattr; 756 rio_queue_t *rqp = (rio_queue_t *)ifq->altq_disc; 757 int error = 0; 758 759 pktattr.pattr_class = m->m_pkthdr.pattr_class; 760 pktattr.pattr_af = m->m_pkthdr.pattr_af; 761 pktattr.pattr_hdr = m->m_pkthdr.pattr_hdr; 762 763 if (rio_addq(rqp->rq_rio, rqp->rq_q, m, &pktattr) == 0) 764 ifq->ifq_len++; 765 else 766 error = ENOBUFS; 767 return error; 768 } 769 770 /* 771 * dequeue routine: 772 * must be called in splnet. 773 * 774 * returns: mbuf dequeued. 775 * NULL when no packet is available in the queue. 776 */ 777 778 static struct mbuf * 779 rio_dequeue(struct ifaltq *ifq, int op) 780 { 781 rio_queue_t *rqp = (rio_queue_t *)ifq->altq_disc; 782 struct mbuf *m = NULL; 783 784 if (op == ALTDQ_POLL) 785 return qhead(rqp->rq_q); 786 787 m = rio_getq(rqp->rq_rio, rqp->rq_q); 788 if (m != NULL) 789 ifq->ifq_len--; 790 return m; 791 } 792 793 #ifdef KLD_MODULE 794 795 static struct altqsw rio_sw = 796 {"rio", rioopen, rioclose, rioioctl}; 797 798 ALTQ_MODULE(altq_rio, ALTQT_RIO, &rio_sw); 799 MODULE_VERSION(altq_rio, 1); 800 MODULE_DEPEND(altq_rio, altq_red, 1, 1, 1); 801 802 #endif /* KLD_MODULE */ 803 #endif /* ALTQ3_COMPAT */ 804 805 #endif /* ALTQ_RIO */ 806