Lines Matching defs:pq
212 struct vm_pagequeue *pq;
219 vm_pageout_init_scan(struct scan_state *ss, struct vm_pagequeue *pq,
223 vm_pagequeue_assert_locked(pq);
228 TAILQ_INSERT_HEAD(&pq->pq_pl, marker, plinks.q);
230 TAILQ_INSERT_AFTER(&pq->pq_pl, after, marker, plinks.q);
234 ss->pq = pq;
238 vm_pagequeue_unlock(pq);
244 struct vm_pagequeue *pq;
246 pq = ss->pq;
247 vm_pagequeue_assert_locked(pq);
251 TAILQ_REMOVE(&pq->pq_pl, ss->marker, plinks.q);
253 pq->pq_pdpages += ss->scanned;
270 struct vm_pagequeue *pq;
274 pq = ss->pq;
279 vm_pagequeue_lock(pq);
296 TAILQ_REMOVE(&pq->pq_pl, m, plinks.q);
300 TAILQ_REMOVE(&pq->pq_pl, marker, plinks.q);
304 TAILQ_INSERT_TAIL(&pq->pq_pl, marker, plinks.q);
306 vm_pagequeue_cnt_add(pq, -ss->bq.bq_cnt);
307 vm_pagequeue_unlock(pq);
695 struct vm_pagequeue *pq;
723 pq = &vmd->vmd_pagequeues[queue];
724 vm_pagequeue_lock(pq);
725 vm_pageout_init_scan(&ss, pq, marker, NULL, pq->pq_cnt);
910 vm_pagequeue_lock(pq);
912 vm_pagequeue_unlock(pq);
960 struct vm_pagequeue *pq;
967 pq = &vmd->vmd_pagequeues[PQ_LAUNDRY];
1084 vm_pagequeue_lock(pq);
1087 vm_pagequeue_lockptr(pq), PVM, "launds", 0);
1106 vm_pagequeue_unlock(pq);
1144 struct vm_pagequeue *pq;
1151 pq = &vmd->vmd_pagequeues[PQ_ACTIVE];
1152 vm_pagequeue_lock(pq);
1160 min_scan = pq->pq_cnt;
1165 if (min_scan > 0 || (page_shortage > 0 && pq->pq_cnt > 0))
1181 max_scan = page_shortage > 0 ? pq->pq_cnt : min_scan;
1183 vm_pageout_init_scan(&ss, pq, marker, &vmd->vmd_clock[0], max_scan);
1186 vm_pagequeue_lock(pq);
1187 TAILQ_REMOVE(&pq->pq_pl, &vmd->vmd_clock[0], plinks.q);
1188 TAILQ_REMOVE(&pq->pq_pl, &vmd->vmd_clock[1], plinks.q);
1189 TAILQ_INSERT_HEAD(&pq->pq_pl, &vmd->vmd_clock[0],
1191 TAILQ_INSERT_TAIL(&pq->pq_pl, &vmd->vmd_clock[1],
1337 vm_pagequeue_lock(pq);
1338 TAILQ_REMOVE(&pq->pq_pl, &vmd->vmd_clock[0], plinks.q);
1339 TAILQ_INSERT_AFTER(&pq->pq_pl, marker, &vmd->vmd_clock[0], plinks.q);
1341 vm_pagequeue_unlock(pq);
1345 vm_pageout_reinsert_inactive_page(struct vm_pagequeue *pq, vm_page_t marker,
1350 vm_pagequeue_assert_locked(pq);
1370 struct vm_pagequeue *pq;
1376 pq = ss->pq;
1381 vm_pagequeue_lock(pq);
1382 delta += vm_pageout_reinsert_inactive_page(pq, marker, m);
1384 vm_pagequeue_lock(pq);
1386 delta += vm_pageout_reinsert_inactive_page(pq, marker, m);
1387 vm_pagequeue_cnt_add(pq, delta);
1388 vm_pagequeue_unlock(pq);
1400 struct vm_pagequeue *pq;
1426 pq = &vmd->vmd_pagequeues[PQ_INACTIVE];
1427 vm_pagequeue_lock(pq);
1428 vm_pageout_init_scan(&ss, pq, marker, NULL, pq->pq_cnt);
1615 vm_pagequeue_lock(pq);
1617 vm_pagequeue_unlock(pq);
1694 struct vm_pagequeue *pq;
1725 pq = &vmd->vmd_pagequeues[PQ_LAUNDRY];
1726 vm_pagequeue_lock(pq);
1728 (pq->pq_cnt > 0 || atomic_load_acq_int(&swapdev_enabled))) {
1740 vm_pagequeue_unlock(pq);