Lines Matching refs:to

14  *    may be used to endorse or promote products derived from this software
130 * to compiler-induced padding and alignment artifacts.
144 * words), we do want to ensure that changes to these values are carefully
178 * Attach the cluster from *m to *n, set up m_ext in *n
243 * Sanitize, 0 to run M_SANITY_ACTION, 1 to garble things so they
263 * No way to recover from that.
287 /* m_tags may only be attached to first mbuf in chain. */
341 * "Move" mbuf pkthdr from "from" to "to".
342 * "from" must have M_PKTHDR set, and "to" must be empty.
345 m_move_pkthdr(struct mbuf *to, struct mbuf *from)
350 M_ASSERTPKTHDR(to);
352 KASSERT(SLIST_EMPTY(&to->m_pkthdr.tags),
353 ("m_move_pkthdr: to has tags"));
359 if (to->m_flags & M_PKTHDR)
360 m_tag_delete_chain(to, NULL);
362 to->m_flags = (from->m_flags & M_COPYFLAGS) | (to->m_flags & M_EXT);
363 if ((to->m_flags & M_EXT) == 0)
364 to->m_data = to->m_pktdat;
365 to->m_pkthdr = from->m_pkthdr; /* especially tags */
371 * Duplicate "from"'s mbuf pkthdr in "to".
372 * "from" must have M_PKTHDR set, and "to" must be empty.
376 m_dup_pkthdr(struct mbuf *to, const struct mbuf *from, int how)
385 * assertions to trip. For now just disable them.
387 M_ASSERTPKTHDR(to);
389 KASSERT(SLIST_EMPTY(&to->m_pkthdr.tags), ("m_dup_pkthdr: to has tags"));
393 if (to->m_flags & M_PKTHDR)
394 m_tag_delete_chain(to, NULL);
396 to->m_flags = (from->m_flags & M_COPYFLAGS) | (to->m_flags & M_EXT);
397 if ((to->m_flags & M_EXT) == 0)
398 to->m_data = to->m_pktdat;
399 to->m_pkthdr = from->m_pkthdr;
400 SLIST_INIT(&to->m_pkthdr.tags);
401 return (m_tag_copy_chain(to, from, how));
406 * allocate new mbuf to prepend to chain,
434 * continuing for "len" bytes. If len is M_COPYALL, copy to end of mbuf.
664 * Concatenate mbuf chain n to m.
782 * If there is room, it will add up to max_protohdr-len extra bytes to the
783 * contiguous region in an attempt to avoid being called next time.
795 * otherwise allocate a new mbuf to prepend to the chain.
840 * the amount of empty space before the data in the new mbuf to be specified
841 * (in the event that the caller expects to prepend later).
885 * attempts to restore the chain to its original state.
889 * the "breaking point" happens to lie within a cluster mbuf. Use the
890 * M_WRITABLE() macro to check for this case.
960 * Routine to copy from device local memory into mbufs.
962 * which to begin copying the data to.
966 void (*copy)(char *from, caddr_t to, u_int len))
1080 * Append the specified data to the indicated mbuf chain,
1084 * Return 1 if able to complete the job; otherwise 0.
1126 * Apply function f to the data in an mbuf chain starting "off" bytes from
1159 * Return a pointer to mbuf/offset of location in mbuf chain.
1356 * First, try to collapse mbufs. Note that we always collapse
1357 * towards the front so we don't need to deal with moving the
1382 * Collapse consecutive mbufs to a cluster.
1410 * No place where we can collapse to a cluster; punt.
1413 * never reallocate the first mbuf to avoid moving the
1423 * Fragment an mbuf chain. There's no reason you'd ever want to do
1435 * (Random values range from 1 to 256)
1589 * to be a cluster. The primary purpose of this work is to create
1591 * secondary goal is to linearize the data so the data can be
1592 * passed to crypto hardware in the most efficient manner possible.
1605 * in front of it that we can use to coalesce. We do
1607 * also w/o having to handle them specially (i.e. convert
1642 * the previous mbuf if possible (since we have to copy
1643 * it anyway, we try to reduce the number of mbufs and
1663 * Allocate new space to hold the copy and copy the data.
1667 * don't know how to break up the non-contiguous memory when