/freebsd-10.1-release/crypto/openssl/crypto/md2/ |
H A D | md2_one.c | 82 unsigned long chunk; 85 chunk = (n > sizeof(temp)) ? sizeof(temp) : n; 86 ebcdic2ascii(temp, d, chunk); 87 MD2_Update(&c, temp, chunk); 88 n -= chunk; 89 d += chunk;
|
/freebsd-10.1-release/contrib/libucl/src/ |
H A D | ucl_parser.c | 45 * @return new position in chunk 47 #define ucl_chunk_skipc(chunk, p) do{ \ 49 (chunk)->line ++; \ 50 (chunk)->column = 0; \ 52 else (chunk)->column ++; \ 54 (chunk)->pos ++; \ 55 (chunk)->remain --; \ 59 ucl_set_err (struct ucl_chunk *chunk, int code, const char *str, UT_string **err) argument 61 if (chunk->pos < chunk 84 struct ucl_chunk *chunk = parser->chunks; local 815 ucl_lex_number(struct ucl_parser *parser, struct ucl_chunk *chunk, ucl_object_t *obj) argument 844 ucl_lex_json_string(struct ucl_parser *parser, struct ucl_chunk *chunk, bool *need_unescape, bool *ucl_escape, bool *var_expand) argument 917 ucl_parse_key(struct ucl_parser *parser, struct ucl_chunk *chunk, bool *next_key, bool *end_of_object) argument 1136 ucl_parse_string_value(struct ucl_parser *parser, struct ucl_chunk *chunk, bool *var_expand, bool *need_unescape) argument 1210 ucl_parse_multiline_string(struct ucl_parser *parser, struct ucl_chunk *chunk, const unsigned char *term, int term_len, unsigned char const **beg, bool *var_expand) argument 1281 ucl_parse_value(struct ucl_parser *parser, struct ucl_chunk *chunk) argument 1453 ucl_parse_after_value(struct ucl_parser *parser, struct ucl_chunk *chunk) argument 1538 ucl_parse_macro_value(struct ucl_parser *parser, struct ucl_chunk *chunk, struct ucl_macro *macro, unsigned char const **macro_start, size_t *macro_len) argument 1619 struct ucl_chunk *chunk = parser->chunks; local 1923 struct ucl_chunk *chunk; local [all...] |
/freebsd-10.1-release/crypto/openssl/crypto/md4/ |
H A D | md4_one.c | 82 unsigned long chunk; 85 chunk = (n > sizeof(temp)) ? sizeof(temp) : n; 86 ebcdic2ascii(temp, d, chunk); 87 MD4_Update(&c, temp, chunk); 88 n -= chunk; 89 d += chunk;
|
/freebsd-10.1-release/crypto/openssl/crypto/md5/ |
H A D | md5_one.c | 82 unsigned long chunk; 85 chunk = (n > sizeof(temp)) ? sizeof(temp) : n; 86 ebcdic2ascii(temp, d, chunk); 87 MD5_Update(&c, temp, chunk); 88 n -= chunk; 89 d += chunk;
|
/freebsd-10.1-release/sys/ofed/drivers/net/mlx4/ |
H A D | icm.c | 48 * per chunk. 55 static void mlx4_free_icm_pages(struct mlx4_dev *dev, struct mlx4_icm_chunk *chunk) argument 59 if (chunk->nsg > 0) 60 pci_unmap_sg(dev->pdev, chunk->mem, chunk->npages, 63 for (i = 0; i < chunk->npages; ++i) 64 __free_pages(sg_page(&chunk->mem[i]), 65 get_order(chunk->mem[i].length)); 68 static void mlx4_free_icm_coherent(struct mlx4_dev *dev, struct mlx4_icm_chunk *chunk) argument 72 for (i = 0; i < chunk 80 struct mlx4_icm_chunk *chunk, *tmp; local 131 struct mlx4_icm_chunk *chunk = NULL; local 310 struct mlx4_icm_chunk *chunk; local [all...] |
H A D | icm.h | 65 struct mlx4_icm_chunk *chunk; member in struct:mlx4_icm_iter 91 iter->chunk = list_empty(&icm->chunk_list) ? 99 return !iter->chunk; 104 if (++iter->page_idx >= iter->chunk->nsg) { 105 if (iter->chunk->list.next == &iter->icm->chunk_list) { 106 iter->chunk = NULL; 110 iter->chunk = list_entry(iter->chunk->list.next, 118 return sg_dma_address(&iter->chunk->mem[iter->page_idx]); 123 return sg_dma_len(&iter->chunk [all...] |
/freebsd-10.1-release/contrib/jemalloc/src/ |
H A D | arena.c | 43 static void arena_avail_insert(arena_t *arena, arena_chunk_t *chunk, 46 static void arena_avail_remove(arena_t *arena, arena_chunk_t *chunk, 52 static void arena_chunk_dealloc(arena_t *arena, arena_chunk_t *chunk); 58 arena_chunk_t *chunk, void *arg); 62 static void arena_run_trim_head(arena_t *arena, arena_chunk_t *chunk, 64 static void arena_run_trim_tail(arena_t *arena, arena_chunk_t *chunk, 72 static void arena_dissociate_bin_run(arena_chunk_t *chunk, arena_run_t *run, 74 static void arena_dalloc_bin_run(arena_t *arena, arena_chunk_t *chunk, 76 static void arena_bin_lower_run(arena_t *arena, arena_chunk_t *chunk, 78 static void arena_ralloc_large_shrink(arena_t *arena, arena_chunk_t *chunk, 209 arena_avail_adjac_pred(arena_chunk_t *chunk, size_t pageind) argument 224 arena_avail_adjac_succ(arena_chunk_t *chunk, size_t pageind, size_t npages) argument 240 arena_avail_adjac(arena_chunk_t *chunk, size_t pageind, size_t npages) argument 248 arena_avail_insert(arena_t *arena, arena_chunk_t *chunk, size_t pageind, size_t npages, bool maybe_adjac_pred, bool maybe_adjac_succ) argument 281 arena_avail_remove(arena_t *arena, arena_chunk_t *chunk, size_t pageind, size_t npages, bool maybe_adjac_pred, bool maybe_adjac_succ) argument 338 arena_chunk_t *chunk = (arena_chunk_t *)CHUNK_ADDR2BASE(run); local 362 arena_run_zero(arena_chunk_t *chunk, size_t run_ind, size_t npages) argument 372 arena_run_page_validate_zeroed(arena_chunk_t *chunk, size_t run_ind) argument 387 arena_chunk_t *chunk; local 519 arena_chunk_t *chunk; local 597 arena_chunk_dealloc(arena_t *arena, arena_chunk_t *chunk) argument 656 arena_chunk_t *chunk; local 711 arena_chunk_purge(arena_t *arena, arena_chunk_t *chunk, bool all) argument 859 chunks_dirty_iter_cb(arena_chunk_tree_t *tree, arena_chunk_t *chunk, void *arg) argument 871 arena_chunk_t *chunk; local 971 arena_chunk_t *chunk; local 1105 arena_run_trim_head(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, size_t oldsize, size_t newsize) argument 1137 arena_run_trim_tail(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, size_t oldsize, size_t newsize, bool dirty) argument 1174 arena_chunk_t *chunk; local 1193 arena_chunk_t *chunk = CHUNK_ADDR2BASE(run); local 1205 arena_chunk_t *chunk = (arena_chunk_t *)CHUNK_ADDR2BASE(run); local 1301 arena_chunk_t *chunk; local 1521 arena_chunk_t *chunk; local 1572 arena_chunk_t *chunk; local 1593 arena_dissociate_bin_run(arena_chunk_t *chunk, arena_run_t *run, arena_bin_t *bin) argument 1616 arena_dalloc_bin_run(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, arena_bin_t *bin) argument 1667 arena_bin_lower_run(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, arena_bin_t *bin) argument 1688 arena_dalloc_bin_locked(arena_t *arena, arena_chunk_t *chunk, void *ptr, arena_chunk_map_t *mapelm) argument 1723 arena_dalloc_bin(arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t pageind, arena_chunk_map_t *mapelm) argument 1738 arena_dalloc_small(arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t pageind) argument 1753 arena_dalloc_large_locked(arena_t *arena, arena_chunk_t *chunk, void *ptr) argument 1774 arena_dalloc_large(arena_t *arena, arena_chunk_t *chunk, void *ptr) argument 1783 arena_ralloc_large_shrink(arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t oldsize, size_t size) argument 1813 arena_ralloc_large_grow(arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t oldsize, size_t size, size_t extra, bool zero) argument 1897 arena_chunk_t *chunk; local [all...] |
H A D | chunk.c | 26 /* Various chunk-related settings. */ 40 extent_tree_t *chunks_ad, void *chunk, size_t size); 57 * the current chunk allocation request is on behalf of the 88 /* Insert the leading space as a smaller chunk. */ 95 /* Insert the trailing space as a smaller chunk. */ 99 * base_node_alloc() can cause a new base chunk to be 213 chunk_record(extent_tree_t *chunks_szad, extent_tree_t *chunks_ad, void *chunk, argument 219 unzeroed = pages_purge(chunk, size); 220 VALGRIND_MAKE_MEM_NOACCESS(chunk, size); 224 * be needed, because base_node_alloc() may cause a new base chunk t 301 chunk_unmap(void *chunk, size_t size) argument 315 chunk_dealloc(void *chunk, size_t size, bool unmap) argument [all...] |
/freebsd-10.1-release/contrib/binutils/libiberty/ |
H A D | objalloc.c | 56 /* This structure appears at the start of each chunk. */ 60 /* Next chunk. */ 62 /* If this chunk contains large objects, this is the value of 63 current_ptr when this chunk was allocated. If this chunk 74 /* We ask for this much memory each time we create a chunk which is to 89 struct objalloc_chunk *chunk; local 102 chunk = (struct objalloc_chunk *) ret->chunks; 103 chunk->next = NULL; 104 chunk 134 struct objalloc_chunk *chunk; local 150 struct objalloc_chunk *chunk; local [all...] |
H A D | obstack.c | 133 /* Initialize an obstack H for use. Specify chunk size SIZE (0 means default). 146 register struct _obstack_chunk *chunk; /* points to new chunk */ 173 chunk = h->chunk = CALL_CHUNKFUN (h, h -> chunk_size); 174 if (!chunk) 176 h->next_free = h->object_base = chunk->contents; 177 h->chunk_limit = chunk->limit 178 = (char *) chunk + h->chunk_size; 179 chunk 145 register struct _obstack_chunk *chunk; /* points to new chunk */ local 190 register struct _obstack_chunk *chunk; /* points to new chunk */ local [all...] |
/freebsd-10.1-release/contrib/gcclibs/libiberty/ |
H A D | objalloc.c | 56 /* This structure appears at the start of each chunk. */ 60 /* Next chunk. */ 62 /* If this chunk contains large objects, this is the value of 63 current_ptr when this chunk was allocated. If this chunk 74 /* We ask for this much memory each time we create a chunk which is to 89 struct objalloc_chunk *chunk; local 102 chunk = (struct objalloc_chunk *) ret->chunks; 103 chunk->next = NULL; 104 chunk 134 struct objalloc_chunk *chunk; local 150 struct objalloc_chunk *chunk; local [all...] |
H A D | obstack.c | 133 /* Initialize an obstack H for use. Specify chunk size SIZE (0 means default). 146 register struct _obstack_chunk *chunk; /* points to new chunk */ 173 chunk = h->chunk = CALL_CHUNKFUN (h, h -> chunk_size); 174 if (!chunk) 176 h->next_free = h->object_base = chunk->contents; 177 h->chunk_limit = chunk->limit 178 = (char *) chunk + h->chunk_size; 179 chunk 145 register struct _obstack_chunk *chunk; /* points to new chunk */ local 190 register struct _obstack_chunk *chunk; /* points to new chunk */ local [all...] |
/freebsd-10.1-release/contrib/jemalloc/include/jemalloc/internal/ |
H A D | arena.h | 57 /* Each element of the chunk map corresponds to one page within the chunk. */ 63 * chunk map overhead accounts for a percentage of memory, rather than 162 /* Arena chunk header. */ 164 /* Arena that owns the chunk. */ 185 * Map of pages within chunk that keeps track of free/large/small. The 186 * first map_bias entries are omitted, since the chunk header does not 188 * for common chunk sizes (e.g. 4 MiB). 344 * In order to avoid rapid chunk allocation/deallocation when an arena 345 * oscillates right on the cusp of needing a new chunk, cach 484 arena_mapp_get(arena_chunk_t *chunk, size_t pageind) argument 494 arena_mapbitsp_get(arena_chunk_t *chunk, size_t pageind) argument 501 arena_mapbits_get(arena_chunk_t *chunk, size_t pageind) argument 508 arena_mapbits_unallocated_size_get(arena_chunk_t *chunk, size_t pageind) argument 518 arena_mapbits_large_size_get(arena_chunk_t *chunk, size_t pageind) argument 529 arena_mapbits_small_runind_get(arena_chunk_t *chunk, size_t pageind) argument 540 arena_mapbits_binind_get(arena_chunk_t *chunk, size_t pageind) argument [all...] |
H A D | chunk_mmap.h | 15 bool chunk_dealloc_mmap(void *chunk, size_t size);
|
H A D | chunk_dss.h | 27 bool chunk_in_dss(void *chunk);
|
/freebsd-10.1-release/contrib/vis/ |
H A D | foldit.c | 45 foldit(const char *chunk, int col, int max, int flags) argument 51 * if this chunk puts us over the limit. 54 cp = chunk;
|
/freebsd-10.1-release/sys/ofed/drivers/infiniband/core/ |
H A D | umem.c | 116 struct ib_umem_chunk *chunk, *tmp; local 119 list_for_each_entry_safe(chunk, tmp, &umem->chunk_list, list) { 120 ib_dma_unmap_sg_attrs(dev, chunk->page_list, 121 chunk->nents, DMA_BIDIRECTIONAL, &chunk->attrs); 122 for (i = 0; i < chunk->nents; ++i) { 123 struct page *page = sg_page(&chunk->page_list[i]); 128 kfree(chunk); 131 struct ib_umem_chunk *chunk, *tmp; 136 list_for_each_entry_safe(chunk, tm 174 struct ib_umem_chunk *chunk; local 518 struct ib_umem_chunk *chunk; local [all...] |
/freebsd-10.1-release/contrib/subversion/subversion/libsvn_subr/ |
H A D | eol.c | 59 apr_uintptr_t chunk = *(const apr_uintptr_t *)buf; 63 apr_uintptr_t r_test = chunk ^ SVN__R_MASK; 64 apr_uintptr_t n_test = chunk ^ SVN__N_MASK; 58 apr_uintptr_t chunk = *(const apr_uintptr_t *)buf; local
|
/freebsd-10.1-release/sys/cddl/contrib/opensolaris/uts/common/fs/zfs/ |
H A D | zap_leaf.c | 28 * chunk number n means l_chunk[n], even though the header precedes it. 44 #define CHAIN_END 0xffff /* end of the chunk chain */ 182 int chunk; local 186 chunk = l->l_phys->l_hdr.lh_freelist; 187 ASSERT3U(chunk, <, ZAP_LEAF_NUMCHUNKS(l)); 188 ASSERT3U(ZAP_LEAF_CHUNK(l, chunk).l_free.lf_type, ==, ZAP_CHUNK_FREE); 190 l->l_phys->l_hdr.lh_freelist = ZAP_LEAF_CHUNK(l, chunk).l_free.lf_next; 194 return (chunk); 198 zap_leaf_chunk_free(zap_leaf_t *l, uint16_t chunk) argument 200 struct zap_leaf_free *zlf = &ZAP_LEAF_CHUNK(l, chunk) 231 uint16_t chunk = zap_leaf_chunk_alloc(l); local 260 uint16_t chunk = *chunkp; local 275 zap_leaf_array_read(zap_leaf_t *l, uint16_t chunk, int array_int_len, int array_len, int buf_int_len, uint64_t buf_len, void *buf) argument 334 zap_leaf_array_match(zap_leaf_t *l, zap_name_t *zn, int chunk, int array_numints) argument 401 uint16_t chunk = *chunkp; local 449 uint16_t chunk; local 577 uint16_t chunk; local 676 uint64_t chunk; local 736 zap_leaf_transfer_array(zap_leaf_t *l, uint16_t chunk, zap_leaf_t *nl) argument 767 uint16_t chunk; local 853 int chunk = l->l_phys->l_hash[i]; local [all...] |
/freebsd-10.1-release/sys/ofed/drivers/infiniband/hw/mthca/ |
H A D | mthca_memfree.h | 76 struct mthca_icm_chunk *chunk; member in struct:mthca_icm_iter 103 iter->chunk = list_empty(&icm->chunk_list) ? 111 return !iter->chunk; 116 if (++iter->page_idx >= iter->chunk->nsg) { 117 if (iter->chunk->list.next == &iter->icm->chunk_list) { 118 iter->chunk = NULL; 122 iter->chunk = list_entry(iter->chunk->list.next, 130 return sg_dma_address(&iter->chunk->mem[iter->page_idx]); 135 return sg_dma_len(&iter->chunk [all...] |
/freebsd-10.1-release/crypto/openssl/crypto/evp/ |
H A D | e_des.c | 141 size_t n, chunk = EVP_MAXCHUNK / 8; local 144 if (inl < chunk) 145 chunk = inl; 147 while (inl && inl >= chunk) { 148 for (n = 0; n < chunk * 8; ++n) { 156 inl -= chunk; 157 in += chunk; 158 out += chunk; 159 if (inl < chunk) 160 chunk [all...] |
/freebsd-10.1-release/crypto/openssl/engines/ |
H A D | e_padlock.c | 926 size_t chunk, allocated = 0; local 931 if ((chunk = ctx->num)) { /* borrow chunk variable */ 936 if (chunk >= AES_BLOCK_SIZE) 940 while (chunk < AES_BLOCK_SIZE && nbytes != 0) { 941 ivp[chunk] = *(out_arg++) = *(in_arg++) ^ ivp[chunk]; 942 chunk++, nbytes--; 944 while (chunk < AES_BLOCK_SIZE && nbytes != 0) { 946 *(out_arg++) = c ^ ivp[chunk]; [all...] |
/freebsd-10.1-release/contrib/serf/buckets/ |
H A D | chunk_buckets.c | 32 serf_bucket_t *chunk; member in struct:__anon4328 46 ctx->chunk = serf_bucket_aggregate_create(allocator); 60 struct iovec vecs[66]; /* 64 + chunk trailer + EOF trailer = 66 */ 85 /* Inserting a 0 byte chunk indicates a terminator, which already happens 87 * EOF chunk to be interpreted by the server as a new request. So, 91 /* Build the chunk header. */ 96 /* Create a copy of the chunk header so we can have multiple chunks 101 serf_bucket_aggregate_append(ctx->chunk, simple_bkt); 103 /* Insert the chunk footer. */ 110 /* Insert the chunk foote [all...] |
/freebsd-10.1-release/contrib/gcclibs/libgomp/ |
H A D | iter.c | 115 /* Initial guess is a C sized chunk positioned nthreads iterations 151 long start, end, chunk, left; local 157 chunk = ws->chunk_size * ws->incr; 161 if (chunk < left) 162 chunk = left; 166 if (chunk > left) 167 chunk = left; 169 end = start + chunk; 187 long start, end, nend, chunk, incr; local 192 chunk [all...] |
/freebsd-10.1-release/gnu/usr.bin/grep/ |
H A D | obstack.c | 143 /* Initialize an obstack H for use. Specify chunk size SIZE (0 means default). 164 register struct _obstack_chunk *chunk; /* points to new chunk */ 196 chunk = h->chunk = CALL_CHUNKFUN (h, h -> chunk_size); 197 if (!chunk) 199 h->next_free = h->object_base = chunk->contents; 200 h->chunk_limit = chunk->limit 201 = (char *) chunk + h->chunk_size; 202 chunk 163 register struct _obstack_chunk *chunk; /* points to new chunk */ local 222 register struct _obstack_chunk *chunk; /* points to new chunk */ local [all...] |